00:00:00.001 Started by upstream project "autotest-spdk-v24.09-vs-dpdk-v22.11" build number 170 00:00:00.001 originally caused by: 00:00:00.001 Started by upstream project "nightly-trigger" build number 3672 00:00:00.001 originally caused by: 00:00:00.001 Started by timer 00:00:00.022 Checking out git https://review.spdk.io/gerrit/a/build_pool/jenkins_build_pool into /var/jenkins_home/workspace/raid-vg-autotest_script/33b20b30f0a51e6b52980845e0f6aa336787973ad45e341fbbf98d1b65b265d4 to read jbp/jenkins/jjb-config/jobs/autotest-downstream/autotest-vg.groovy 00:00:00.024 The recommended git tool is: git 00:00:00.025 using credential 00000000-0000-0000-0000-000000000002 00:00:00.028 > git rev-parse --resolve-git-dir /var/jenkins_home/workspace/raid-vg-autotest_script/33b20b30f0a51e6b52980845e0f6aa336787973ad45e341fbbf98d1b65b265d4/jbp/.git # timeout=10 00:00:00.044 Fetching changes from the remote Git repository 00:00:00.054 > git config remote.origin.url https://review.spdk.io/gerrit/a/build_pool/jenkins_build_pool # timeout=10 00:00:00.067 Using shallow fetch with depth 1 00:00:00.067 Fetching upstream changes from https://review.spdk.io/gerrit/a/build_pool/jenkins_build_pool 00:00:00.067 > git --version # timeout=10 00:00:00.082 > git --version # 'git version 2.39.2' 00:00:00.082 using GIT_ASKPASS to set credentials SPDKCI HTTPS Credentials 00:00:00.107 Setting http proxy: proxy-dmz.intel.com:911 00:00:00.107 > git fetch --tags --force --progress --depth=1 -- https://review.spdk.io/gerrit/a/build_pool/jenkins_build_pool refs/heads/master # timeout=5 00:00:02.687 > git rev-parse origin/FETCH_HEAD^{commit} # timeout=10 00:00:02.700 > git rev-parse FETCH_HEAD^{commit} # timeout=10 00:00:02.713 Checking out Revision db4637e8b949f278f369ec13f70585206ccd9507 (FETCH_HEAD) 00:00:02.713 > git config core.sparsecheckout # timeout=10 00:00:02.724 > git read-tree -mu HEAD # timeout=10 00:00:02.743 > git checkout -f db4637e8b949f278f369ec13f70585206ccd9507 # timeout=5 00:00:02.771 Commit message: "jenkins/jjb-config: Add missing SPDK_TEST_NVME_INTERRUPT flag" 00:00:02.771 > git rev-list --no-walk db4637e8b949f278f369ec13f70585206ccd9507 # timeout=10 00:00:02.986 [Pipeline] Start of Pipeline 00:00:03.001 [Pipeline] library 00:00:03.003 Loading library shm_lib@master 00:00:03.003 Library shm_lib@master is cached. Copying from home. 00:00:03.020 [Pipeline] node 00:00:03.034 Running on VM-host-WFP7 in /var/jenkins/workspace/raid-vg-autotest 00:00:03.036 [Pipeline] { 00:00:03.044 [Pipeline] catchError 00:00:03.046 [Pipeline] { 00:00:03.055 [Pipeline] wrap 00:00:03.061 [Pipeline] { 00:00:03.068 [Pipeline] stage 00:00:03.069 [Pipeline] { (Prologue) 00:00:03.082 [Pipeline] echo 00:00:03.083 Node: VM-host-WFP7 00:00:03.087 [Pipeline] cleanWs 00:00:03.097 [WS-CLEANUP] Deleting project workspace... 00:00:03.097 [WS-CLEANUP] Deferred wipeout is used... 00:00:03.104 [WS-CLEANUP] done 00:00:03.295 [Pipeline] setCustomBuildProperty 00:00:03.391 [Pipeline] httpRequest 00:00:04.422 [Pipeline] echo 00:00:04.424 Sorcerer 10.211.164.101 is alive 00:00:04.433 [Pipeline] retry 00:00:04.435 [Pipeline] { 00:00:04.448 [Pipeline] httpRequest 00:00:04.454 HttpMethod: GET 00:00:04.455 URL: http://10.211.164.101/packages/jbp_db4637e8b949f278f369ec13f70585206ccd9507.tar.gz 00:00:04.455 Sending request to url: http://10.211.164.101/packages/jbp_db4637e8b949f278f369ec13f70585206ccd9507.tar.gz 00:00:04.456 Response Code: HTTP/1.1 200 OK 00:00:04.457 Success: Status code 200 is in the accepted range: 200,404 00:00:04.457 Saving response body to /var/jenkins/workspace/raid-vg-autotest/jbp_db4637e8b949f278f369ec13f70585206ccd9507.tar.gz 00:00:04.601 [Pipeline] } 00:00:04.619 [Pipeline] // retry 00:00:04.627 [Pipeline] sh 00:00:04.912 + tar --no-same-owner -xf jbp_db4637e8b949f278f369ec13f70585206ccd9507.tar.gz 00:00:04.927 [Pipeline] httpRequest 00:00:05.299 [Pipeline] echo 00:00:05.300 Sorcerer 10.211.164.101 is alive 00:00:05.307 [Pipeline] retry 00:00:05.308 [Pipeline] { 00:00:05.318 [Pipeline] httpRequest 00:00:05.323 HttpMethod: GET 00:00:05.323 URL: http://10.211.164.101/packages/spdk_b18e1bd6297ec2f89ab275de3193457af1c946df.tar.gz 00:00:05.324 Sending request to url: http://10.211.164.101/packages/spdk_b18e1bd6297ec2f89ab275de3193457af1c946df.tar.gz 00:00:05.325 Response Code: HTTP/1.1 200 OK 00:00:05.326 Success: Status code 200 is in the accepted range: 200,404 00:00:05.326 Saving response body to /var/jenkins/workspace/raid-vg-autotest/spdk_b18e1bd6297ec2f89ab275de3193457af1c946df.tar.gz 00:01:19.624 [Pipeline] } 00:01:19.642 [Pipeline] // retry 00:01:19.651 [Pipeline] sh 00:01:19.933 + tar --no-same-owner -xf spdk_b18e1bd6297ec2f89ab275de3193457af1c946df.tar.gz 00:01:22.478 [Pipeline] sh 00:01:22.764 + git -C spdk log --oneline -n5 00:01:22.764 b18e1bd62 version: v24.09.1-pre 00:01:22.764 19524ad45 version: v24.09 00:01:22.764 9756b40a3 dpdk: update submodule to include alarm_cancel fix 00:01:22.764 a808500d2 test/nvmf: disable nvmf_shutdown_tc4 on e810 00:01:22.764 3024272c6 bdev/nvme: take nvme_ctrlr.mutex when setting keys 00:01:22.783 [Pipeline] withCredentials 00:01:22.794 > git --version # timeout=10 00:01:22.808 > git --version # 'git version 2.39.2' 00:01:22.825 Masking supported pattern matches of $GIT_PASSWORD or $GIT_ASKPASS 00:01:22.827 [Pipeline] { 00:01:22.835 [Pipeline] retry 00:01:22.837 [Pipeline] { 00:01:22.852 [Pipeline] sh 00:01:23.133 + git ls-remote http://dpdk.org/git/dpdk-stable v22.11.4 00:01:23.404 [Pipeline] } 00:01:23.421 [Pipeline] // retry 00:01:23.426 [Pipeline] } 00:01:23.441 [Pipeline] // withCredentials 00:01:23.450 [Pipeline] httpRequest 00:01:23.865 [Pipeline] echo 00:01:23.867 Sorcerer 10.211.164.101 is alive 00:01:23.879 [Pipeline] retry 00:01:23.881 [Pipeline] { 00:01:23.897 [Pipeline] httpRequest 00:01:23.902 HttpMethod: GET 00:01:23.903 URL: http://10.211.164.101/packages/dpdk_fee0f13c213d0584f0c42a51d0e0625d99a0b2f1.tar.gz 00:01:23.903 Sending request to url: http://10.211.164.101/packages/dpdk_fee0f13c213d0584f0c42a51d0e0625d99a0b2f1.tar.gz 00:01:23.904 Response Code: HTTP/1.1 200 OK 00:01:23.904 Success: Status code 200 is in the accepted range: 200,404 00:01:23.905 Saving response body to /var/jenkins/workspace/raid-vg-autotest/dpdk_fee0f13c213d0584f0c42a51d0e0625d99a0b2f1.tar.gz 00:01:31.914 [Pipeline] } 00:01:31.925 [Pipeline] // retry 00:01:31.930 [Pipeline] sh 00:01:32.213 + tar --no-same-owner -xf dpdk_fee0f13c213d0584f0c42a51d0e0625d99a0b2f1.tar.gz 00:01:33.604 [Pipeline] sh 00:01:33.889 + git -C dpdk log --oneline -n5 00:01:33.889 caf0f5d395 version: 22.11.4 00:01:33.889 7d6f1cc05f Revert "net/iavf: fix abnormal disable HW interrupt" 00:01:33.889 dc9c799c7d vhost: fix missing spinlock unlock 00:01:33.889 4307659a90 net/mlx5: fix LACP redirection in Rx domain 00:01:33.889 6ef77f2a5e net/gve: fix RX buffer size alignment 00:01:33.908 [Pipeline] writeFile 00:01:33.924 [Pipeline] sh 00:01:34.211 + jbp/jenkins/jjb-config/jobs/scripts/autorun_quirks.sh 00:01:34.224 [Pipeline] sh 00:01:34.509 + cat autorun-spdk.conf 00:01:34.509 SPDK_RUN_FUNCTIONAL_TEST=1 00:01:34.509 SPDK_RUN_ASAN=1 00:01:34.509 SPDK_RUN_UBSAN=1 00:01:34.509 SPDK_TEST_RAID=1 00:01:34.509 SPDK_TEST_NATIVE_DPDK=v22.11.4 00:01:34.509 SPDK_RUN_EXTERNAL_DPDK=/home/vagrant/spdk_repo/dpdk/build 00:01:34.509 SPDK_ABI_DIR=/home/vagrant/spdk_repo/spdk-abi 00:01:34.518 RUN_NIGHTLY=1 00:01:34.520 [Pipeline] } 00:01:34.534 [Pipeline] // stage 00:01:34.548 [Pipeline] stage 00:01:34.551 [Pipeline] { (Run VM) 00:01:34.563 [Pipeline] sh 00:01:34.848 + jbp/jenkins/jjb-config/jobs/scripts/prepare_nvme.sh 00:01:34.848 + echo 'Start stage prepare_nvme.sh' 00:01:34.848 Start stage prepare_nvme.sh 00:01:34.848 + [[ -n 6 ]] 00:01:34.848 + disk_prefix=ex6 00:01:34.848 + [[ -n /var/jenkins/workspace/raid-vg-autotest ]] 00:01:34.848 + [[ -e /var/jenkins/workspace/raid-vg-autotest/autorun-spdk.conf ]] 00:01:34.848 + source /var/jenkins/workspace/raid-vg-autotest/autorun-spdk.conf 00:01:34.848 ++ SPDK_RUN_FUNCTIONAL_TEST=1 00:01:34.848 ++ SPDK_RUN_ASAN=1 00:01:34.848 ++ SPDK_RUN_UBSAN=1 00:01:34.848 ++ SPDK_TEST_RAID=1 00:01:34.848 ++ SPDK_TEST_NATIVE_DPDK=v22.11.4 00:01:34.848 ++ SPDK_RUN_EXTERNAL_DPDK=/home/vagrant/spdk_repo/dpdk/build 00:01:34.848 ++ SPDK_ABI_DIR=/home/vagrant/spdk_repo/spdk-abi 00:01:34.848 ++ RUN_NIGHTLY=1 00:01:34.848 + cd /var/jenkins/workspace/raid-vg-autotest 00:01:34.848 + nvme_files=() 00:01:34.848 + declare -A nvme_files 00:01:34.848 + backend_dir=/var/lib/libvirt/images/backends 00:01:34.848 + nvme_files['nvme.img']=5G 00:01:34.848 + nvme_files['nvme-cmb.img']=5G 00:01:34.848 + nvme_files['nvme-multi0.img']=4G 00:01:34.848 + nvme_files['nvme-multi1.img']=4G 00:01:34.848 + nvme_files['nvme-multi2.img']=4G 00:01:34.848 + nvme_files['nvme-openstack.img']=8G 00:01:34.848 + nvme_files['nvme-zns.img']=5G 00:01:34.848 + (( SPDK_TEST_NVME_PMR == 1 )) 00:01:34.848 + (( SPDK_TEST_FTL == 1 )) 00:01:34.848 + (( SPDK_TEST_NVME_FDP == 1 )) 00:01:34.848 + [[ ! -d /var/lib/libvirt/images/backends ]] 00:01:34.848 + for nvme in "${!nvme_files[@]}" 00:01:34.848 + sudo -E spdk/scripts/vagrant/create_nvme_img.sh -n /var/lib/libvirt/images/backends/ex6-nvme-multi2.img -s 4G 00:01:34.848 Formatting '/var/lib/libvirt/images/backends/ex6-nvme-multi2.img', fmt=raw size=4294967296 preallocation=falloc 00:01:34.848 + for nvme in "${!nvme_files[@]}" 00:01:34.848 + sudo -E spdk/scripts/vagrant/create_nvme_img.sh -n /var/lib/libvirt/images/backends/ex6-nvme-cmb.img -s 5G 00:01:34.848 Formatting '/var/lib/libvirt/images/backends/ex6-nvme-cmb.img', fmt=raw size=5368709120 preallocation=falloc 00:01:34.848 + for nvme in "${!nvme_files[@]}" 00:01:34.848 + sudo -E spdk/scripts/vagrant/create_nvme_img.sh -n /var/lib/libvirt/images/backends/ex6-nvme-openstack.img -s 8G 00:01:34.848 Formatting '/var/lib/libvirt/images/backends/ex6-nvme-openstack.img', fmt=raw size=8589934592 preallocation=falloc 00:01:34.848 + for nvme in "${!nvme_files[@]}" 00:01:34.848 + sudo -E spdk/scripts/vagrant/create_nvme_img.sh -n /var/lib/libvirt/images/backends/ex6-nvme-zns.img -s 5G 00:01:34.848 Formatting '/var/lib/libvirt/images/backends/ex6-nvme-zns.img', fmt=raw size=5368709120 preallocation=falloc 00:01:34.848 + for nvme in "${!nvme_files[@]}" 00:01:34.848 + sudo -E spdk/scripts/vagrant/create_nvme_img.sh -n /var/lib/libvirt/images/backends/ex6-nvme-multi1.img -s 4G 00:01:34.848 Formatting '/var/lib/libvirt/images/backends/ex6-nvme-multi1.img', fmt=raw size=4294967296 preallocation=falloc 00:01:34.848 + for nvme in "${!nvme_files[@]}" 00:01:34.848 + sudo -E spdk/scripts/vagrant/create_nvme_img.sh -n /var/lib/libvirt/images/backends/ex6-nvme-multi0.img -s 4G 00:01:34.848 Formatting '/var/lib/libvirt/images/backends/ex6-nvme-multi0.img', fmt=raw size=4294967296 preallocation=falloc 00:01:34.848 + for nvme in "${!nvme_files[@]}" 00:01:34.848 + sudo -E spdk/scripts/vagrant/create_nvme_img.sh -n /var/lib/libvirt/images/backends/ex6-nvme.img -s 5G 00:01:35.108 Formatting '/var/lib/libvirt/images/backends/ex6-nvme.img', fmt=raw size=5368709120 preallocation=falloc 00:01:35.108 ++ sudo grep -rl ex6-nvme.img /etc/libvirt/qemu 00:01:35.108 + echo 'End stage prepare_nvme.sh' 00:01:35.108 End stage prepare_nvme.sh 00:01:35.120 [Pipeline] sh 00:01:35.404 + DISTRO=fedora39 CPUS=10 RAM=12288 jbp/jenkins/jjb-config/jobs/scripts/vagrant_create_vm.sh 00:01:35.404 Setup: -n 10 -s 12288 -x http://proxy-dmz.intel.com:911 -p libvirt --qemu-emulator=/usr/local/qemu/vanilla-v8.0.0/bin/qemu-system-x86_64 -b /var/lib/libvirt/images/backends/ex6-nvme.img -b /var/lib/libvirt/images/backends/ex6-nvme-multi0.img,nvme,/var/lib/libvirt/images/backends/ex6-nvme-multi1.img:/var/lib/libvirt/images/backends/ex6-nvme-multi2.img -H -a -v -f fedora39 00:01:35.404 00:01:35.404 DIR=/var/jenkins/workspace/raid-vg-autotest/spdk/scripts/vagrant 00:01:35.404 SPDK_DIR=/var/jenkins/workspace/raid-vg-autotest/spdk 00:01:35.404 VAGRANT_TARGET=/var/jenkins/workspace/raid-vg-autotest 00:01:35.404 HELP=0 00:01:35.404 DRY_RUN=0 00:01:35.404 NVME_FILE=/var/lib/libvirt/images/backends/ex6-nvme.img,/var/lib/libvirt/images/backends/ex6-nvme-multi0.img, 00:01:35.404 NVME_DISKS_TYPE=nvme,nvme, 00:01:35.404 NVME_AUTO_CREATE=0 00:01:35.404 NVME_DISKS_NAMESPACES=,/var/lib/libvirt/images/backends/ex6-nvme-multi1.img:/var/lib/libvirt/images/backends/ex6-nvme-multi2.img, 00:01:35.404 NVME_CMB=,, 00:01:35.404 NVME_PMR=,, 00:01:35.404 NVME_ZNS=,, 00:01:35.404 NVME_MS=,, 00:01:35.404 NVME_FDP=,, 00:01:35.404 SPDK_VAGRANT_DISTRO=fedora39 00:01:35.404 SPDK_VAGRANT_VMCPU=10 00:01:35.404 SPDK_VAGRANT_VMRAM=12288 00:01:35.404 SPDK_VAGRANT_PROVIDER=libvirt 00:01:35.404 SPDK_VAGRANT_HTTP_PROXY=http://proxy-dmz.intel.com:911 00:01:35.404 SPDK_QEMU_EMULATOR=/usr/local/qemu/vanilla-v8.0.0/bin/qemu-system-x86_64 00:01:35.404 SPDK_OPENSTACK_NETWORK=0 00:01:35.404 VAGRANT_PACKAGE_BOX=0 00:01:35.404 VAGRANTFILE=/var/jenkins/workspace/raid-vg-autotest/spdk/scripts/vagrant/Vagrantfile 00:01:35.404 FORCE_DISTRO=true 00:01:35.404 VAGRANT_BOX_VERSION= 00:01:35.404 EXTRA_VAGRANTFILES= 00:01:35.404 NIC_MODEL=virtio 00:01:35.404 00:01:35.404 mkdir: created directory '/var/jenkins/workspace/raid-vg-autotest/fedora39-libvirt' 00:01:35.404 /var/jenkins/workspace/raid-vg-autotest/fedora39-libvirt /var/jenkins/workspace/raid-vg-autotest 00:01:37.311 Bringing machine 'default' up with 'libvirt' provider... 00:01:37.572 ==> default: Creating image (snapshot of base box volume). 00:01:37.833 ==> default: Creating domain with the following settings... 00:01:37.833 ==> default: -- Name: fedora39-39-1.5-1721788873-2326_default_1732728188_d18d1080aa958676dde4 00:01:37.833 ==> default: -- Domain type: kvm 00:01:37.833 ==> default: -- Cpus: 10 00:01:37.833 ==> default: -- Feature: acpi 00:01:37.833 ==> default: -- Feature: apic 00:01:37.833 ==> default: -- Feature: pae 00:01:37.833 ==> default: -- Memory: 12288M 00:01:37.833 ==> default: -- Memory Backing: hugepages: 00:01:37.833 ==> default: -- Management MAC: 00:01:37.833 ==> default: -- Loader: 00:01:37.833 ==> default: -- Nvram: 00:01:37.833 ==> default: -- Base box: spdk/fedora39 00:01:37.833 ==> default: -- Storage pool: default 00:01:37.833 ==> default: -- Image: /var/lib/libvirt/images/fedora39-39-1.5-1721788873-2326_default_1732728188_d18d1080aa958676dde4.img (20G) 00:01:37.833 ==> default: -- Volume Cache: default 00:01:37.833 ==> default: -- Kernel: 00:01:37.833 ==> default: -- Initrd: 00:01:37.833 ==> default: -- Graphics Type: vnc 00:01:37.833 ==> default: -- Graphics Port: -1 00:01:37.833 ==> default: -- Graphics IP: 127.0.0.1 00:01:37.833 ==> default: -- Graphics Password: Not defined 00:01:37.833 ==> default: -- Video Type: cirrus 00:01:37.833 ==> default: -- Video VRAM: 9216 00:01:37.833 ==> default: -- Sound Type: 00:01:37.833 ==> default: -- Keymap: en-us 00:01:37.833 ==> default: -- TPM Path: 00:01:37.833 ==> default: -- INPUT: type=mouse, bus=ps2 00:01:37.833 ==> default: -- Command line args: 00:01:37.833 ==> default: -> value=-device, 00:01:37.833 ==> default: -> value=nvme,id=nvme-0,serial=12340,addr=0x10, 00:01:37.833 ==> default: -> value=-drive, 00:01:37.833 ==> default: -> value=format=raw,file=/var/lib/libvirt/images/backends/ex6-nvme.img,if=none,id=nvme-0-drive0, 00:01:37.833 ==> default: -> value=-device, 00:01:37.833 ==> default: -> value=nvme-ns,drive=nvme-0-drive0,bus=nvme-0,nsid=1,zoned=false,logical_block_size=4096,physical_block_size=4096, 00:01:37.833 ==> default: -> value=-device, 00:01:37.833 ==> default: -> value=nvme,id=nvme-1,serial=12341,addr=0x11, 00:01:37.833 ==> default: -> value=-drive, 00:01:37.833 ==> default: -> value=format=raw,file=/var/lib/libvirt/images/backends/ex6-nvme-multi0.img,if=none,id=nvme-1-drive0, 00:01:37.833 ==> default: -> value=-device, 00:01:37.833 ==> default: -> value=nvme-ns,drive=nvme-1-drive0,bus=nvme-1,nsid=1,zoned=false,logical_block_size=4096,physical_block_size=4096, 00:01:37.833 ==> default: -> value=-drive, 00:01:37.833 ==> default: -> value=format=raw,file=/var/lib/libvirt/images/backends/ex6-nvme-multi1.img,if=none,id=nvme-1-drive1, 00:01:37.833 ==> default: -> value=-device, 00:01:37.833 ==> default: -> value=nvme-ns,drive=nvme-1-drive1,bus=nvme-1,nsid=2,zoned=false,logical_block_size=4096,physical_block_size=4096, 00:01:37.833 ==> default: -> value=-drive, 00:01:37.833 ==> default: -> value=format=raw,file=/var/lib/libvirt/images/backends/ex6-nvme-multi2.img,if=none,id=nvme-1-drive2, 00:01:37.833 ==> default: -> value=-device, 00:01:37.833 ==> default: -> value=nvme-ns,drive=nvme-1-drive2,bus=nvme-1,nsid=3,zoned=false,logical_block_size=4096,physical_block_size=4096, 00:01:37.833 ==> default: Creating shared folders metadata... 00:01:38.101 ==> default: Starting domain. 00:01:40.017 ==> default: Waiting for domain to get an IP address... 00:01:54.961 ==> default: Waiting for SSH to become available... 00:01:55.902 ==> default: Configuring and enabling network interfaces... 00:02:02.473 default: SSH address: 192.168.121.187:22 00:02:02.473 default: SSH username: vagrant 00:02:02.473 default: SSH auth method: private key 00:02:05.765 ==> default: Rsyncing folder: /mnt/jenkins_nvme/jenkins/workspace/raid-vg-autotest/spdk/ => /home/vagrant/spdk_repo/spdk 00:02:12.359 ==> default: Rsyncing folder: /mnt/jenkins_nvme/jenkins/workspace/raid-vg-autotest/dpdk/ => /home/vagrant/spdk_repo/dpdk 00:02:18.938 ==> default: Mounting SSHFS shared folder... 00:02:20.848 ==> default: Mounting folder via SSHFS: /mnt/jenkins_nvme/jenkins/workspace/raid-vg-autotest/fedora39-libvirt/output => /home/vagrant/spdk_repo/output 00:02:20.848 ==> default: Checking Mount.. 00:02:22.229 ==> default: Folder Successfully Mounted! 00:02:22.229 ==> default: Running provisioner: file... 00:02:23.611 default: ~/.gitconfig => .gitconfig 00:02:24.181 00:02:24.181 SUCCESS! 00:02:24.181 00:02:24.181 cd to /var/jenkins/workspace/raid-vg-autotest/fedora39-libvirt and type "vagrant ssh" to use. 00:02:24.181 Use vagrant "suspend" and vagrant "resume" to stop and start. 00:02:24.181 Use vagrant "destroy" followed by "rm -rf /var/jenkins/workspace/raid-vg-autotest/fedora39-libvirt" to destroy all trace of vm. 00:02:24.181 00:02:24.192 [Pipeline] } 00:02:24.207 [Pipeline] // stage 00:02:24.217 [Pipeline] dir 00:02:24.218 Running in /var/jenkins/workspace/raid-vg-autotest/fedora39-libvirt 00:02:24.220 [Pipeline] { 00:02:24.233 [Pipeline] catchError 00:02:24.235 [Pipeline] { 00:02:24.248 [Pipeline] sh 00:02:24.532 + vagrant ssh-config --host vagrant 00:02:24.532 + sed -ne /^Host/,$p 00:02:24.532 + tee ssh_conf 00:02:27.073 Host vagrant 00:02:27.073 HostName 192.168.121.187 00:02:27.073 User vagrant 00:02:27.073 Port 22 00:02:27.073 UserKnownHostsFile /dev/null 00:02:27.073 StrictHostKeyChecking no 00:02:27.073 PasswordAuthentication no 00:02:27.073 IdentityFile /var/lib/libvirt/images/.vagrant.d/boxes/spdk-VAGRANTSLASH-fedora39/39-1.5-1721788873-2326/libvirt/fedora39 00:02:27.073 IdentitiesOnly yes 00:02:27.073 LogLevel FATAL 00:02:27.073 ForwardAgent yes 00:02:27.073 ForwardX11 yes 00:02:27.073 00:02:27.088 [Pipeline] withEnv 00:02:27.091 [Pipeline] { 00:02:27.108 [Pipeline] sh 00:02:27.397 + /usr/local/bin/ssh -t -F ssh_conf vagrant@vagrant #!/bin/bash 00:02:27.397 source /etc/os-release 00:02:27.397 [[ -e /image.version ]] && img=$(< /image.version) 00:02:27.397 # Minimal, systemd-like check. 00:02:27.397 if [[ -e /.dockerenv ]]; then 00:02:27.397 # Clear garbage from the node's name: 00:02:27.397 # agt-er_autotest_547-896 -> autotest_547-896 00:02:27.397 # $HOSTNAME is the actual container id 00:02:27.397 agent=$HOSTNAME@${DOCKER_SWARM_PLUGIN_JENKINS_AGENT_NAME#*_} 00:02:27.397 if grep -q "/etc/hostname" /proc/self/mountinfo; then 00:02:27.397 # We can assume this is a mount from a host where container is running, 00:02:27.397 # so fetch its hostname to easily identify the target swarm worker. 00:02:27.397 container="$(< /etc/hostname) ($agent)" 00:02:27.397 else 00:02:27.397 # Fallback 00:02:27.397 container=$agent 00:02:27.397 fi 00:02:27.397 fi 00:02:27.397 echo "${NAME} ${VERSION_ID}|$(uname -r)|${img:-N/A}|${container:-N/A}" 00:02:27.397 00:02:27.675 [Pipeline] } 00:02:27.686 [Pipeline] // withEnv 00:02:27.691 [Pipeline] setCustomBuildProperty 00:02:27.700 [Pipeline] stage 00:02:27.702 [Pipeline] { (Tests) 00:02:27.711 [Pipeline] sh 00:02:27.987 + scp -F ssh_conf -r /var/jenkins/workspace/raid-vg-autotest/jbp/jenkins/jjb-config/jobs/scripts/autoruner.sh vagrant@vagrant:./ 00:02:28.263 [Pipeline] sh 00:02:28.568 + scp -F ssh_conf -r /var/jenkins/workspace/raid-vg-autotest/jbp/jenkins/jjb-config/jobs/scripts/pkgdep-autoruner.sh vagrant@vagrant:./ 00:02:28.858 [Pipeline] timeout 00:02:28.858 Timeout set to expire in 1 hr 30 min 00:02:28.860 [Pipeline] { 00:02:28.874 [Pipeline] sh 00:02:29.158 + /usr/local/bin/ssh -t -F ssh_conf vagrant@vagrant git -C spdk_repo/spdk reset --hard 00:02:29.729 HEAD is now at b18e1bd62 version: v24.09.1-pre 00:02:29.743 [Pipeline] sh 00:02:30.033 + /usr/local/bin/ssh -t -F ssh_conf vagrant@vagrant sudo chown vagrant:vagrant spdk_repo 00:02:30.308 [Pipeline] sh 00:02:30.598 + scp -F ssh_conf -r /var/jenkins/workspace/raid-vg-autotest/autorun-spdk.conf vagrant@vagrant:spdk_repo 00:02:30.875 [Pipeline] sh 00:02:31.160 + /usr/local/bin/ssh -t -F ssh_conf vagrant@vagrant JOB_BASE_NAME=raid-vg-autotest ./autoruner.sh spdk_repo 00:02:31.421 ++ readlink -f spdk_repo 00:02:31.421 + DIR_ROOT=/home/vagrant/spdk_repo 00:02:31.421 + [[ -n /home/vagrant/spdk_repo ]] 00:02:31.421 + DIR_SPDK=/home/vagrant/spdk_repo/spdk 00:02:31.421 + DIR_OUTPUT=/home/vagrant/spdk_repo/output 00:02:31.421 + [[ -d /home/vagrant/spdk_repo/spdk ]] 00:02:31.421 + [[ ! -d /home/vagrant/spdk_repo/output ]] 00:02:31.421 + [[ -d /home/vagrant/spdk_repo/output ]] 00:02:31.421 + [[ raid-vg-autotest == pkgdep-* ]] 00:02:31.421 + cd /home/vagrant/spdk_repo 00:02:31.421 + source /etc/os-release 00:02:31.421 ++ NAME='Fedora Linux' 00:02:31.421 ++ VERSION='39 (Cloud Edition)' 00:02:31.421 ++ ID=fedora 00:02:31.421 ++ VERSION_ID=39 00:02:31.421 ++ VERSION_CODENAME= 00:02:31.421 ++ PLATFORM_ID=platform:f39 00:02:31.421 ++ PRETTY_NAME='Fedora Linux 39 (Cloud Edition)' 00:02:31.421 ++ ANSI_COLOR='0;38;2;60;110;180' 00:02:31.421 ++ LOGO=fedora-logo-icon 00:02:31.421 ++ CPE_NAME=cpe:/o:fedoraproject:fedora:39 00:02:31.421 ++ HOME_URL=https://fedoraproject.org/ 00:02:31.421 ++ DOCUMENTATION_URL=https://docs.fedoraproject.org/en-US/fedora/f39/system-administrators-guide/ 00:02:31.421 ++ SUPPORT_URL=https://ask.fedoraproject.org/ 00:02:31.421 ++ BUG_REPORT_URL=https://bugzilla.redhat.com/ 00:02:31.421 ++ REDHAT_BUGZILLA_PRODUCT=Fedora 00:02:31.421 ++ REDHAT_BUGZILLA_PRODUCT_VERSION=39 00:02:31.421 ++ REDHAT_SUPPORT_PRODUCT=Fedora 00:02:31.421 ++ REDHAT_SUPPORT_PRODUCT_VERSION=39 00:02:31.421 ++ SUPPORT_END=2024-11-12 00:02:31.421 ++ VARIANT='Cloud Edition' 00:02:31.421 ++ VARIANT_ID=cloud 00:02:31.421 + uname -a 00:02:31.421 Linux fedora39-cloud-1721788873-2326 6.8.9-200.fc39.x86_64 #1 SMP PREEMPT_DYNAMIC Wed Jul 24 03:04:40 UTC 2024 x86_64 GNU/Linux 00:02:31.421 + sudo /home/vagrant/spdk_repo/spdk/scripts/setup.sh status 00:02:31.992 0000:00:03.0 (1af4 1001): Active devices: mount@vda:vda2,mount@vda:vda3,mount@vda:vda5, so not binding PCI dev 00:02:31.992 Hugepages 00:02:31.992 node hugesize free / total 00:02:31.992 node0 1048576kB 0 / 0 00:02:31.992 node0 2048kB 0 / 0 00:02:31.992 00:02:31.992 Type BDF Vendor Device NUMA Driver Device Block devices 00:02:31.992 virtio 0000:00:03.0 1af4 1001 unknown virtio-pci - vda 00:02:31.992 NVMe 0000:00:10.0 1b36 0010 unknown nvme nvme0 nvme0n1 00:02:32.252 NVMe 0000:00:11.0 1b36 0010 unknown nvme nvme1 nvme1n1 nvme1n2 nvme1n3 00:02:32.252 + rm -f /tmp/spdk-ld-path 00:02:32.252 + source autorun-spdk.conf 00:02:32.252 ++ SPDK_RUN_FUNCTIONAL_TEST=1 00:02:32.252 ++ SPDK_RUN_ASAN=1 00:02:32.252 ++ SPDK_RUN_UBSAN=1 00:02:32.252 ++ SPDK_TEST_RAID=1 00:02:32.252 ++ SPDK_TEST_NATIVE_DPDK=v22.11.4 00:02:32.252 ++ SPDK_RUN_EXTERNAL_DPDK=/home/vagrant/spdk_repo/dpdk/build 00:02:32.252 ++ SPDK_ABI_DIR=/home/vagrant/spdk_repo/spdk-abi 00:02:32.252 ++ RUN_NIGHTLY=1 00:02:32.252 + (( SPDK_TEST_NVME_CMB == 1 || SPDK_TEST_NVME_PMR == 1 )) 00:02:32.252 + [[ -n '' ]] 00:02:32.252 + sudo git config --global --add safe.directory /home/vagrant/spdk_repo/spdk 00:02:32.252 + for M in /var/spdk/build-*-manifest.txt 00:02:32.252 + [[ -f /var/spdk/build-kernel-manifest.txt ]] 00:02:32.252 + cp /var/spdk/build-kernel-manifest.txt /home/vagrant/spdk_repo/output/ 00:02:32.252 + for M in /var/spdk/build-*-manifest.txt 00:02:32.252 + [[ -f /var/spdk/build-pkg-manifest.txt ]] 00:02:32.252 + cp /var/spdk/build-pkg-manifest.txt /home/vagrant/spdk_repo/output/ 00:02:32.252 + for M in /var/spdk/build-*-manifest.txt 00:02:32.252 + [[ -f /var/spdk/build-repo-manifest.txt ]] 00:02:32.252 + cp /var/spdk/build-repo-manifest.txt /home/vagrant/spdk_repo/output/ 00:02:32.252 ++ uname 00:02:32.252 + [[ Linux == \L\i\n\u\x ]] 00:02:32.252 + sudo dmesg -T 00:02:32.252 + sudo dmesg --clear 00:02:32.252 + dmesg_pid=6165 00:02:32.252 + [[ Fedora Linux == FreeBSD ]] 00:02:32.252 + export UNBIND_ENTIRE_IOMMU_GROUP=yes 00:02:32.252 + UNBIND_ENTIRE_IOMMU_GROUP=yes 00:02:32.252 + sudo dmesg -Tw 00:02:32.252 + [[ -e /var/spdk/dependencies/vhost/spdk_test_image.qcow2 ]] 00:02:32.252 + [[ -x /usr/src/fio-static/fio ]] 00:02:32.252 + export FIO_BIN=/usr/src/fio-static/fio 00:02:32.252 + FIO_BIN=/usr/src/fio-static/fio 00:02:32.252 + [[ '' == \/\q\e\m\u\_\v\f\i\o\/* ]] 00:02:32.252 + [[ ! -v VFIO_QEMU_BIN ]] 00:02:32.252 + [[ -e /usr/local/qemu/vfio-user-latest ]] 00:02:32.252 + export VFIO_QEMU_BIN=/usr/local/qemu/vfio-user-latest/bin/qemu-system-x86_64 00:02:32.252 + VFIO_QEMU_BIN=/usr/local/qemu/vfio-user-latest/bin/qemu-system-x86_64 00:02:32.252 + [[ -e /usr/local/qemu/vanilla-latest ]] 00:02:32.252 + export QEMU_BIN=/usr/local/qemu/vanilla-latest/bin/qemu-system-x86_64 00:02:32.252 + QEMU_BIN=/usr/local/qemu/vanilla-latest/bin/qemu-system-x86_64 00:02:32.252 + spdk/autorun.sh /home/vagrant/spdk_repo/autorun-spdk.conf 00:02:32.252 Test configuration: 00:02:32.252 SPDK_RUN_FUNCTIONAL_TEST=1 00:02:32.252 SPDK_RUN_ASAN=1 00:02:32.252 SPDK_RUN_UBSAN=1 00:02:32.252 SPDK_TEST_RAID=1 00:02:32.252 SPDK_TEST_NATIVE_DPDK=v22.11.4 00:02:32.252 SPDK_RUN_EXTERNAL_DPDK=/home/vagrant/spdk_repo/dpdk/build 00:02:32.252 SPDK_ABI_DIR=/home/vagrant/spdk_repo/spdk-abi 00:02:32.513 RUN_NIGHTLY=1 17:24:03 -- common/autotest_common.sh@1680 -- $ [[ n == y ]] 00:02:32.513 17:24:03 -- common/autobuild_common.sh@15 -- $ source /home/vagrant/spdk_repo/spdk/scripts/common.sh 00:02:32.513 17:24:03 -- scripts/common.sh@15 -- $ shopt -s extglob 00:02:32.513 17:24:03 -- scripts/common.sh@544 -- $ [[ -e /bin/wpdk_common.sh ]] 00:02:32.513 17:24:03 -- scripts/common.sh@552 -- $ [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:02:32.513 17:24:03 -- scripts/common.sh@553 -- $ source /etc/opt/spdk-pkgdep/paths/export.sh 00:02:32.513 17:24:03 -- paths/export.sh@2 -- $ PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/vagrant/.local/bin:/home/vagrant/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:02:32.513 17:24:03 -- paths/export.sh@3 -- $ PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/vagrant/.local/bin:/home/vagrant/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:02:32.513 17:24:03 -- paths/export.sh@4 -- $ PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/vagrant/.local/bin:/home/vagrant/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:02:32.513 17:24:03 -- paths/export.sh@5 -- $ export PATH 00:02:32.513 17:24:03 -- paths/export.sh@6 -- $ echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/vagrant/.local/bin:/home/vagrant/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:02:32.513 17:24:03 -- common/autobuild_common.sh@478 -- $ out=/home/vagrant/spdk_repo/spdk/../output 00:02:32.513 17:24:03 -- common/autobuild_common.sh@479 -- $ date +%s 00:02:32.513 17:24:03 -- common/autobuild_common.sh@479 -- $ mktemp -dt spdk_1732728243.XXXXXX 00:02:32.513 17:24:03 -- common/autobuild_common.sh@479 -- $ SPDK_WORKSPACE=/tmp/spdk_1732728243.15dnDe 00:02:32.513 17:24:03 -- common/autobuild_common.sh@481 -- $ [[ -n '' ]] 00:02:32.513 17:24:03 -- common/autobuild_common.sh@485 -- $ '[' -n v22.11.4 ']' 00:02:32.513 17:24:03 -- common/autobuild_common.sh@486 -- $ dirname /home/vagrant/spdk_repo/dpdk/build 00:02:32.513 17:24:03 -- common/autobuild_common.sh@486 -- $ scanbuild_exclude=' --exclude /home/vagrant/spdk_repo/dpdk' 00:02:32.513 17:24:03 -- common/autobuild_common.sh@492 -- $ scanbuild_exclude+=' --exclude /home/vagrant/spdk_repo/spdk/xnvme --exclude /tmp' 00:02:32.513 17:24:03 -- common/autobuild_common.sh@494 -- $ scanbuild='scan-build -o /home/vagrant/spdk_repo/spdk/../output/scan-build-tmp --exclude /home/vagrant/spdk_repo/dpdk --exclude /home/vagrant/spdk_repo/spdk/xnvme --exclude /tmp --status-bugs' 00:02:32.513 17:24:03 -- common/autobuild_common.sh@495 -- $ get_config_params 00:02:32.513 17:24:03 -- common/autotest_common.sh@407 -- $ xtrace_disable 00:02:32.513 17:24:03 -- common/autotest_common.sh@10 -- $ set +x 00:02:32.513 17:24:03 -- common/autobuild_common.sh@495 -- $ config_params='--enable-debug --enable-werror --with-rdma --with-idxd --with-fio=/usr/src/fio --with-iscsi-initiator --disable-unit-tests --enable-ubsan --enable-asan --enable-coverage --with-ublk --with-raid5f --with-dpdk=/home/vagrant/spdk_repo/dpdk/build' 00:02:32.513 17:24:03 -- common/autobuild_common.sh@497 -- $ start_monitor_resources 00:02:32.513 17:24:03 -- pm/common@17 -- $ local monitor 00:02:32.513 17:24:03 -- pm/common@19 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:02:32.513 17:24:03 -- pm/common@19 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:02:32.513 17:24:03 -- pm/common@25 -- $ sleep 1 00:02:32.513 17:24:03 -- pm/common@21 -- $ date +%s 00:02:32.513 17:24:03 -- pm/common@21 -- $ date +%s 00:02:32.513 17:24:03 -- pm/common@21 -- $ /home/vagrant/spdk_repo/spdk/scripts/perf/pm/collect-cpu-load -d /home/vagrant/spdk_repo/spdk/../output/power -l -p monitor.autobuild.sh.1732728243 00:02:32.513 17:24:03 -- pm/common@21 -- $ /home/vagrant/spdk_repo/spdk/scripts/perf/pm/collect-vmstat -d /home/vagrant/spdk_repo/spdk/../output/power -l -p monitor.autobuild.sh.1732728243 00:02:32.513 Redirecting to /home/vagrant/spdk_repo/spdk/../output/power/monitor.autobuild.sh.1732728243_collect-cpu-load.pm.log 00:02:32.513 Redirecting to /home/vagrant/spdk_repo/spdk/../output/power/monitor.autobuild.sh.1732728243_collect-vmstat.pm.log 00:02:33.455 17:24:04 -- common/autobuild_common.sh@498 -- $ trap stop_monitor_resources EXIT 00:02:33.455 17:24:04 -- spdk/autobuild.sh@11 -- $ SPDK_TEST_AUTOBUILD= 00:02:33.455 17:24:04 -- spdk/autobuild.sh@12 -- $ umask 022 00:02:33.455 17:24:04 -- spdk/autobuild.sh@13 -- $ cd /home/vagrant/spdk_repo/spdk 00:02:33.455 17:24:04 -- spdk/autobuild.sh@16 -- $ date -u 00:02:33.455 Wed Nov 27 05:24:04 PM UTC 2024 00:02:33.455 17:24:04 -- spdk/autobuild.sh@17 -- $ git describe --tags 00:02:33.455 v24.09-rc1-9-gb18e1bd62 00:02:33.455 17:24:04 -- spdk/autobuild.sh@19 -- $ '[' 1 -eq 1 ']' 00:02:33.455 17:24:04 -- spdk/autobuild.sh@20 -- $ run_test asan echo 'using asan' 00:02:33.455 17:24:04 -- common/autotest_common.sh@1101 -- $ '[' 3 -le 1 ']' 00:02:33.455 17:24:04 -- common/autotest_common.sh@1107 -- $ xtrace_disable 00:02:33.455 17:24:04 -- common/autotest_common.sh@10 -- $ set +x 00:02:33.455 ************************************ 00:02:33.455 START TEST asan 00:02:33.455 ************************************ 00:02:33.455 using asan 00:02:33.455 17:24:04 asan -- common/autotest_common.sh@1125 -- $ echo 'using asan' 00:02:33.455 00:02:33.455 real 0m0.000s 00:02:33.455 user 0m0.000s 00:02:33.455 sys 0m0.000s 00:02:33.455 17:24:04 asan -- common/autotest_common.sh@1126 -- $ xtrace_disable 00:02:33.455 17:24:04 asan -- common/autotest_common.sh@10 -- $ set +x 00:02:33.455 ************************************ 00:02:33.455 END TEST asan 00:02:33.455 ************************************ 00:02:33.716 17:24:04 -- spdk/autobuild.sh@23 -- $ '[' 1 -eq 1 ']' 00:02:33.716 17:24:04 -- spdk/autobuild.sh@24 -- $ run_test ubsan echo 'using ubsan' 00:02:33.716 17:24:04 -- common/autotest_common.sh@1101 -- $ '[' 3 -le 1 ']' 00:02:33.717 17:24:04 -- common/autotest_common.sh@1107 -- $ xtrace_disable 00:02:33.717 17:24:04 -- common/autotest_common.sh@10 -- $ set +x 00:02:33.717 ************************************ 00:02:33.717 START TEST ubsan 00:02:33.717 ************************************ 00:02:33.717 using ubsan 00:02:33.717 17:24:04 ubsan -- common/autotest_common.sh@1125 -- $ echo 'using ubsan' 00:02:33.717 00:02:33.717 real 0m0.000s 00:02:33.717 user 0m0.000s 00:02:33.717 sys 0m0.000s 00:02:33.717 17:24:04 ubsan -- common/autotest_common.sh@1126 -- $ xtrace_disable 00:02:33.717 17:24:04 ubsan -- common/autotest_common.sh@10 -- $ set +x 00:02:33.717 ************************************ 00:02:33.717 END TEST ubsan 00:02:33.717 ************************************ 00:02:33.717 17:24:04 -- spdk/autobuild.sh@27 -- $ '[' -n v22.11.4 ']' 00:02:33.717 17:24:04 -- spdk/autobuild.sh@28 -- $ build_native_dpdk 00:02:33.717 17:24:04 -- common/autobuild_common.sh@442 -- $ run_test build_native_dpdk _build_native_dpdk 00:02:33.717 17:24:04 -- common/autotest_common.sh@1101 -- $ '[' 2 -le 1 ']' 00:02:33.717 17:24:04 -- common/autotest_common.sh@1107 -- $ xtrace_disable 00:02:33.717 17:24:04 -- common/autotest_common.sh@10 -- $ set +x 00:02:33.717 ************************************ 00:02:33.717 START TEST build_native_dpdk 00:02:33.717 ************************************ 00:02:33.717 17:24:04 build_native_dpdk -- common/autotest_common.sh@1125 -- $ _build_native_dpdk 00:02:33.717 17:24:04 build_native_dpdk -- common/autobuild_common.sh@48 -- $ local external_dpdk_dir 00:02:33.717 17:24:04 build_native_dpdk -- common/autobuild_common.sh@49 -- $ local external_dpdk_base_dir 00:02:33.717 17:24:04 build_native_dpdk -- common/autobuild_common.sh@50 -- $ local compiler_version 00:02:33.717 17:24:04 build_native_dpdk -- common/autobuild_common.sh@51 -- $ local compiler 00:02:33.717 17:24:04 build_native_dpdk -- common/autobuild_common.sh@52 -- $ local dpdk_kmods 00:02:33.717 17:24:04 build_native_dpdk -- common/autobuild_common.sh@53 -- $ local repo=dpdk 00:02:33.717 17:24:04 build_native_dpdk -- common/autobuild_common.sh@55 -- $ compiler=gcc 00:02:33.717 17:24:04 build_native_dpdk -- common/autobuild_common.sh@61 -- $ export CC=gcc 00:02:33.717 17:24:04 build_native_dpdk -- common/autobuild_common.sh@61 -- $ CC=gcc 00:02:33.717 17:24:04 build_native_dpdk -- common/autobuild_common.sh@63 -- $ [[ gcc != *clang* ]] 00:02:33.717 17:24:04 build_native_dpdk -- common/autobuild_common.sh@63 -- $ [[ gcc != *gcc* ]] 00:02:33.717 17:24:04 build_native_dpdk -- common/autobuild_common.sh@68 -- $ gcc -dumpversion 00:02:33.717 17:24:04 build_native_dpdk -- common/autobuild_common.sh@68 -- $ compiler_version=13 00:02:33.717 17:24:04 build_native_dpdk -- common/autobuild_common.sh@69 -- $ compiler_version=13 00:02:33.717 17:24:04 build_native_dpdk -- common/autobuild_common.sh@70 -- $ external_dpdk_dir=/home/vagrant/spdk_repo/dpdk/build 00:02:33.717 17:24:04 build_native_dpdk -- common/autobuild_common.sh@71 -- $ dirname /home/vagrant/spdk_repo/dpdk/build 00:02:33.717 17:24:04 build_native_dpdk -- common/autobuild_common.sh@71 -- $ external_dpdk_base_dir=/home/vagrant/spdk_repo/dpdk 00:02:33.717 17:24:04 build_native_dpdk -- common/autobuild_common.sh@73 -- $ [[ ! -d /home/vagrant/spdk_repo/dpdk ]] 00:02:33.717 17:24:04 build_native_dpdk -- common/autobuild_common.sh@82 -- $ orgdir=/home/vagrant/spdk_repo/spdk 00:02:33.717 17:24:04 build_native_dpdk -- common/autobuild_common.sh@83 -- $ git -C /home/vagrant/spdk_repo/dpdk log --oneline -n 5 00:02:33.717 caf0f5d395 version: 22.11.4 00:02:33.717 7d6f1cc05f Revert "net/iavf: fix abnormal disable HW interrupt" 00:02:33.717 dc9c799c7d vhost: fix missing spinlock unlock 00:02:33.717 4307659a90 net/mlx5: fix LACP redirection in Rx domain 00:02:33.717 6ef77f2a5e net/gve: fix RX buffer size alignment 00:02:33.717 17:24:04 build_native_dpdk -- common/autobuild_common.sh@85 -- $ dpdk_cflags='-fPIC -g -fcommon' 00:02:33.717 17:24:04 build_native_dpdk -- common/autobuild_common.sh@86 -- $ dpdk_ldflags= 00:02:33.717 17:24:04 build_native_dpdk -- common/autobuild_common.sh@87 -- $ dpdk_ver=22.11.4 00:02:33.717 17:24:04 build_native_dpdk -- common/autobuild_common.sh@89 -- $ [[ gcc == *gcc* ]] 00:02:33.717 17:24:04 build_native_dpdk -- common/autobuild_common.sh@89 -- $ [[ 13 -ge 5 ]] 00:02:33.717 17:24:04 build_native_dpdk -- common/autobuild_common.sh@90 -- $ dpdk_cflags+=' -Werror' 00:02:33.717 17:24:04 build_native_dpdk -- common/autobuild_common.sh@93 -- $ [[ gcc == *gcc* ]] 00:02:33.717 17:24:04 build_native_dpdk -- common/autobuild_common.sh@93 -- $ [[ 13 -ge 10 ]] 00:02:33.717 17:24:04 build_native_dpdk -- common/autobuild_common.sh@94 -- $ dpdk_cflags+=' -Wno-stringop-overflow' 00:02:33.717 17:24:04 build_native_dpdk -- common/autobuild_common.sh@100 -- $ DPDK_DRIVERS=("bus" "bus/pci" "bus/vdev" "mempool/ring" "net/i40e" "net/i40e/base") 00:02:33.717 17:24:04 build_native_dpdk -- common/autobuild_common.sh@102 -- $ local mlx5_libs_added=n 00:02:33.717 17:24:04 build_native_dpdk -- common/autobuild_common.sh@103 -- $ [[ 0 -eq 1 ]] 00:02:33.717 17:24:04 build_native_dpdk -- common/autobuild_common.sh@103 -- $ [[ 0 -eq 1 ]] 00:02:33.717 17:24:04 build_native_dpdk -- common/autobuild_common.sh@139 -- $ [[ 0 -eq 1 ]] 00:02:33.717 17:24:04 build_native_dpdk -- common/autobuild_common.sh@167 -- $ cd /home/vagrant/spdk_repo/dpdk 00:02:33.717 17:24:04 build_native_dpdk -- common/autobuild_common.sh@168 -- $ uname -s 00:02:33.717 17:24:04 build_native_dpdk -- common/autobuild_common.sh@168 -- $ '[' Linux = Linux ']' 00:02:33.717 17:24:04 build_native_dpdk -- common/autobuild_common.sh@169 -- $ lt 22.11.4 21.11.0 00:02:33.717 17:24:04 build_native_dpdk -- scripts/common.sh@373 -- $ cmp_versions 22.11.4 '<' 21.11.0 00:02:33.717 17:24:04 build_native_dpdk -- scripts/common.sh@333 -- $ local ver1 ver1_l 00:02:33.717 17:24:04 build_native_dpdk -- scripts/common.sh@334 -- $ local ver2 ver2_l 00:02:33.717 17:24:04 build_native_dpdk -- scripts/common.sh@336 -- $ IFS=.-: 00:02:33.717 17:24:04 build_native_dpdk -- scripts/common.sh@336 -- $ read -ra ver1 00:02:33.717 17:24:04 build_native_dpdk -- scripts/common.sh@337 -- $ IFS=.-: 00:02:33.717 17:24:04 build_native_dpdk -- scripts/common.sh@337 -- $ read -ra ver2 00:02:33.717 17:24:04 build_native_dpdk -- scripts/common.sh@338 -- $ local 'op=<' 00:02:33.717 17:24:04 build_native_dpdk -- scripts/common.sh@340 -- $ ver1_l=3 00:02:33.717 17:24:04 build_native_dpdk -- scripts/common.sh@341 -- $ ver2_l=3 00:02:33.717 17:24:04 build_native_dpdk -- scripts/common.sh@343 -- $ local lt=0 gt=0 eq=0 v 00:02:33.717 17:24:04 build_native_dpdk -- scripts/common.sh@344 -- $ case "$op" in 00:02:33.717 17:24:04 build_native_dpdk -- scripts/common.sh@345 -- $ : 1 00:02:33.717 17:24:04 build_native_dpdk -- scripts/common.sh@364 -- $ (( v = 0 )) 00:02:33.717 17:24:04 build_native_dpdk -- scripts/common.sh@364 -- $ (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:02:33.717 17:24:04 build_native_dpdk -- scripts/common.sh@365 -- $ decimal 22 00:02:33.717 17:24:04 build_native_dpdk -- scripts/common.sh@353 -- $ local d=22 00:02:33.717 17:24:04 build_native_dpdk -- scripts/common.sh@354 -- $ [[ 22 =~ ^[0-9]+$ ]] 00:02:33.717 17:24:04 build_native_dpdk -- scripts/common.sh@355 -- $ echo 22 00:02:33.717 17:24:04 build_native_dpdk -- scripts/common.sh@365 -- $ ver1[v]=22 00:02:33.717 17:24:04 build_native_dpdk -- scripts/common.sh@366 -- $ decimal 21 00:02:33.717 17:24:04 build_native_dpdk -- scripts/common.sh@353 -- $ local d=21 00:02:33.717 17:24:04 build_native_dpdk -- scripts/common.sh@354 -- $ [[ 21 =~ ^[0-9]+$ ]] 00:02:33.717 17:24:04 build_native_dpdk -- scripts/common.sh@355 -- $ echo 21 00:02:33.717 17:24:04 build_native_dpdk -- scripts/common.sh@366 -- $ ver2[v]=21 00:02:33.717 17:24:04 build_native_dpdk -- scripts/common.sh@367 -- $ (( ver1[v] > ver2[v] )) 00:02:33.717 17:24:04 build_native_dpdk -- scripts/common.sh@367 -- $ return 1 00:02:33.717 17:24:04 build_native_dpdk -- common/autobuild_common.sh@173 -- $ patch -p1 00:02:33.717 patching file config/rte_config.h 00:02:33.717 Hunk #1 succeeded at 60 (offset 1 line). 00:02:33.717 17:24:04 build_native_dpdk -- common/autobuild_common.sh@176 -- $ lt 22.11.4 24.07.0 00:02:33.717 17:24:04 build_native_dpdk -- scripts/common.sh@373 -- $ cmp_versions 22.11.4 '<' 24.07.0 00:02:33.717 17:24:04 build_native_dpdk -- scripts/common.sh@333 -- $ local ver1 ver1_l 00:02:33.717 17:24:04 build_native_dpdk -- scripts/common.sh@334 -- $ local ver2 ver2_l 00:02:33.717 17:24:04 build_native_dpdk -- scripts/common.sh@336 -- $ IFS=.-: 00:02:33.717 17:24:04 build_native_dpdk -- scripts/common.sh@336 -- $ read -ra ver1 00:02:33.717 17:24:04 build_native_dpdk -- scripts/common.sh@337 -- $ IFS=.-: 00:02:33.717 17:24:04 build_native_dpdk -- scripts/common.sh@337 -- $ read -ra ver2 00:02:33.717 17:24:04 build_native_dpdk -- scripts/common.sh@338 -- $ local 'op=<' 00:02:33.717 17:24:04 build_native_dpdk -- scripts/common.sh@340 -- $ ver1_l=3 00:02:33.717 17:24:04 build_native_dpdk -- scripts/common.sh@341 -- $ ver2_l=3 00:02:33.717 17:24:04 build_native_dpdk -- scripts/common.sh@343 -- $ local lt=0 gt=0 eq=0 v 00:02:33.717 17:24:04 build_native_dpdk -- scripts/common.sh@344 -- $ case "$op" in 00:02:33.717 17:24:04 build_native_dpdk -- scripts/common.sh@345 -- $ : 1 00:02:33.717 17:24:04 build_native_dpdk -- scripts/common.sh@364 -- $ (( v = 0 )) 00:02:33.717 17:24:04 build_native_dpdk -- scripts/common.sh@364 -- $ (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:02:33.717 17:24:04 build_native_dpdk -- scripts/common.sh@365 -- $ decimal 22 00:02:33.717 17:24:04 build_native_dpdk -- scripts/common.sh@353 -- $ local d=22 00:02:33.717 17:24:04 build_native_dpdk -- scripts/common.sh@354 -- $ [[ 22 =~ ^[0-9]+$ ]] 00:02:33.717 17:24:04 build_native_dpdk -- scripts/common.sh@355 -- $ echo 22 00:02:33.717 17:24:04 build_native_dpdk -- scripts/common.sh@365 -- $ ver1[v]=22 00:02:33.717 17:24:04 build_native_dpdk -- scripts/common.sh@366 -- $ decimal 24 00:02:33.717 17:24:04 build_native_dpdk -- scripts/common.sh@353 -- $ local d=24 00:02:33.717 17:24:04 build_native_dpdk -- scripts/common.sh@354 -- $ [[ 24 =~ ^[0-9]+$ ]] 00:02:33.717 17:24:04 build_native_dpdk -- scripts/common.sh@355 -- $ echo 24 00:02:33.717 17:24:04 build_native_dpdk -- scripts/common.sh@366 -- $ ver2[v]=24 00:02:33.717 17:24:04 build_native_dpdk -- scripts/common.sh@367 -- $ (( ver1[v] > ver2[v] )) 00:02:33.717 17:24:04 build_native_dpdk -- scripts/common.sh@368 -- $ (( ver1[v] < ver2[v] )) 00:02:33.717 17:24:04 build_native_dpdk -- scripts/common.sh@368 -- $ return 0 00:02:33.717 17:24:04 build_native_dpdk -- common/autobuild_common.sh@177 -- $ patch -p1 00:02:33.717 patching file lib/pcapng/rte_pcapng.c 00:02:33.717 Hunk #1 succeeded at 110 (offset -18 lines). 00:02:33.717 17:24:04 build_native_dpdk -- common/autobuild_common.sh@179 -- $ ge 22.11.4 24.07.0 00:02:33.717 17:24:04 build_native_dpdk -- scripts/common.sh@376 -- $ cmp_versions 22.11.4 '>=' 24.07.0 00:02:33.717 17:24:04 build_native_dpdk -- scripts/common.sh@333 -- $ local ver1 ver1_l 00:02:33.717 17:24:04 build_native_dpdk -- scripts/common.sh@334 -- $ local ver2 ver2_l 00:02:33.717 17:24:04 build_native_dpdk -- scripts/common.sh@336 -- $ IFS=.-: 00:02:33.717 17:24:04 build_native_dpdk -- scripts/common.sh@336 -- $ read -ra ver1 00:02:33.717 17:24:04 build_native_dpdk -- scripts/common.sh@337 -- $ IFS=.-: 00:02:33.717 17:24:04 build_native_dpdk -- scripts/common.sh@337 -- $ read -ra ver2 00:02:33.717 17:24:04 build_native_dpdk -- scripts/common.sh@338 -- $ local 'op=>=' 00:02:33.717 17:24:04 build_native_dpdk -- scripts/common.sh@340 -- $ ver1_l=3 00:02:33.717 17:24:04 build_native_dpdk -- scripts/common.sh@341 -- $ ver2_l=3 00:02:33.717 17:24:04 build_native_dpdk -- scripts/common.sh@343 -- $ local lt=0 gt=0 eq=0 v 00:02:33.717 17:24:04 build_native_dpdk -- scripts/common.sh@344 -- $ case "$op" in 00:02:33.718 17:24:04 build_native_dpdk -- scripts/common.sh@348 -- $ : 1 00:02:33.718 17:24:04 build_native_dpdk -- scripts/common.sh@364 -- $ (( v = 0 )) 00:02:33.718 17:24:04 build_native_dpdk -- scripts/common.sh@364 -- $ (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:02:33.718 17:24:04 build_native_dpdk -- scripts/common.sh@365 -- $ decimal 22 00:02:33.718 17:24:04 build_native_dpdk -- scripts/common.sh@353 -- $ local d=22 00:02:33.718 17:24:04 build_native_dpdk -- scripts/common.sh@354 -- $ [[ 22 =~ ^[0-9]+$ ]] 00:02:33.718 17:24:04 build_native_dpdk -- scripts/common.sh@355 -- $ echo 22 00:02:33.718 17:24:04 build_native_dpdk -- scripts/common.sh@365 -- $ ver1[v]=22 00:02:33.718 17:24:04 build_native_dpdk -- scripts/common.sh@366 -- $ decimal 24 00:02:33.718 17:24:04 build_native_dpdk -- scripts/common.sh@353 -- $ local d=24 00:02:33.718 17:24:04 build_native_dpdk -- scripts/common.sh@354 -- $ [[ 24 =~ ^[0-9]+$ ]] 00:02:33.718 17:24:04 build_native_dpdk -- scripts/common.sh@355 -- $ echo 24 00:02:33.718 17:24:04 build_native_dpdk -- scripts/common.sh@366 -- $ ver2[v]=24 00:02:33.718 17:24:04 build_native_dpdk -- scripts/common.sh@367 -- $ (( ver1[v] > ver2[v] )) 00:02:33.718 17:24:04 build_native_dpdk -- scripts/common.sh@368 -- $ (( ver1[v] < ver2[v] )) 00:02:33.718 17:24:04 build_native_dpdk -- scripts/common.sh@368 -- $ return 1 00:02:33.718 17:24:04 build_native_dpdk -- common/autobuild_common.sh@183 -- $ dpdk_kmods=false 00:02:33.718 17:24:04 build_native_dpdk -- common/autobuild_common.sh@184 -- $ uname -s 00:02:33.718 17:24:04 build_native_dpdk -- common/autobuild_common.sh@184 -- $ '[' Linux = FreeBSD ']' 00:02:33.718 17:24:04 build_native_dpdk -- common/autobuild_common.sh@188 -- $ printf %s, bus bus/pci bus/vdev mempool/ring net/i40e net/i40e/base 00:02:33.718 17:24:04 build_native_dpdk -- common/autobuild_common.sh@188 -- $ meson build-tmp --prefix=/home/vagrant/spdk_repo/dpdk/build --libdir lib -Denable_docs=false -Denable_kmods=false -Dtests=false -Dc_link_args= '-Dc_args=-fPIC -g -fcommon -Werror -Wno-stringop-overflow' -Dmachine=native -Denable_drivers=bus,bus/pci,bus/vdev,mempool/ring,net/i40e,net/i40e/base, 00:02:40.295 The Meson build system 00:02:40.295 Version: 1.5.0 00:02:40.295 Source dir: /home/vagrant/spdk_repo/dpdk 00:02:40.295 Build dir: /home/vagrant/spdk_repo/dpdk/build-tmp 00:02:40.295 Build type: native build 00:02:40.295 Program cat found: YES (/usr/bin/cat) 00:02:40.295 Project name: DPDK 00:02:40.295 Project version: 22.11.4 00:02:40.295 C compiler for the host machine: gcc (gcc 13.3.1 "gcc (GCC) 13.3.1 20240522 (Red Hat 13.3.1-1)") 00:02:40.295 C linker for the host machine: gcc ld.bfd 2.40-14 00:02:40.295 Host machine cpu family: x86_64 00:02:40.295 Host machine cpu: x86_64 00:02:40.295 Message: ## Building in Developer Mode ## 00:02:40.295 Program pkg-config found: YES (/usr/bin/pkg-config) 00:02:40.295 Program check-symbols.sh found: YES (/home/vagrant/spdk_repo/dpdk/buildtools/check-symbols.sh) 00:02:40.295 Program options-ibverbs-static.sh found: YES (/home/vagrant/spdk_repo/dpdk/buildtools/options-ibverbs-static.sh) 00:02:40.295 Program objdump found: YES (/usr/bin/objdump) 00:02:40.295 Program python3 found: YES (/usr/bin/python3) 00:02:40.295 Program cat found: YES (/usr/bin/cat) 00:02:40.295 config/meson.build:83: WARNING: The "machine" option is deprecated. Please use "cpu_instruction_set" instead. 00:02:40.295 Checking for size of "void *" : 8 00:02:40.295 Checking for size of "void *" : 8 (cached) 00:02:40.295 Library m found: YES 00:02:40.295 Library numa found: YES 00:02:40.295 Has header "numaif.h" : YES 00:02:40.295 Library fdt found: NO 00:02:40.295 Library execinfo found: NO 00:02:40.295 Has header "execinfo.h" : YES 00:02:40.295 Found pkg-config: YES (/usr/bin/pkg-config) 1.9.5 00:02:40.295 Run-time dependency libarchive found: NO (tried pkgconfig) 00:02:40.295 Run-time dependency libbsd found: NO (tried pkgconfig) 00:02:40.295 Run-time dependency jansson found: NO (tried pkgconfig) 00:02:40.295 Run-time dependency openssl found: YES 3.1.1 00:02:40.295 Run-time dependency libpcap found: YES 1.10.4 00:02:40.295 Has header "pcap.h" with dependency libpcap: YES 00:02:40.295 Compiler for C supports arguments -Wcast-qual: YES 00:02:40.295 Compiler for C supports arguments -Wdeprecated: YES 00:02:40.295 Compiler for C supports arguments -Wformat: YES 00:02:40.295 Compiler for C supports arguments -Wformat-nonliteral: NO 00:02:40.295 Compiler for C supports arguments -Wformat-security: NO 00:02:40.295 Compiler for C supports arguments -Wmissing-declarations: YES 00:02:40.295 Compiler for C supports arguments -Wmissing-prototypes: YES 00:02:40.295 Compiler for C supports arguments -Wnested-externs: YES 00:02:40.295 Compiler for C supports arguments -Wold-style-definition: YES 00:02:40.295 Compiler for C supports arguments -Wpointer-arith: YES 00:02:40.295 Compiler for C supports arguments -Wsign-compare: YES 00:02:40.295 Compiler for C supports arguments -Wstrict-prototypes: YES 00:02:40.295 Compiler for C supports arguments -Wundef: YES 00:02:40.295 Compiler for C supports arguments -Wwrite-strings: YES 00:02:40.295 Compiler for C supports arguments -Wno-address-of-packed-member: YES 00:02:40.295 Compiler for C supports arguments -Wno-packed-not-aligned: YES 00:02:40.295 Compiler for C supports arguments -Wno-missing-field-initializers: YES 00:02:40.295 Compiler for C supports arguments -Wno-zero-length-bounds: YES 00:02:40.295 Compiler for C supports arguments -mavx512f: YES 00:02:40.295 Checking if "AVX512 checking" compiles: YES 00:02:40.295 Fetching value of define "__SSE4_2__" : 1 00:02:40.295 Fetching value of define "__AES__" : 1 00:02:40.295 Fetching value of define "__AVX__" : 1 00:02:40.295 Fetching value of define "__AVX2__" : 1 00:02:40.295 Fetching value of define "__AVX512BW__" : 1 00:02:40.295 Fetching value of define "__AVX512CD__" : 1 00:02:40.295 Fetching value of define "__AVX512DQ__" : 1 00:02:40.295 Fetching value of define "__AVX512F__" : 1 00:02:40.295 Fetching value of define "__AVX512VL__" : 1 00:02:40.295 Fetching value of define "__PCLMUL__" : 1 00:02:40.295 Fetching value of define "__RDRND__" : 1 00:02:40.295 Fetching value of define "__RDSEED__" : 1 00:02:40.295 Fetching value of define "__VPCLMULQDQ__" : (undefined) 00:02:40.295 Compiler for C supports arguments -Wno-format-truncation: YES 00:02:40.295 Message: lib/kvargs: Defining dependency "kvargs" 00:02:40.295 Message: lib/telemetry: Defining dependency "telemetry" 00:02:40.295 Checking for function "getentropy" : YES 00:02:40.295 Message: lib/eal: Defining dependency "eal" 00:02:40.295 Message: lib/ring: Defining dependency "ring" 00:02:40.295 Message: lib/rcu: Defining dependency "rcu" 00:02:40.295 Message: lib/mempool: Defining dependency "mempool" 00:02:40.295 Message: lib/mbuf: Defining dependency "mbuf" 00:02:40.295 Fetching value of define "__PCLMUL__" : 1 (cached) 00:02:40.295 Fetching value of define "__AVX512F__" : 1 (cached) 00:02:40.295 Fetching value of define "__AVX512BW__" : 1 (cached) 00:02:40.295 Fetching value of define "__AVX512DQ__" : 1 (cached) 00:02:40.295 Fetching value of define "__AVX512VL__" : 1 (cached) 00:02:40.295 Fetching value of define "__VPCLMULQDQ__" : (undefined) (cached) 00:02:40.295 Compiler for C supports arguments -mpclmul: YES 00:02:40.295 Compiler for C supports arguments -maes: YES 00:02:40.295 Compiler for C supports arguments -mavx512f: YES (cached) 00:02:40.295 Compiler for C supports arguments -mavx512bw: YES 00:02:40.295 Compiler for C supports arguments -mavx512dq: YES 00:02:40.295 Compiler for C supports arguments -mavx512vl: YES 00:02:40.295 Compiler for C supports arguments -mvpclmulqdq: YES 00:02:40.295 Compiler for C supports arguments -mavx2: YES 00:02:40.295 Compiler for C supports arguments -mavx: YES 00:02:40.295 Message: lib/net: Defining dependency "net" 00:02:40.295 Message: lib/meter: Defining dependency "meter" 00:02:40.295 Message: lib/ethdev: Defining dependency "ethdev" 00:02:40.295 Message: lib/pci: Defining dependency "pci" 00:02:40.295 Message: lib/cmdline: Defining dependency "cmdline" 00:02:40.295 Message: lib/metrics: Defining dependency "metrics" 00:02:40.295 Message: lib/hash: Defining dependency "hash" 00:02:40.295 Message: lib/timer: Defining dependency "timer" 00:02:40.295 Fetching value of define "__AVX2__" : 1 (cached) 00:02:40.295 Fetching value of define "__AVX512F__" : 1 (cached) 00:02:40.295 Fetching value of define "__AVX512VL__" : 1 (cached) 00:02:40.295 Fetching value of define "__AVX512CD__" : 1 (cached) 00:02:40.295 Fetching value of define "__AVX512BW__" : 1 (cached) 00:02:40.295 Message: lib/acl: Defining dependency "acl" 00:02:40.295 Message: lib/bbdev: Defining dependency "bbdev" 00:02:40.295 Message: lib/bitratestats: Defining dependency "bitratestats" 00:02:40.295 Run-time dependency libelf found: YES 0.191 00:02:40.295 Message: lib/bpf: Defining dependency "bpf" 00:02:40.295 Message: lib/cfgfile: Defining dependency "cfgfile" 00:02:40.295 Message: lib/compressdev: Defining dependency "compressdev" 00:02:40.295 Message: lib/cryptodev: Defining dependency "cryptodev" 00:02:40.295 Message: lib/distributor: Defining dependency "distributor" 00:02:40.295 Message: lib/efd: Defining dependency "efd" 00:02:40.295 Message: lib/eventdev: Defining dependency "eventdev" 00:02:40.295 Message: lib/gpudev: Defining dependency "gpudev" 00:02:40.295 Message: lib/gro: Defining dependency "gro" 00:02:40.295 Message: lib/gso: Defining dependency "gso" 00:02:40.295 Message: lib/ip_frag: Defining dependency "ip_frag" 00:02:40.295 Message: lib/jobstats: Defining dependency "jobstats" 00:02:40.295 Message: lib/latencystats: Defining dependency "latencystats" 00:02:40.295 Message: lib/lpm: Defining dependency "lpm" 00:02:40.295 Fetching value of define "__AVX512F__" : 1 (cached) 00:02:40.295 Fetching value of define "__AVX512DQ__" : 1 (cached) 00:02:40.295 Fetching value of define "__AVX512IFMA__" : (undefined) 00:02:40.295 Compiler for C supports arguments -mavx512f -mavx512dq -mavx512ifma: YES 00:02:40.295 Message: lib/member: Defining dependency "member" 00:02:40.295 Message: lib/pcapng: Defining dependency "pcapng" 00:02:40.295 Compiler for C supports arguments -Wno-cast-qual: YES 00:02:40.295 Message: lib/power: Defining dependency "power" 00:02:40.295 Message: lib/rawdev: Defining dependency "rawdev" 00:02:40.295 Message: lib/regexdev: Defining dependency "regexdev" 00:02:40.295 Message: lib/dmadev: Defining dependency "dmadev" 00:02:40.295 Message: lib/rib: Defining dependency "rib" 00:02:40.295 Message: lib/reorder: Defining dependency "reorder" 00:02:40.295 Message: lib/sched: Defining dependency "sched" 00:02:40.295 Message: lib/security: Defining dependency "security" 00:02:40.295 Message: lib/stack: Defining dependency "stack" 00:02:40.295 Has header "linux/userfaultfd.h" : YES 00:02:40.295 Message: lib/vhost: Defining dependency "vhost" 00:02:40.295 Message: lib/ipsec: Defining dependency "ipsec" 00:02:40.295 Fetching value of define "__AVX512F__" : 1 (cached) 00:02:40.295 Fetching value of define "__AVX512DQ__" : 1 (cached) 00:02:40.295 Fetching value of define "__AVX512BW__" : 1 (cached) 00:02:40.295 Message: lib/fib: Defining dependency "fib" 00:02:40.295 Message: lib/port: Defining dependency "port" 00:02:40.295 Message: lib/pdump: Defining dependency "pdump" 00:02:40.295 Message: lib/table: Defining dependency "table" 00:02:40.295 Message: lib/pipeline: Defining dependency "pipeline" 00:02:40.295 Message: lib/graph: Defining dependency "graph" 00:02:40.295 Message: lib/node: Defining dependency "node" 00:02:40.295 Compiler for C supports arguments -Wno-format-truncation: YES (cached) 00:02:40.295 Message: drivers/bus/pci: Defining dependency "bus_pci" 00:02:40.295 Message: drivers/bus/vdev: Defining dependency "bus_vdev" 00:02:40.295 Message: drivers/mempool/ring: Defining dependency "mempool_ring" 00:02:40.295 Compiler for C supports arguments -Wno-sign-compare: YES 00:02:40.295 Compiler for C supports arguments -Wno-unused-value: YES 00:02:40.295 Compiler for C supports arguments -Wno-format: YES 00:02:40.295 Compiler for C supports arguments -Wno-format-security: YES 00:02:40.295 Compiler for C supports arguments -Wno-format-nonliteral: YES 00:02:40.296 Compiler for C supports arguments -Wno-strict-aliasing: YES 00:02:40.866 Compiler for C supports arguments -Wno-unused-but-set-variable: YES 00:02:40.866 Compiler for C supports arguments -Wno-unused-parameter: YES 00:02:40.866 Fetching value of define "__AVX2__" : 1 (cached) 00:02:40.866 Fetching value of define "__AVX512F__" : 1 (cached) 00:02:40.866 Fetching value of define "__AVX512BW__" : 1 (cached) 00:02:40.866 Compiler for C supports arguments -mavx512f: YES (cached) 00:02:40.866 Compiler for C supports arguments -mavx512bw: YES (cached) 00:02:40.866 Compiler for C supports arguments -march=skylake-avx512: YES 00:02:40.866 Message: drivers/net/i40e: Defining dependency "net_i40e" 00:02:40.866 Program doxygen found: YES (/usr/local/bin/doxygen) 00:02:40.866 Configuring doxy-api.conf using configuration 00:02:40.866 Program sphinx-build found: NO 00:02:40.866 Configuring rte_build_config.h using configuration 00:02:40.866 Message: 00:02:40.866 ================= 00:02:40.866 Applications Enabled 00:02:40.866 ================= 00:02:40.866 00:02:40.866 apps: 00:02:40.866 dumpcap, pdump, proc-info, test-acl, test-bbdev, test-cmdline, test-compress-perf, test-crypto-perf, 00:02:40.866 test-eventdev, test-fib, test-flow-perf, test-gpudev, test-pipeline, test-pmd, test-regex, test-sad, 00:02:40.866 test-security-perf, 00:02:40.866 00:02:40.866 Message: 00:02:40.866 ================= 00:02:40.866 Libraries Enabled 00:02:40.866 ================= 00:02:40.866 00:02:40.866 libs: 00:02:40.866 kvargs, telemetry, eal, ring, rcu, mempool, mbuf, net, 00:02:40.866 meter, ethdev, pci, cmdline, metrics, hash, timer, acl, 00:02:40.866 bbdev, bitratestats, bpf, cfgfile, compressdev, cryptodev, distributor, efd, 00:02:40.866 eventdev, gpudev, gro, gso, ip_frag, jobstats, latencystats, lpm, 00:02:40.866 member, pcapng, power, rawdev, regexdev, dmadev, rib, reorder, 00:02:40.866 sched, security, stack, vhost, ipsec, fib, port, pdump, 00:02:40.866 table, pipeline, graph, node, 00:02:40.866 00:02:40.866 Message: 00:02:40.866 =============== 00:02:40.866 Drivers Enabled 00:02:40.866 =============== 00:02:40.866 00:02:40.866 common: 00:02:40.866 00:02:40.866 bus: 00:02:40.866 pci, vdev, 00:02:40.866 mempool: 00:02:40.866 ring, 00:02:40.866 dma: 00:02:40.867 00:02:40.867 net: 00:02:40.867 i40e, 00:02:40.867 raw: 00:02:40.867 00:02:40.867 crypto: 00:02:40.867 00:02:40.867 compress: 00:02:40.867 00:02:40.867 regex: 00:02:40.867 00:02:40.867 vdpa: 00:02:40.867 00:02:40.867 event: 00:02:40.867 00:02:40.867 baseband: 00:02:40.867 00:02:40.867 gpu: 00:02:40.867 00:02:40.867 00:02:40.867 Message: 00:02:40.867 ================= 00:02:40.867 Content Skipped 00:02:40.867 ================= 00:02:40.867 00:02:40.867 apps: 00:02:40.867 00:02:40.867 libs: 00:02:40.867 kni: explicitly disabled via build config (deprecated lib) 00:02:40.867 flow_classify: explicitly disabled via build config (deprecated lib) 00:02:40.867 00:02:40.867 drivers: 00:02:40.867 common/cpt: not in enabled drivers build config 00:02:40.867 common/dpaax: not in enabled drivers build config 00:02:40.867 common/iavf: not in enabled drivers build config 00:02:40.867 common/idpf: not in enabled drivers build config 00:02:40.867 common/mvep: not in enabled drivers build config 00:02:40.867 common/octeontx: not in enabled drivers build config 00:02:40.867 bus/auxiliary: not in enabled drivers build config 00:02:40.867 bus/dpaa: not in enabled drivers build config 00:02:40.867 bus/fslmc: not in enabled drivers build config 00:02:40.867 bus/ifpga: not in enabled drivers build config 00:02:40.867 bus/vmbus: not in enabled drivers build config 00:02:40.867 common/cnxk: not in enabled drivers build config 00:02:40.867 common/mlx5: not in enabled drivers build config 00:02:40.867 common/qat: not in enabled drivers build config 00:02:40.867 common/sfc_efx: not in enabled drivers build config 00:02:40.867 mempool/bucket: not in enabled drivers build config 00:02:40.867 mempool/cnxk: not in enabled drivers build config 00:02:40.867 mempool/dpaa: not in enabled drivers build config 00:02:40.867 mempool/dpaa2: not in enabled drivers build config 00:02:40.867 mempool/octeontx: not in enabled drivers build config 00:02:40.867 mempool/stack: not in enabled drivers build config 00:02:40.867 dma/cnxk: not in enabled drivers build config 00:02:40.867 dma/dpaa: not in enabled drivers build config 00:02:40.867 dma/dpaa2: not in enabled drivers build config 00:02:40.867 dma/hisilicon: not in enabled drivers build config 00:02:40.867 dma/idxd: not in enabled drivers build config 00:02:40.867 dma/ioat: not in enabled drivers build config 00:02:40.867 dma/skeleton: not in enabled drivers build config 00:02:40.867 net/af_packet: not in enabled drivers build config 00:02:40.867 net/af_xdp: not in enabled drivers build config 00:02:40.867 net/ark: not in enabled drivers build config 00:02:40.867 net/atlantic: not in enabled drivers build config 00:02:40.867 net/avp: not in enabled drivers build config 00:02:40.867 net/axgbe: not in enabled drivers build config 00:02:40.867 net/bnx2x: not in enabled drivers build config 00:02:40.867 net/bnxt: not in enabled drivers build config 00:02:40.867 net/bonding: not in enabled drivers build config 00:02:40.867 net/cnxk: not in enabled drivers build config 00:02:40.867 net/cxgbe: not in enabled drivers build config 00:02:40.867 net/dpaa: not in enabled drivers build config 00:02:40.867 net/dpaa2: not in enabled drivers build config 00:02:40.867 net/e1000: not in enabled drivers build config 00:02:40.867 net/ena: not in enabled drivers build config 00:02:40.867 net/enetc: not in enabled drivers build config 00:02:40.867 net/enetfec: not in enabled drivers build config 00:02:40.867 net/enic: not in enabled drivers build config 00:02:40.867 net/failsafe: not in enabled drivers build config 00:02:40.867 net/fm10k: not in enabled drivers build config 00:02:40.867 net/gve: not in enabled drivers build config 00:02:40.867 net/hinic: not in enabled drivers build config 00:02:40.867 net/hns3: not in enabled drivers build config 00:02:40.867 net/iavf: not in enabled drivers build config 00:02:40.867 net/ice: not in enabled drivers build config 00:02:40.867 net/idpf: not in enabled drivers build config 00:02:40.867 net/igc: not in enabled drivers build config 00:02:40.867 net/ionic: not in enabled drivers build config 00:02:40.867 net/ipn3ke: not in enabled drivers build config 00:02:40.867 net/ixgbe: not in enabled drivers build config 00:02:40.867 net/kni: not in enabled drivers build config 00:02:40.867 net/liquidio: not in enabled drivers build config 00:02:40.867 net/mana: not in enabled drivers build config 00:02:40.867 net/memif: not in enabled drivers build config 00:02:40.867 net/mlx4: not in enabled drivers build config 00:02:40.867 net/mlx5: not in enabled drivers build config 00:02:40.867 net/mvneta: not in enabled drivers build config 00:02:40.867 net/mvpp2: not in enabled drivers build config 00:02:40.867 net/netvsc: not in enabled drivers build config 00:02:40.867 net/nfb: not in enabled drivers build config 00:02:40.867 net/nfp: not in enabled drivers build config 00:02:40.867 net/ngbe: not in enabled drivers build config 00:02:40.867 net/null: not in enabled drivers build config 00:02:40.867 net/octeontx: not in enabled drivers build config 00:02:40.867 net/octeon_ep: not in enabled drivers build config 00:02:40.867 net/pcap: not in enabled drivers build config 00:02:40.867 net/pfe: not in enabled drivers build config 00:02:40.867 net/qede: not in enabled drivers build config 00:02:40.867 net/ring: not in enabled drivers build config 00:02:40.867 net/sfc: not in enabled drivers build config 00:02:40.867 net/softnic: not in enabled drivers build config 00:02:40.867 net/tap: not in enabled drivers build config 00:02:40.867 net/thunderx: not in enabled drivers build config 00:02:40.867 net/txgbe: not in enabled drivers build config 00:02:40.867 net/vdev_netvsc: not in enabled drivers build config 00:02:40.867 net/vhost: not in enabled drivers build config 00:02:40.867 net/virtio: not in enabled drivers build config 00:02:40.867 net/vmxnet3: not in enabled drivers build config 00:02:40.867 raw/cnxk_bphy: not in enabled drivers build config 00:02:40.867 raw/cnxk_gpio: not in enabled drivers build config 00:02:40.867 raw/dpaa2_cmdif: not in enabled drivers build config 00:02:40.867 raw/ifpga: not in enabled drivers build config 00:02:40.867 raw/ntb: not in enabled drivers build config 00:02:40.867 raw/skeleton: not in enabled drivers build config 00:02:40.867 crypto/armv8: not in enabled drivers build config 00:02:40.867 crypto/bcmfs: not in enabled drivers build config 00:02:40.867 crypto/caam_jr: not in enabled drivers build config 00:02:40.867 crypto/ccp: not in enabled drivers build config 00:02:40.867 crypto/cnxk: not in enabled drivers build config 00:02:40.867 crypto/dpaa_sec: not in enabled drivers build config 00:02:40.867 crypto/dpaa2_sec: not in enabled drivers build config 00:02:40.867 crypto/ipsec_mb: not in enabled drivers build config 00:02:40.867 crypto/mlx5: not in enabled drivers build config 00:02:40.867 crypto/mvsam: not in enabled drivers build config 00:02:40.867 crypto/nitrox: not in enabled drivers build config 00:02:40.867 crypto/null: not in enabled drivers build config 00:02:40.867 crypto/octeontx: not in enabled drivers build config 00:02:40.867 crypto/openssl: not in enabled drivers build config 00:02:40.867 crypto/scheduler: not in enabled drivers build config 00:02:40.867 crypto/uadk: not in enabled drivers build config 00:02:40.867 crypto/virtio: not in enabled drivers build config 00:02:40.867 compress/isal: not in enabled drivers build config 00:02:40.867 compress/mlx5: not in enabled drivers build config 00:02:40.867 compress/octeontx: not in enabled drivers build config 00:02:40.867 compress/zlib: not in enabled drivers build config 00:02:40.867 regex/mlx5: not in enabled drivers build config 00:02:40.867 regex/cn9k: not in enabled drivers build config 00:02:40.867 vdpa/ifc: not in enabled drivers build config 00:02:40.867 vdpa/mlx5: not in enabled drivers build config 00:02:40.867 vdpa/sfc: not in enabled drivers build config 00:02:40.867 event/cnxk: not in enabled drivers build config 00:02:40.867 event/dlb2: not in enabled drivers build config 00:02:40.867 event/dpaa: not in enabled drivers build config 00:02:40.867 event/dpaa2: not in enabled drivers build config 00:02:40.867 event/dsw: not in enabled drivers build config 00:02:40.867 event/opdl: not in enabled drivers build config 00:02:40.867 event/skeleton: not in enabled drivers build config 00:02:40.867 event/sw: not in enabled drivers build config 00:02:40.867 event/octeontx: not in enabled drivers build config 00:02:40.867 baseband/acc: not in enabled drivers build config 00:02:40.867 baseband/fpga_5gnr_fec: not in enabled drivers build config 00:02:40.867 baseband/fpga_lte_fec: not in enabled drivers build config 00:02:40.867 baseband/la12xx: not in enabled drivers build config 00:02:40.867 baseband/null: not in enabled drivers build config 00:02:40.867 baseband/turbo_sw: not in enabled drivers build config 00:02:40.867 gpu/cuda: not in enabled drivers build config 00:02:40.867 00:02:40.867 00:02:40.867 Build targets in project: 311 00:02:40.867 00:02:40.867 DPDK 22.11.4 00:02:40.867 00:02:40.867 User defined options 00:02:40.867 libdir : lib 00:02:40.867 prefix : /home/vagrant/spdk_repo/dpdk/build 00:02:40.867 c_args : -fPIC -g -fcommon -Werror -Wno-stringop-overflow 00:02:40.867 c_link_args : 00:02:40.867 enable_docs : false 00:02:40.867 enable_drivers: bus,bus/pci,bus/vdev,mempool/ring,net/i40e,net/i40e/base, 00:02:40.867 enable_kmods : false 00:02:40.867 machine : native 00:02:40.867 tests : false 00:02:40.867 00:02:40.867 Found ninja-1.11.1.git.kitware.jobserver-1 at /usr/local/bin/ninja 00:02:40.867 WARNING: Running the setup command as `meson [options]` instead of `meson setup [options]` is ambiguous and deprecated. 00:02:40.867 17:24:11 build_native_dpdk -- common/autobuild_common.sh@192 -- $ ninja -C /home/vagrant/spdk_repo/dpdk/build-tmp -j10 00:02:40.867 ninja: Entering directory `/home/vagrant/spdk_repo/dpdk/build-tmp' 00:02:40.867 [1/740] Generating lib/rte_telemetry_mingw with a custom command 00:02:40.867 [2/740] Generating lib/rte_kvargs_def with a custom command 00:02:40.867 [3/740] Generating lib/rte_telemetry_def with a custom command 00:02:40.867 [4/740] Generating lib/rte_kvargs_mingw with a custom command 00:02:40.867 [5/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_class.c.o 00:02:40.867 [6/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_debug.c.o 00:02:41.127 [7/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_errno.c.o 00:02:41.127 [8/740] Compiling C object lib/librte_kvargs.a.p/kvargs_rte_kvargs.c.o 00:02:41.127 [9/740] Linking static target lib/librte_kvargs.a 00:02:41.127 [10/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_config.c.o 00:02:41.127 [11/740] Compiling C object lib/librte_telemetry.a.p/telemetry_telemetry_data.c.o 00:02:41.127 [12/740] Compiling C object lib/librte_telemetry.a.p/telemetry_telemetry_legacy.c.o 00:02:41.127 [13/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_bus.c.o 00:02:41.127 [14/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_hexdump.c.o 00:02:41.127 [15/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_launch.c.o 00:02:41.127 [16/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_devargs.c.o 00:02:41.127 [17/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_dev.c.o 00:02:41.127 [18/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_interrupts.c.o 00:02:41.127 [19/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_mcfg.c.o 00:02:41.127 [20/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_log.c.o 00:02:41.127 [21/740] Generating lib/kvargs.sym_chk with a custom command (wrapped by meson to capture output) 00:02:41.127 [22/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_lcore.c.o 00:02:41.127 [23/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_string_fns.c.o 00:02:41.127 [24/740] Linking target lib/librte_kvargs.so.23.0 00:02:41.387 [25/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_memalloc.c.o 00:02:41.387 [26/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_tailqs.c.o 00:02:41.387 [27/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_timer.c.o 00:02:41.387 [28/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_memzone.c.o 00:02:41.387 [29/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_fbarray.c.o 00:02:41.387 [30/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_thread.c.o 00:02:41.387 [31/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_uuid.c.o 00:02:41.387 [32/740] Compiling C object lib/librte_telemetry.a.p/telemetry_telemetry.c.o 00:02:41.387 [33/740] Linking static target lib/librte_telemetry.a 00:02:41.387 [34/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_trace_points.c.o 00:02:41.387 [35/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_memory.c.o 00:02:41.387 [36/740] Compiling C object lib/librte_eal.a.p/eal_common_rte_reciprocal.c.o 00:02:41.387 [37/740] Compiling C object lib/librte_eal.a.p/eal_common_rte_version.c.o 00:02:41.387 [38/740] Compiling C object lib/librte_eal.a.p/eal_common_malloc_elem.c.o 00:02:41.387 [39/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_cpuflags.c.o 00:02:41.646 [40/740] Generating symbol file lib/librte_kvargs.so.23.0.p/librte_kvargs.so.23.0.symbols 00:02:41.646 [41/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_hypervisor.c.o 00:02:41.646 [42/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_dynmem.c.o 00:02:41.646 [43/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_trace_ctf.c.o 00:02:41.646 [44/740] Generating lib/telemetry.sym_chk with a custom command (wrapped by meson to capture output) 00:02:41.646 [45/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_trace.c.o 00:02:41.646 [46/740] Compiling C object lib/librte_eal.a.p/eal_common_malloc_heap.c.o 00:02:41.646 [47/740] Linking target lib/librte_telemetry.so.23.0 00:02:41.646 [48/740] Compiling C object lib/librte_eal.a.p/eal_common_rte_random.c.o 00:02:41.646 [49/740] Compiling C object lib/librte_eal.a.p/eal_common_rte_service.c.o 00:02:41.905 [50/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_trace_utils.c.o 00:02:41.905 [51/740] Compiling C object lib/librte_eal.a.p/eal_common_rte_keepalive.c.o 00:02:41.905 [52/740] Compiling C object lib/librte_eal.a.p/eal_unix_eal_debug.c.o 00:02:41.905 [53/740] Compiling C object lib/librte_eal.a.p/eal_common_hotplug_mp.c.o 00:02:41.905 [54/740] Compiling C object lib/librte_eal.a.p/eal_unix_eal_file.c.o 00:02:41.905 [55/740] Generating symbol file lib/librte_telemetry.so.23.0.p/librte_telemetry.so.23.0.symbols 00:02:41.905 [56/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_proc.c.o 00:02:41.905 [57/740] Compiling C object lib/librte_eal.a.p/eal_unix_eal_firmware.c.o 00:02:41.905 [58/740] Compiling C object lib/librte_eal.a.p/eal_unix_eal_filesystem.c.o 00:02:41.905 [59/740] Compiling C object lib/librte_eal.a.p/eal_unix_eal_unix_timer.c.o 00:02:41.905 [60/740] Compiling C object lib/librte_eal.a.p/eal_unix_eal_unix_thread.c.o 00:02:41.905 [61/740] Compiling C object lib/librte_eal.a.p/eal_unix_eal_unix_memory.c.o 00:02:41.905 [62/740] Compiling C object lib/librte_eal.a.p/eal_common_malloc_mp.c.o 00:02:41.905 [63/740] Compiling C object lib/librte_eal.a.p/eal_linux_eal_cpuflags.c.o 00:02:41.905 [64/740] Compiling C object lib/librte_eal.a.p/eal_unix_rte_thread.c.o 00:02:41.905 [65/740] Compiling C object lib/librte_eal.a.p/eal_common_rte_malloc.c.o 00:02:41.906 [66/740] Compiling C object lib/librte_eal.a.p/eal_linux_eal_log.c.o 00:02:41.906 [67/740] Compiling C object lib/librte_eal.a.p/eal_linux_eal_lcore.c.o 00:02:41.906 [68/740] Compiling C object lib/librte_eal.a.p/eal_linux_eal_alarm.c.o 00:02:41.906 [69/740] Compiling C object lib/librte_eal.a.p/eal_linux_eal_thread.c.o 00:02:42.165 [70/740] Compiling C object lib/librte_eal.a.p/eal_linux_eal_dev.c.o 00:02:42.165 [71/740] Compiling C object lib/librte_eal.a.p/eal_linux_eal_timer.c.o 00:02:42.165 [72/740] Compiling C object lib/librte_eal.a.p/eal_linux_eal_vfio_mp_sync.c.o 00:02:42.165 [73/740] Compiling C object lib/librte_eal.a.p/eal_linux_eal_hugepage_info.c.o 00:02:42.165 [74/740] Compiling C object lib/librte_eal.a.p/eal_x86_rte_cpuflags.c.o 00:02:42.165 [75/740] Compiling C object lib/librte_eal.a.p/eal_x86_rte_hypervisor.c.o 00:02:42.165 [76/740] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_options.c.o 00:02:42.165 [77/740] Compiling C object lib/librte_eal.a.p/eal_x86_rte_spinlock.c.o 00:02:42.165 [78/740] Generating lib/rte_eal_def with a custom command 00:02:42.165 [79/740] Compiling C object lib/librte_eal.a.p/eal_x86_rte_cycles.c.o 00:02:42.165 [80/740] Generating lib/rte_eal_mingw with a custom command 00:02:42.165 [81/740] Generating lib/rte_ring_def with a custom command 00:02:42.165 [82/740] Generating lib/rte_ring_mingw with a custom command 00:02:42.165 [83/740] Compiling C object lib/librte_eal.a.p/eal_linux_eal.c.o 00:02:42.165 [84/740] Generating lib/rte_rcu_def with a custom command 00:02:42.165 [85/740] Generating lib/rte_rcu_mingw with a custom command 00:02:42.165 [86/740] Compiling C object lib/librte_eal.a.p/eal_x86_rte_power_intrinsics.c.o 00:02:42.165 [87/740] Compiling C object lib/librte_ring.a.p/ring_rte_ring.c.o 00:02:42.165 [88/740] Linking static target lib/librte_ring.a 00:02:42.165 [89/740] Compiling C object lib/librte_eal.a.p/eal_linux_eal_interrupts.c.o 00:02:42.425 [90/740] Generating lib/rte_mempool_def with a custom command 00:02:42.425 [91/740] Generating lib/rte_mempool_mingw with a custom command 00:02:42.425 [92/740] Compiling C object lib/librte_eal.a.p/eal_linux_eal_memalloc.c.o 00:02:42.425 [93/740] Compiling C object lib/librte_eal.a.p/eal_linux_eal_memory.c.o 00:02:42.425 [94/740] Generating lib/ring.sym_chk with a custom command (wrapped by meson to capture output) 00:02:42.425 [95/740] Compiling C object lib/librte_mempool.a.p/mempool_rte_mempool_ops_default.c.o 00:02:42.425 [96/740] Compiling C object lib/librte_eal.a.p/eal_linux_eal_vfio.c.o 00:02:42.425 [97/740] Generating lib/rte_mbuf_def with a custom command 00:02:42.685 [98/740] Compiling C object lib/librte_mempool.a.p/mempool_rte_mempool_ops.c.o 00:02:42.685 [99/740] Linking static target lib/librte_eal.a 00:02:42.685 [100/740] Generating lib/rte_mbuf_mingw with a custom command 00:02:42.685 [101/740] Compiling C object lib/librte_mempool.a.p/mempool_mempool_trace_points.c.o 00:02:42.685 [102/740] Compiling C object lib/librte_mbuf.a.p/mbuf_rte_mbuf_ptype.c.o 00:02:42.685 [103/740] Compiling C object lib/librte_mbuf.a.p/mbuf_rte_mbuf_pool_ops.c.o 00:02:42.685 [104/740] Compiling C object lib/librte_rcu.a.p/rcu_rte_rcu_qsbr.c.o 00:02:42.685 [105/740] Linking static target lib/librte_rcu.a 00:02:42.945 [106/740] Compiling C object lib/librte_mempool.a.p/mempool_rte_mempool.c.o 00:02:42.945 [107/740] Linking static target lib/librte_mempool.a 00:02:42.945 [108/740] Compiling C object lib/librte_mbuf.a.p/mbuf_rte_mbuf_dyn.c.o 00:02:42.945 [109/740] Compiling C object lib/librte_net.a.p/net_rte_ether.c.o 00:02:42.945 [110/740] Generating lib/rte_net_def with a custom command 00:02:42.945 [111/740] Compiling C object lib/net/libnet_crc_avx512_lib.a.p/net_crc_avx512.c.o 00:02:42.945 [112/740] Linking static target lib/net/libnet_crc_avx512_lib.a 00:02:42.945 [113/740] Generating lib/rte_net_mingw with a custom command 00:02:42.945 [114/740] Generating lib/rte_meter_def with a custom command 00:02:42.945 [115/740] Compiling C object lib/librte_net.a.p/net_rte_arp.c.o 00:02:42.945 [116/740] Compiling C object lib/librte_net.a.p/net_rte_net_crc.c.o 00:02:42.945 [117/740] Generating lib/rcu.sym_chk with a custom command (wrapped by meson to capture output) 00:02:42.945 [118/740] Generating lib/rte_meter_mingw with a custom command 00:02:43.205 [119/740] Compiling C object lib/librte_meter.a.p/meter_rte_meter.c.o 00:02:43.205 [120/740] Linking static target lib/librte_meter.a 00:02:43.205 [121/740] Compiling C object lib/librte_net.a.p/net_rte_net.c.o 00:02:43.205 [122/740] Compiling C object lib/librte_net.a.p/net_net_crc_sse.c.o 00:02:43.205 [123/740] Linking static target lib/librte_net.a 00:02:43.205 [124/740] Generating lib/meter.sym_chk with a custom command (wrapped by meson to capture output) 00:02:43.464 [125/740] Compiling C object lib/librte_ethdev.a.p/ethdev_ethdev_profile.c.o 00:02:43.464 [126/740] Compiling C object lib/librte_mbuf.a.p/mbuf_rte_mbuf.c.o 00:02:43.464 [127/740] Linking static target lib/librte_mbuf.a 00:02:43.464 [128/740] Compiling C object lib/librte_ethdev.a.p/ethdev_ethdev_trace_points.c.o 00:02:43.464 [129/740] Generating lib/net.sym_chk with a custom command (wrapped by meson to capture output) 00:02:43.464 [130/740] Generating lib/mempool.sym_chk with a custom command (wrapped by meson to capture output) 00:02:43.464 [131/740] Compiling C object lib/librte_ethdev.a.p/ethdev_ethdev_private.c.o 00:02:43.464 [132/740] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_class_eth.c.o 00:02:43.464 [133/740] Compiling C object lib/librte_ethdev.a.p/ethdev_ethdev_driver.c.o 00:02:43.724 [134/740] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_ethdev_cman.c.o 00:02:43.724 [135/740] Generating lib/mbuf.sym_chk with a custom command (wrapped by meson to capture output) 00:02:43.997 [136/740] Compiling C object lib/librte_ethdev.a.p/ethdev_sff_telemetry.c.o 00:02:43.997 [137/740] Generating lib/rte_ethdev_def with a custom command 00:02:43.997 [138/740] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_mtr.c.o 00:02:43.997 [139/740] Generating lib/rte_ethdev_mingw with a custom command 00:02:43.997 [140/740] Generating lib/rte_pci_def with a custom command 00:02:43.997 [141/740] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_tm.c.o 00:02:43.997 [142/740] Generating lib/rte_pci_mingw with a custom command 00:02:43.997 [143/740] Compiling C object lib/librte_pci.a.p/pci_rte_pci.c.o 00:02:43.997 [144/740] Linking static target lib/librte_pci.a 00:02:43.997 [145/740] Compiling C object lib/librte_ethdev.a.p/ethdev_sff_8079.c.o 00:02:43.997 [146/740] Compiling C object lib/librte_ethdev.a.p/ethdev_sff_common.c.o 00:02:43.997 [147/740] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline.c.o 00:02:44.275 [148/740] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_cirbuf.c.o 00:02:44.275 [149/740] Generating lib/pci.sym_chk with a custom command (wrapped by meson to capture output) 00:02:44.275 [150/740] Compiling C object lib/librte_ethdev.a.p/ethdev_sff_8472.c.o 00:02:44.275 [151/740] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse.c.o 00:02:44.275 [152/740] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse_num.c.o 00:02:44.275 [153/740] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse_portlist.c.o 00:02:44.275 [154/740] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse_string.c.o 00:02:44.275 [155/740] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_socket.c.o 00:02:44.275 [156/740] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_vt100.c.o 00:02:44.275 [157/740] Generating lib/rte_cmdline_def with a custom command 00:02:44.275 [158/740] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_os_unix.c.o 00:02:44.275 [159/740] Generating lib/rte_cmdline_mingw with a custom command 00:02:44.275 [160/740] Generating lib/rte_metrics_def with a custom command 00:02:44.275 [161/740] Generating lib/rte_metrics_mingw with a custom command 00:02:44.275 [162/740] Compiling C object lib/librte_ethdev.a.p/ethdev_sff_8636.c.o 00:02:44.534 [163/740] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse_etheraddr.c.o 00:02:44.534 [164/740] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse_ipaddr.c.o 00:02:44.534 [165/740] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_rdline.c.o 00:02:44.534 [166/740] Linking static target lib/librte_cmdline.a 00:02:44.535 [167/740] Compiling C object lib/librte_metrics.a.p/metrics_rte_metrics.c.o 00:02:44.535 [168/740] Generating lib/rte_hash_def with a custom command 00:02:44.535 [169/740] Generating lib/rte_hash_mingw with a custom command 00:02:44.535 [170/740] Generating lib/rte_timer_def with a custom command 00:02:44.535 [171/740] Generating lib/rte_timer_mingw with a custom command 00:02:44.535 [172/740] Compiling C object lib/librte_hash.a.p/hash_rte_fbk_hash.c.o 00:02:44.535 [173/740] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_flow.c.o 00:02:44.794 [174/740] Compiling C object lib/librte_metrics.a.p/metrics_rte_metrics_telemetry.c.o 00:02:44.794 [175/740] Linking static target lib/librte_metrics.a 00:02:44.794 [176/740] Compiling C object lib/librte_timer.a.p/timer_rte_timer.c.o 00:02:44.794 [177/740] Linking static target lib/librte_timer.a 00:02:45.054 [178/740] Generating lib/metrics.sym_chk with a custom command (wrapped by meson to capture output) 00:02:45.054 [179/740] Compiling C object lib/librte_acl.a.p/acl_acl_gen.c.o 00:02:45.054 [180/740] Compiling C object lib/librte_hash.a.p/hash_rte_thash.c.o 00:02:45.054 [181/740] Generating lib/timer.sym_chk with a custom command (wrapped by meson to capture output) 00:02:45.054 [182/740] Generating lib/cmdline.sym_chk with a custom command (wrapped by meson to capture output) 00:02:45.313 [183/740] Compiling C object lib/librte_acl.a.p/acl_acl_run_scalar.c.o 00:02:45.313 [184/740] Generating lib/rte_acl_def with a custom command 00:02:45.313 [185/740] Generating lib/rte_acl_mingw with a custom command 00:02:45.313 [186/740] Compiling C object lib/librte_acl.a.p/acl_tb_mem.c.o 00:02:45.314 [187/740] Generating lib/rte_bbdev_def with a custom command 00:02:45.314 [188/740] Generating lib/rte_bbdev_mingw with a custom command 00:02:45.314 [189/740] Compiling C object lib/librte_acl.a.p/acl_rte_acl.c.o 00:02:45.314 [190/740] Generating lib/rte_bitratestats_def with a custom command 00:02:45.314 [191/740] Generating lib/rte_bitratestats_mingw with a custom command 00:02:45.573 [192/740] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_ethdev.c.o 00:02:45.573 [193/740] Linking static target lib/librte_ethdev.a 00:02:45.833 [194/740] Compiling C object lib/librte_bitratestats.a.p/bitratestats_rte_bitrate.c.o 00:02:45.833 [195/740] Linking static target lib/librte_bitratestats.a 00:02:45.833 [196/740] Compiling C object lib/librte_bpf.a.p/bpf_bpf.c.o 00:02:45.833 [197/740] Compiling C object lib/librte_acl.a.p/acl_acl_bld.c.o 00:02:45.833 [198/740] Compiling C object lib/librte_bbdev.a.p/bbdev_rte_bbdev.c.o 00:02:46.093 [199/740] Generating lib/bitratestats.sym_chk with a custom command (wrapped by meson to capture output) 00:02:46.093 [200/740] Linking static target lib/librte_bbdev.a 00:02:46.093 [201/740] Compiling C object lib/librte_bpf.a.p/bpf_bpf_dump.c.o 00:02:46.353 [202/740] Compiling C object lib/librte_bpf.a.p/bpf_bpf_load.c.o 00:02:46.353 [203/740] Compiling C object lib/librte_bpf.a.p/bpf_bpf_exec.c.o 00:02:46.353 [204/740] Generating lib/bbdev.sym_chk with a custom command (wrapped by meson to capture output) 00:02:46.613 [205/740] Compiling C object lib/librte_hash.a.p/hash_rte_cuckoo_hash.c.o 00:02:46.613 [206/740] Linking static target lib/librte_hash.a 00:02:46.613 [207/740] Compiling C object lib/librte_acl.a.p/acl_acl_run_sse.c.o 00:02:46.613 [208/740] Compiling C object lib/librte_bpf.a.p/bpf_bpf_stub.c.o 00:02:46.871 [209/740] Compiling C object lib/librte_bpf.a.p/bpf_bpf_pkt.c.o 00:02:46.871 [210/740] Generating lib/rte_bpf_def with a custom command 00:02:46.871 [211/740] Generating lib/rte_bpf_mingw with a custom command 00:02:47.130 [212/740] Compiling C object lib/librte_bpf.a.p/bpf_bpf_load_elf.c.o 00:02:47.130 [213/740] Generating lib/rte_cfgfile_def with a custom command 00:02:47.130 [214/740] Generating lib/rte_cfgfile_mingw with a custom command 00:02:47.130 [215/740] Compiling C object lib/librte_bpf.a.p/bpf_bpf_convert.c.o 00:02:47.130 [216/740] Compiling C object lib/librte_acl.a.p/acl_acl_run_avx2.c.o 00:02:47.130 [217/740] Compiling C object lib/librte_bpf.a.p/bpf_bpf_validate.c.o 00:02:47.130 [218/740] Compiling C object lib/librte_cfgfile.a.p/cfgfile_rte_cfgfile.c.o 00:02:47.130 [219/740] Linking static target lib/librte_cfgfile.a 00:02:47.130 [220/740] Generating lib/rte_compressdev_def with a custom command 00:02:47.389 [221/740] Generating lib/rte_compressdev_mingw with a custom command 00:02:47.389 [222/740] Generating lib/hash.sym_chk with a custom command (wrapped by meson to capture output) 00:02:47.389 [223/740] Compiling C object lib/librte_bpf.a.p/bpf_bpf_jit_x86.c.o 00:02:47.389 [224/740] Linking static target lib/librte_bpf.a 00:02:47.389 [225/740] Compiling C object lib/librte_compressdev.a.p/compressdev_rte_compressdev_pmd.c.o 00:02:47.389 [226/740] Generating lib/cfgfile.sym_chk with a custom command (wrapped by meson to capture output) 00:02:47.648 [227/740] Generating lib/rte_cryptodev_def with a custom command 00:02:47.648 [228/740] Generating lib/rte_cryptodev_mingw with a custom command 00:02:47.648 [229/740] Compiling C object lib/librte_acl.a.p/acl_acl_run_avx512.c.o 00:02:47.648 [230/740] Linking static target lib/librte_acl.a 00:02:47.648 [231/740] Compiling C object lib/librte_compressdev.a.p/compressdev_rte_compressdev.c.o 00:02:47.648 [232/740] Generating lib/bpf.sym_chk with a custom command (wrapped by meson to capture output) 00:02:47.648 [233/740] Compiling C object lib/librte_cryptodev.a.p/cryptodev_cryptodev_pmd.c.o 00:02:47.648 [234/740] Compiling C object lib/librte_compressdev.a.p/compressdev_rte_comp.c.o 00:02:47.648 [235/740] Linking static target lib/librte_compressdev.a 00:02:47.648 [236/740] Generating lib/rte_distributor_def with a custom command 00:02:47.648 [237/740] Generating lib/rte_distributor_mingw with a custom command 00:02:47.648 [238/740] Compiling C object lib/librte_cryptodev.a.p/cryptodev_cryptodev_trace_points.c.o 00:02:47.908 [239/740] Generating lib/rte_efd_def with a custom command 00:02:47.908 [240/740] Generating lib/rte_efd_mingw with a custom command 00:02:47.908 [241/740] Generating lib/acl.sym_chk with a custom command (wrapped by meson to capture output) 00:02:47.908 [242/740] Compiling C object lib/librte_distributor.a.p/distributor_rte_distributor_match_sse.c.o 00:02:48.190 [243/740] Compiling C object lib/librte_distributor.a.p/distributor_rte_distributor_single.c.o 00:02:48.190 [244/740] Compiling C object lib/librte_eventdev.a.p/eventdev_eventdev_private.c.o 00:02:48.190 [245/740] Compiling C object lib/librte_distributor.a.p/distributor_rte_distributor.c.o 00:02:48.190 [246/740] Linking static target lib/librte_distributor.a 00:02:48.190 [247/740] Generating lib/eal.sym_chk with a custom command (wrapped by meson to capture output) 00:02:48.190 [248/740] Linking target lib/librte_eal.so.23.0 00:02:48.190 [249/740] Compiling C object lib/librte_eventdev.a.p/eventdev_eventdev_trace_points.c.o 00:02:48.452 [250/740] Generating symbol file lib/librte_eal.so.23.0.p/librte_eal.so.23.0.symbols 00:02:48.452 [251/740] Generating lib/compressdev.sym_chk with a custom command (wrapped by meson to capture output) 00:02:48.452 [252/740] Generating lib/distributor.sym_chk with a custom command (wrapped by meson to capture output) 00:02:48.452 [253/740] Linking target lib/librte_ring.so.23.0 00:02:48.452 [254/740] Linking target lib/librte_meter.so.23.0 00:02:48.452 [255/740] Linking target lib/librte_pci.so.23.0 00:02:48.452 [256/740] Generating symbol file lib/librte_ring.so.23.0.p/librte_ring.so.23.0.symbols 00:02:48.452 [257/740] Generating symbol file lib/librte_pci.so.23.0.p/librte_pci.so.23.0.symbols 00:02:48.452 [258/740] Generating symbol file lib/librte_meter.so.23.0.p/librte_meter.so.23.0.symbols 00:02:48.452 [259/740] Linking target lib/librte_rcu.so.23.0 00:02:48.452 [260/740] Linking target lib/librte_mempool.so.23.0 00:02:48.452 [261/740] Linking target lib/librte_timer.so.23.0 00:02:48.711 [262/740] Compiling C object lib/librte_eventdev.a.p/eventdev_rte_event_ring.c.o 00:02:48.711 [263/740] Linking target lib/librte_acl.so.23.0 00:02:48.711 [264/740] Generating symbol file lib/librte_rcu.so.23.0.p/librte_rcu.so.23.0.symbols 00:02:48.711 [265/740] Generating symbol file lib/librte_mempool.so.23.0.p/librte_mempool.so.23.0.symbols 00:02:48.711 [266/740] Generating symbol file lib/librte_timer.so.23.0.p/librte_timer.so.23.0.symbols 00:02:48.711 [267/740] Linking target lib/librte_cfgfile.so.23.0 00:02:48.711 [268/740] Linking target lib/librte_mbuf.so.23.0 00:02:48.711 [269/740] Generating symbol file lib/librte_acl.so.23.0.p/librte_acl.so.23.0.symbols 00:02:48.711 [270/740] Compiling C object lib/librte_efd.a.p/efd_rte_efd.c.o 00:02:48.711 [271/740] Generating lib/rte_eventdev_def with a custom command 00:02:48.711 [272/740] Linking static target lib/librte_efd.a 00:02:48.711 [273/740] Generating lib/rte_eventdev_mingw with a custom command 00:02:48.711 [274/740] Generating symbol file lib/librte_mbuf.so.23.0.p/librte_mbuf.so.23.0.symbols 00:02:48.970 [275/740] Linking target lib/librte_net.so.23.0 00:02:48.970 [276/740] Compiling C object lib/librte_eventdev.a.p/eventdev_rte_event_crypto_adapter.c.o 00:02:48.970 [277/740] Generating symbol file lib/librte_net.so.23.0.p/librte_net.so.23.0.symbols 00:02:48.970 [278/740] Linking target lib/librte_bbdev.so.23.0 00:02:48.970 [279/740] Linking target lib/librte_cmdline.so.23.0 00:02:48.970 [280/740] Generating lib/efd.sym_chk with a custom command (wrapped by meson to capture output) 00:02:48.970 [281/740] Linking target lib/librte_hash.so.23.0 00:02:48.970 [282/740] Linking target lib/librte_compressdev.so.23.0 00:02:48.970 [283/740] Linking target lib/librte_distributor.so.23.0 00:02:48.970 [284/740] Generating lib/rte_gpudev_def with a custom command 00:02:48.970 [285/740] Generating lib/rte_gpudev_mingw with a custom command 00:02:49.229 [286/740] Generating symbol file lib/librte_hash.so.23.0.p/librte_hash.so.23.0.symbols 00:02:49.229 [287/740] Compiling C object lib/librte_eventdev.a.p/eventdev_rte_event_eth_tx_adapter.c.o 00:02:49.229 [288/740] Linking target lib/librte_efd.so.23.0 00:02:49.229 [289/740] Compiling C object lib/librte_cryptodev.a.p/cryptodev_rte_cryptodev.c.o 00:02:49.229 [290/740] Generating lib/ethdev.sym_chk with a custom command (wrapped by meson to capture output) 00:02:49.229 [291/740] Linking static target lib/librte_cryptodev.a 00:02:49.229 [292/740] Linking target lib/librte_ethdev.so.23.0 00:02:49.229 [293/740] Compiling C object lib/librte_gpudev.a.p/gpudev_gpudev.c.o 00:02:49.488 [294/740] Linking static target lib/librte_gpudev.a 00:02:49.488 [295/740] Generating symbol file lib/librte_ethdev.so.23.0.p/librte_ethdev.so.23.0.symbols 00:02:49.488 [296/740] Linking target lib/librte_metrics.so.23.0 00:02:49.488 [297/740] Compiling C object lib/librte_gro.a.p/gro_gro_tcp4.c.o 00:02:49.488 [298/740] Compiling C object lib/librte_gro.a.p/gro_rte_gro.c.o 00:02:49.488 [299/740] Linking target lib/librte_bpf.so.23.0 00:02:49.488 [300/740] Generating symbol file lib/librte_metrics.so.23.0.p/librte_metrics.so.23.0.symbols 00:02:49.488 [301/740] Compiling C object lib/librte_eventdev.a.p/eventdev_rte_event_timer_adapter.c.o 00:02:49.488 [302/740] Linking target lib/librte_bitratestats.so.23.0 00:02:49.488 [303/740] Generating lib/rte_gro_def with a custom command 00:02:49.747 [304/740] Generating lib/rte_gro_mingw with a custom command 00:02:49.747 [305/740] Compiling C object lib/librte_gro.a.p/gro_gro_udp4.c.o 00:02:49.747 [306/740] Generating symbol file lib/librte_bpf.so.23.0.p/librte_bpf.so.23.0.symbols 00:02:49.747 [307/740] Compiling C object lib/librte_gro.a.p/gro_gro_vxlan_tcp4.c.o 00:02:49.747 [308/740] Compiling C object lib/librte_eventdev.a.p/eventdev_rte_eventdev.c.o 00:02:50.006 [309/740] Compiling C object lib/librte_gso.a.p/gso_gso_udp4.c.o 00:02:50.006 [310/740] Compiling C object lib/librte_gso.a.p/gso_gso_tcp4.c.o 00:02:50.006 [311/740] Generating lib/rte_gso_def with a custom command 00:02:50.006 [312/740] Generating lib/gpudev.sym_chk with a custom command (wrapped by meson to capture output) 00:02:50.006 [313/740] Generating lib/rte_gso_mingw with a custom command 00:02:50.006 [314/740] Linking target lib/librte_gpudev.so.23.0 00:02:50.006 [315/740] Compiling C object lib/librte_gso.a.p/gso_gso_tunnel_udp4.c.o 00:02:50.006 [316/740] Compiling C object lib/librte_gro.a.p/gro_gro_vxlan_udp4.c.o 00:02:50.006 [317/740] Compiling C object lib/librte_gso.a.p/gso_gso_tunnel_tcp4.c.o 00:02:50.006 [318/740] Compiling C object lib/librte_gso.a.p/gso_rte_gso.c.o 00:02:50.006 [319/740] Linking static target lib/librte_gro.a 00:02:50.264 [320/740] Compiling C object lib/librte_eventdev.a.p/eventdev_rte_event_eth_rx_adapter.c.o 00:02:50.264 [321/740] Linking static target lib/librte_eventdev.a 00:02:50.264 [322/740] Compiling C object lib/librte_gso.a.p/gso_gso_common.c.o 00:02:50.264 [323/740] Linking static target lib/librte_gso.a 00:02:50.264 [324/740] Generating lib/gro.sym_chk with a custom command (wrapped by meson to capture output) 00:02:50.264 [325/740] Linking target lib/librte_gro.so.23.0 00:02:50.264 [326/740] Generating lib/rte_ip_frag_def with a custom command 00:02:50.264 [327/740] Generating lib/rte_ip_frag_mingw with a custom command 00:02:50.264 [328/740] Generating lib/gso.sym_chk with a custom command (wrapped by meson to capture output) 00:02:50.523 [329/740] Linking target lib/librte_gso.so.23.0 00:02:50.523 [330/740] Generating lib/rte_jobstats_def with a custom command 00:02:50.523 [331/740] Compiling C object lib/librte_ip_frag.a.p/ip_frag_rte_ipv6_reassembly.c.o 00:02:50.523 [332/740] Compiling C object lib/librte_ip_frag.a.p/ip_frag_rte_ipv4_reassembly.c.o 00:02:50.523 [333/740] Generating lib/rte_jobstats_mingw with a custom command 00:02:50.523 [334/740] Generating lib/rte_latencystats_def with a custom command 00:02:50.523 [335/740] Compiling C object lib/librte_jobstats.a.p/jobstats_rte_jobstats.c.o 00:02:50.523 [336/740] Generating lib/rte_latencystats_mingw with a custom command 00:02:50.523 [337/740] Linking static target lib/librte_jobstats.a 00:02:50.523 [338/740] Compiling C object lib/librte_ip_frag.a.p/ip_frag_rte_ipv6_fragmentation.c.o 00:02:50.523 [339/740] Compiling C object lib/librte_ip_frag.a.p/ip_frag_rte_ip_frag_common.c.o 00:02:50.523 [340/740] Generating lib/rte_lpm_def with a custom command 00:02:50.523 [341/740] Compiling C object lib/librte_ip_frag.a.p/ip_frag_ip_frag_internal.c.o 00:02:50.523 [342/740] Generating lib/rte_lpm_mingw with a custom command 00:02:50.782 [343/740] Compiling C object lib/librte_ip_frag.a.p/ip_frag_rte_ipv4_fragmentation.c.o 00:02:50.782 [344/740] Linking static target lib/librte_ip_frag.a 00:02:50.782 [345/740] Generating lib/jobstats.sym_chk with a custom command (wrapped by meson to capture output) 00:02:50.782 [346/740] Linking target lib/librte_jobstats.so.23.0 00:02:50.782 [347/740] Generating lib/ip_frag.sym_chk with a custom command (wrapped by meson to capture output) 00:02:51.041 [348/740] Linking target lib/librte_ip_frag.so.23.0 00:02:51.041 [349/740] Compiling C object lib/librte_latencystats.a.p/latencystats_rte_latencystats.c.o 00:02:51.041 [350/740] Linking static target lib/librte_latencystats.a 00:02:51.041 [351/740] Compiling C object lib/librte_member.a.p/member_rte_member.c.o 00:02:51.041 [352/740] Compiling C object lib/librte_lpm.a.p/lpm_rte_lpm.c.o 00:02:51.041 [353/740] Compiling C object lib/member/libsketch_avx512_tmp.a.p/rte_member_sketch_avx512.c.o 00:02:51.041 [354/740] Linking static target lib/member/libsketch_avx512_tmp.a 00:02:51.041 [355/740] Generating lib/rte_member_def with a custom command 00:02:51.041 [356/740] Generating lib/rte_member_mingw with a custom command 00:02:51.041 [357/740] Generating symbol file lib/librte_ip_frag.so.23.0.p/librte_ip_frag.so.23.0.symbols 00:02:51.041 [358/740] Generating lib/rte_pcapng_def with a custom command 00:02:51.041 [359/740] Generating lib/rte_pcapng_mingw with a custom command 00:02:51.041 [360/740] Generating lib/cryptodev.sym_chk with a custom command (wrapped by meson to capture output) 00:02:51.041 [361/740] Generating lib/latencystats.sym_chk with a custom command (wrapped by meson to capture output) 00:02:51.041 [362/740] Linking target lib/librte_cryptodev.so.23.0 00:02:51.041 [363/740] Linking target lib/librte_latencystats.so.23.0 00:02:51.041 [364/740] Compiling C object lib/librte_power.a.p/power_guest_channel.c.o 00:02:51.300 [365/740] Compiling C object lib/librte_power.a.p/power_power_common.c.o 00:02:51.300 [366/740] Generating symbol file lib/librte_cryptodev.so.23.0.p/librte_cryptodev.so.23.0.symbols 00:02:51.300 [367/740] Compiling C object lib/librte_power.a.p/power_power_kvm_vm.c.o 00:02:51.300 [368/740] Compiling C object lib/librte_power.a.p/power_rte_power.c.o 00:02:51.300 [369/740] Compiling C object lib/librte_lpm.a.p/lpm_rte_lpm6.c.o 00:02:51.300 [370/740] Linking static target lib/librte_lpm.a 00:02:51.559 [371/740] Compiling C object lib/librte_member.a.p/member_rte_member_vbf.c.o 00:02:51.559 [372/740] Compiling C object lib/librte_member.a.p/member_rte_member_ht.c.o 00:02:51.559 [373/740] Compiling C object lib/librte_power.a.p/power_rte_power_empty_poll.c.o 00:02:51.559 [374/740] Generating lib/rte_power_def with a custom command 00:02:51.559 [375/740] Compiling C object lib/librte_power.a.p/power_power_acpi_cpufreq.c.o 00:02:51.559 [376/740] Generating lib/rte_power_mingw with a custom command 00:02:51.559 [377/740] Generating lib/rte_rawdev_def with a custom command 00:02:51.559 [378/740] Generating lib/rte_rawdev_mingw with a custom command 00:02:51.559 [379/740] Compiling C object lib/librte_power.a.p/power_power_cppc_cpufreq.c.o 00:02:51.817 [380/740] Generating lib/rte_regexdev_def with a custom command 00:02:51.817 [381/740] Generating lib/lpm.sym_chk with a custom command (wrapped by meson to capture output) 00:02:51.817 [382/740] Generating lib/rte_regexdev_mingw with a custom command 00:02:51.817 [383/740] Compiling C object lib/librte_pcapng.a.p/pcapng_rte_pcapng.c.o 00:02:51.817 [384/740] Linking static target lib/librte_pcapng.a 00:02:51.817 [385/740] Linking target lib/librte_lpm.so.23.0 00:02:51.817 [386/740] Compiling C object lib/librte_power.a.p/power_power_pstate_cpufreq.c.o 00:02:51.817 [387/740] Generating lib/rte_dmadev_def with a custom command 00:02:51.817 [388/740] Generating lib/rte_dmadev_mingw with a custom command 00:02:51.817 [389/740] Generating lib/eventdev.sym_chk with a custom command (wrapped by meson to capture output) 00:02:51.817 [390/740] Generating symbol file lib/librte_lpm.so.23.0.p/librte_lpm.so.23.0.symbols 00:02:51.817 [391/740] Compiling C object lib/librte_power.a.p/power_rte_power_intel_uncore.c.o 00:02:51.817 [392/740] Linking target lib/librte_eventdev.so.23.0 00:02:51.817 [393/740] Generating lib/rte_rib_def with a custom command 00:02:51.817 [394/740] Compiling C object lib/librte_rawdev.a.p/rawdev_rte_rawdev.c.o 00:02:51.817 [395/740] Linking static target lib/librte_rawdev.a 00:02:51.817 [396/740] Generating lib/rte_rib_mingw with a custom command 00:02:52.077 [397/740] Generating lib/pcapng.sym_chk with a custom command (wrapped by meson to capture output) 00:02:52.077 [398/740] Generating symbol file lib/librte_eventdev.so.23.0.p/librte_eventdev.so.23.0.symbols 00:02:52.077 [399/740] Linking target lib/librte_pcapng.so.23.0 00:02:52.077 [400/740] Generating lib/rte_reorder_def with a custom command 00:02:52.077 [401/740] Compiling C object lib/librte_power.a.p/power_rte_power_pmd_mgmt.c.o 00:02:52.077 [402/740] Linking static target lib/librte_power.a 00:02:52.077 [403/740] Generating lib/rte_reorder_mingw with a custom command 00:02:52.077 [404/740] Compiling C object lib/librte_dmadev.a.p/dmadev_rte_dmadev.c.o 00:02:52.077 [405/740] Linking static target lib/librte_dmadev.a 00:02:52.077 [406/740] Generating symbol file lib/librte_pcapng.so.23.0.p/librte_pcapng.so.23.0.symbols 00:02:52.077 [407/740] Compiling C object lib/librte_regexdev.a.p/regexdev_rte_regexdev.c.o 00:02:52.077 [408/740] Linking static target lib/librte_regexdev.a 00:02:52.336 [409/740] Compiling C object lib/librte_sched.a.p/sched_rte_red.c.o 00:02:52.336 [410/740] Generating lib/rawdev.sym_chk with a custom command (wrapped by meson to capture output) 00:02:52.336 [411/740] Compiling C object lib/librte_sched.a.p/sched_rte_approx.c.o 00:02:52.336 [412/740] Linking target lib/librte_rawdev.so.23.0 00:02:52.336 [413/740] Generating lib/rte_sched_def with a custom command 00:02:52.336 [414/740] Generating lib/rte_sched_mingw with a custom command 00:02:52.336 [415/740] Compiling C object lib/librte_rib.a.p/rib_rte_rib.c.o 00:02:52.336 [416/740] Compiling C object lib/librte_sched.a.p/sched_rte_pie.c.o 00:02:52.336 [417/740] Generating lib/rte_security_def with a custom command 00:02:52.336 [418/740] Generating lib/rte_security_mingw with a custom command 00:02:52.336 [419/740] Compiling C object lib/librte_member.a.p/member_rte_member_sketch.c.o 00:02:52.336 [420/740] Linking static target lib/librte_member.a 00:02:52.336 [421/740] Compiling C object lib/librte_reorder.a.p/reorder_rte_reorder.c.o 00:02:52.336 [422/740] Linking static target lib/librte_reorder.a 00:02:52.336 [423/740] Generating lib/dmadev.sym_chk with a custom command (wrapped by meson to capture output) 00:02:52.596 [424/740] Compiling C object lib/librte_stack.a.p/stack_rte_stack_std.c.o 00:02:52.596 [425/740] Compiling C object lib/librte_stack.a.p/stack_rte_stack.c.o 00:02:52.596 [426/740] Linking target lib/librte_dmadev.so.23.0 00:02:52.596 [427/740] Generating lib/rte_stack_def with a custom command 00:02:52.596 [428/740] Generating lib/rte_stack_mingw with a custom command 00:02:52.596 [429/740] Compiling C object lib/librte_stack.a.p/stack_rte_stack_lf.c.o 00:02:52.596 [430/740] Linking static target lib/librte_stack.a 00:02:52.596 [431/740] Compiling C object lib/librte_rib.a.p/rib_rte_rib6.c.o 00:02:52.596 [432/740] Linking static target lib/librte_rib.a 00:02:52.596 [433/740] Generating symbol file lib/librte_dmadev.so.23.0.p/librte_dmadev.so.23.0.symbols 00:02:52.596 [434/740] Generating lib/reorder.sym_chk with a custom command (wrapped by meson to capture output) 00:02:52.596 [435/740] Compiling C object lib/librte_vhost.a.p/vhost_fd_man.c.o 00:02:52.596 [436/740] Generating lib/regexdev.sym_chk with a custom command (wrapped by meson to capture output) 00:02:52.596 [437/740] Linking target lib/librte_reorder.so.23.0 00:02:52.596 [438/740] Generating lib/member.sym_chk with a custom command (wrapped by meson to capture output) 00:02:52.596 [439/740] Generating lib/stack.sym_chk with a custom command (wrapped by meson to capture output) 00:02:52.596 [440/740] Linking target lib/librte_regexdev.so.23.0 00:02:52.855 [441/740] Linking target lib/librte_member.so.23.0 00:02:52.855 [442/740] Linking target lib/librte_stack.so.23.0 00:02:52.855 [443/740] Generating lib/power.sym_chk with a custom command (wrapped by meson to capture output) 00:02:52.855 [444/740] Linking target lib/librte_power.so.23.0 00:02:52.855 [445/740] Compiling C object lib/librte_security.a.p/security_rte_security.c.o 00:02:52.855 [446/740] Linking static target lib/librte_security.a 00:02:52.855 [447/740] Generating lib/rib.sym_chk with a custom command (wrapped by meson to capture output) 00:02:52.855 [448/740] Linking target lib/librte_rib.so.23.0 00:02:53.114 [449/740] Generating symbol file lib/librte_rib.so.23.0.p/librte_rib.so.23.0.symbols 00:02:53.114 [450/740] Generating lib/rte_vhost_def with a custom command 00:02:53.114 [451/740] Compiling C object lib/librte_vhost.a.p/vhost_vdpa.c.o 00:02:53.114 [452/740] Generating lib/rte_vhost_mingw with a custom command 00:02:53.114 [453/740] Compiling C object lib/librte_vhost.a.p/vhost_iotlb.c.o 00:02:53.114 [454/740] Generating lib/security.sym_chk with a custom command (wrapped by meson to capture output) 00:02:53.373 [455/740] Linking target lib/librte_security.so.23.0 00:02:53.373 [456/740] Compiling C object lib/librte_vhost.a.p/vhost_socket.c.o 00:02:53.373 [457/740] Generating symbol file lib/librte_security.so.23.0.p/librte_security.so.23.0.symbols 00:02:53.373 [458/740] Compiling C object lib/librte_sched.a.p/sched_rte_sched.c.o 00:02:53.373 [459/740] Linking static target lib/librte_sched.a 00:02:53.632 [460/740] Compiling C object lib/librte_ipsec.a.p/ipsec_ses.c.o 00:02:53.632 [461/740] Compiling C object lib/librte_ipsec.a.p/ipsec_sa.c.o 00:02:53.632 [462/740] Generating lib/rte_ipsec_def with a custom command 00:02:53.632 [463/740] Generating lib/rte_ipsec_mingw with a custom command 00:02:53.632 [464/740] Compiling C object lib/librte_vhost.a.p/vhost_vhost_user.c.o 00:02:53.891 [465/740] Generating lib/sched.sym_chk with a custom command (wrapped by meson to capture output) 00:02:53.891 [466/740] Compiling C object lib/librte_vhost.a.p/vhost_vhost.c.o 00:02:53.891 [467/740] Linking target lib/librte_sched.so.23.0 00:02:53.891 [468/740] Compiling C object lib/librte_fib.a.p/fib_rte_fib.c.o 00:02:53.891 [469/740] Generating symbol file lib/librte_sched.so.23.0.p/librte_sched.so.23.0.symbols 00:02:53.891 [470/740] Compiling C object lib/librte_ipsec.a.p/ipsec_ipsec_sad.c.o 00:02:53.891 [471/740] Compiling C object lib/librte_ipsec.a.p/ipsec_ipsec_telemetry.c.o 00:02:54.150 [472/740] Generating lib/rte_fib_def with a custom command 00:02:54.150 [473/740] Generating lib/rte_fib_mingw with a custom command 00:02:54.150 [474/740] Compiling C object lib/librte_fib.a.p/fib_rte_fib6.c.o 00:02:54.409 [475/740] Compiling C object lib/librte_fib.a.p/fib_dir24_8_avx512.c.o 00:02:54.409 [476/740] Compiling C object lib/librte_fib.a.p/fib_trie_avx512.c.o 00:02:54.409 [477/740] Compiling C object lib/librte_ipsec.a.p/ipsec_esp_inb.c.o 00:02:54.409 [478/740] Compiling C object lib/librte_ipsec.a.p/ipsec_esp_outb.c.o 00:02:54.409 [479/740] Linking static target lib/librte_ipsec.a 00:02:54.409 [480/740] Compiling C object lib/librte_fib.a.p/fib_trie.c.o 00:02:54.668 [481/740] Compiling C object lib/librte_fib.a.p/fib_dir24_8.c.o 00:02:54.668 [482/740] Linking static target lib/librte_fib.a 00:02:54.668 [483/740] Compiling C object lib/librte_port.a.p/port_rte_port_fd.c.o 00:02:54.668 [484/740] Compiling C object lib/librte_port.a.p/port_rte_port_ethdev.c.o 00:02:54.668 [485/740] Generating lib/ipsec.sym_chk with a custom command (wrapped by meson to capture output) 00:02:54.668 [486/740] Linking target lib/librte_ipsec.so.23.0 00:02:54.927 [487/740] Compiling C object lib/librte_port.a.p/port_rte_port_sched.c.o 00:02:54.927 [488/740] Compiling C object lib/librte_port.a.p/port_rte_port_frag.c.o 00:02:54.927 [489/740] Generating lib/fib.sym_chk with a custom command (wrapped by meson to capture output) 00:02:54.927 [490/740] Linking target lib/librte_fib.so.23.0 00:02:54.927 [491/740] Compiling C object lib/librte_port.a.p/port_rte_port_ras.c.o 00:02:55.494 [492/740] Compiling C object lib/librte_port.a.p/port_rte_swx_port_ethdev.c.o 00:02:55.494 [493/740] Compiling C object lib/librte_port.a.p/port_rte_port_sym_crypto.c.o 00:02:55.494 [494/740] Generating lib/rte_port_def with a custom command 00:02:55.494 [495/740] Generating lib/rte_port_mingw with a custom command 00:02:55.494 [496/740] Compiling C object lib/librte_port.a.p/port_rte_port_source_sink.c.o 00:02:55.494 [497/740] Generating lib/rte_pdump_def with a custom command 00:02:55.494 [498/740] Compiling C object lib/librte_port.a.p/port_rte_swx_port_fd.c.o 00:02:55.494 [499/740] Generating lib/rte_pdump_mingw with a custom command 00:02:55.494 [500/740] Compiling C object lib/librte_port.a.p/port_rte_port_eventdev.c.o 00:02:55.494 [501/740] Compiling C object lib/librte_table.a.p/table_rte_swx_keycmp.c.o 00:02:55.752 [502/740] Compiling C object lib/librte_port.a.p/port_rte_swx_port_source_sink.c.o 00:02:55.752 [503/740] Compiling C object lib/librte_table.a.p/table_rte_swx_table_learner.c.o 00:02:55.752 [504/740] Compiling C object lib/librte_port.a.p/port_rte_swx_port_ring.c.o 00:02:55.752 [505/740] Compiling C object lib/librte_table.a.p/table_rte_swx_table_em.c.o 00:02:56.010 [506/740] Compiling C object lib/librte_port.a.p/port_rte_port_ring.c.o 00:02:56.010 [507/740] Compiling C object lib/librte_table.a.p/table_rte_swx_table_selector.c.o 00:02:56.010 [508/740] Linking static target lib/librte_port.a 00:02:56.010 [509/740] Compiling C object lib/librte_table.a.p/table_rte_swx_table_wm.c.o 00:02:56.010 [510/740] Compiling C object lib/librte_table.a.p/table_rte_table_array.c.o 00:02:56.010 [511/740] Compiling C object lib/librte_table.a.p/table_rte_table_hash_cuckoo.c.o 00:02:56.268 [512/740] Compiling C object lib/librte_table.a.p/table_rte_table_acl.c.o 00:02:56.268 [513/740] Compiling C object lib/librte_pdump.a.p/pdump_rte_pdump.c.o 00:02:56.268 [514/740] Linking static target lib/librte_pdump.a 00:02:56.268 [515/740] Generating lib/port.sym_chk with a custom command (wrapped by meson to capture output) 00:02:56.528 [516/740] Linking target lib/librte_port.so.23.0 00:02:56.528 [517/740] Generating lib/pdump.sym_chk with a custom command (wrapped by meson to capture output) 00:02:56.528 [518/740] Generating symbol file lib/librte_port.so.23.0.p/librte_port.so.23.0.symbols 00:02:56.528 [519/740] Linking target lib/librte_pdump.so.23.0 00:02:56.528 [520/740] Compiling C object lib/librte_table.a.p/table_rte_table_hash_ext.c.o 00:02:56.528 [521/740] Compiling C object lib/librte_table.a.p/table_rte_table_lpm.c.o 00:02:56.528 [522/740] Generating lib/rte_table_def with a custom command 00:02:56.528 [523/740] Generating lib/rte_table_mingw with a custom command 00:02:56.528 [524/740] Compiling C object lib/librte_table.a.p/table_rte_table_hash_key8.c.o 00:02:56.787 [525/740] Compiling C object lib/librte_table.a.p/table_rte_table_hash_key16.c.o 00:02:56.787 [526/740] Compiling C object lib/librte_table.a.p/table_rte_table_stub.c.o 00:02:56.787 [527/740] Compiling C object lib/librte_table.a.p/table_rte_table_lpm_ipv6.c.o 00:02:57.045 [528/740] Compiling C object lib/librte_table.a.p/table_rte_table_hash_lru.c.o 00:02:57.045 [529/740] Compiling C object lib/librte_pipeline.a.p/pipeline_rte_port_in_action.c.o 00:02:57.045 [530/740] Generating lib/rte_pipeline_def with a custom command 00:02:57.045 [531/740] Compiling C object lib/librte_table.a.p/table_rte_table_hash_key32.c.o 00:02:57.045 [532/740] Linking static target lib/librte_table.a 00:02:57.045 [533/740] Generating lib/rte_pipeline_mingw with a custom command 00:02:57.303 [534/740] Compiling C object lib/librte_pipeline.a.p/pipeline_rte_pipeline.c.o 00:02:57.303 [535/740] Compiling C object lib/librte_graph.a.p/graph_node.c.o 00:02:57.561 [536/740] Compiling C object lib/librte_graph.a.p/graph_graph.c.o 00:02:57.561 [537/740] Compiling C object lib/librte_graph.a.p/graph_graph_ops.c.o 00:02:57.561 [538/740] Compiling C object lib/librte_vhost.a.p/vhost_vhost_crypto.c.o 00:02:57.561 [539/740] Generating lib/rte_graph_def with a custom command 00:02:57.561 [540/740] Generating lib/table.sym_chk with a custom command (wrapped by meson to capture output) 00:02:57.561 [541/740] Generating lib/rte_graph_mingw with a custom command 00:02:57.561 [542/740] Linking target lib/librte_table.so.23.0 00:02:57.819 [543/740] Compiling C object lib/librte_graph.a.p/graph_graph_debug.c.o 00:02:57.819 [544/740] Generating symbol file lib/librte_table.so.23.0.p/librte_table.so.23.0.symbols 00:02:57.819 [545/740] Compiling C object lib/librte_pipeline.a.p/pipeline_rte_swx_ctl.c.o 00:02:57.819 [546/740] Compiling C object lib/librte_graph.a.p/graph_graph_stats.c.o 00:02:58.078 [547/740] Compiling C object lib/librte_graph.a.p/graph_graph_populate.c.o 00:02:58.078 [548/740] Linking static target lib/librte_graph.a 00:02:58.078 [549/740] Compiling C object lib/librte_node.a.p/node_ethdev_ctrl.c.o 00:02:58.346 [550/740] Compiling C object lib/librte_node.a.p/node_ethdev_tx.c.o 00:02:58.346 [551/740] Compiling C object lib/librte_node.a.p/node_ethdev_rx.c.o 00:02:58.346 [552/740] Compiling C object lib/librte_node.a.p/node_null.c.o 00:02:58.346 [553/740] Compiling C object lib/librte_node.a.p/node_log.c.o 00:02:58.346 [554/740] Generating lib/rte_node_def with a custom command 00:02:58.346 [555/740] Generating lib/rte_node_mingw with a custom command 00:02:58.628 [556/740] Generating lib/graph.sym_chk with a custom command (wrapped by meson to capture output) 00:02:58.628 [557/740] Compiling C object lib/librte_pipeline.a.p/pipeline_rte_swx_pipeline_spec.c.o 00:02:58.628 [558/740] Linking target lib/librte_graph.so.23.0 00:02:58.628 [559/740] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_pci_common.c.o 00:02:58.628 [560/740] Compiling C object lib/librte_node.a.p/node_pkt_drop.c.o 00:02:58.628 [561/740] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_pci_params.c.o 00:02:58.628 [562/740] Compiling C object lib/librte_node.a.p/node_ip4_lookup.c.o 00:02:58.628 [563/740] Generating symbol file lib/librte_graph.so.23.0.p/librte_graph.so.23.0.symbols 00:02:58.628 [564/740] Generating drivers/rte_bus_pci_def with a custom command 00:02:58.628 [565/740] Generating drivers/rte_bus_pci_mingw with a custom command 00:02:58.628 [566/740] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_pci_common_uio.c.o 00:02:58.886 [567/740] Compiling C object drivers/libtmp_rte_bus_vdev.a.p/bus_vdev_vdev_params.c.o 00:02:58.886 [568/740] Generating drivers/rte_bus_vdev_def with a custom command 00:02:58.886 [569/740] Generating drivers/rte_bus_vdev_mingw with a custom command 00:02:58.886 [570/740] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_linux_pci_uio.c.o 00:02:58.886 [571/740] Compiling C object lib/librte_node.a.p/node_pkt_cls.c.o 00:02:58.886 [572/740] Generating drivers/rte_mempool_ring_def with a custom command 00:02:58.886 [573/740] Generating drivers/rte_mempool_ring_mingw with a custom command 00:02:58.886 [574/740] Compiling C object lib/librte_node.a.p/node_ip4_rewrite.c.o 00:02:58.886 [575/740] Linking static target lib/librte_node.a 00:02:58.886 [576/740] Compiling C object drivers/libtmp_rte_bus_vdev.a.p/bus_vdev_vdev.c.o 00:02:58.886 [577/740] Linking static target drivers/libtmp_rte_bus_vdev.a 00:02:58.886 [578/740] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_linux_pci_vfio.c.o 00:02:59.145 [579/740] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_linux_pci.c.o 00:02:59.145 [580/740] Linking static target drivers/libtmp_rte_bus_pci.a 00:02:59.145 [581/740] Generating lib/node.sym_chk with a custom command (wrapped by meson to capture output) 00:02:59.145 [582/740] Generating drivers/rte_bus_vdev.pmd.c with a custom command 00:02:59.145 [583/740] Compiling C object drivers/librte_bus_vdev.a.p/meson-generated_.._rte_bus_vdev.pmd.c.o 00:02:59.145 [584/740] Linking static target drivers/librte_bus_vdev.a 00:02:59.145 [585/740] Linking target lib/librte_node.so.23.0 00:02:59.145 [586/740] Compiling C object drivers/librte_bus_vdev.so.23.0.p/meson-generated_.._rte_bus_vdev.pmd.c.o 00:02:59.145 [587/740] Generating drivers/rte_bus_pci.pmd.c with a custom command 00:02:59.145 [588/740] Compiling C object drivers/librte_bus_pci.a.p/meson-generated_.._rte_bus_pci.pmd.c.o 00:02:59.145 [589/740] Linking static target drivers/librte_bus_pci.a 00:02:59.403 [590/740] Generating drivers/rte_bus_vdev.sym_chk with a custom command (wrapped by meson to capture output) 00:02:59.403 [591/740] Compiling C object drivers/librte_bus_pci.so.23.0.p/meson-generated_.._rte_bus_pci.pmd.c.o 00:02:59.403 [592/740] Linking target drivers/librte_bus_vdev.so.23.0 00:02:59.403 [593/740] Generating symbol file drivers/librte_bus_vdev.so.23.0.p/librte_bus_vdev.so.23.0.symbols 00:02:59.403 [594/740] Compiling C object drivers/net/i40e/base/libi40e_base.a.p/i40e_dcb.c.o 00:02:59.403 [595/740] Compiling C object drivers/net/i40e/base/libi40e_base.a.p/i40e_diag.c.o 00:02:59.661 [596/740] Compiling C object drivers/net/i40e/base/libi40e_base.a.p/i40e_adminq.c.o 00:02:59.661 [597/740] Generating drivers/rte_bus_pci.sym_chk with a custom command (wrapped by meson to capture output) 00:02:59.661 [598/740] Linking target drivers/librte_bus_pci.so.23.0 00:02:59.661 [599/740] Compiling C object drivers/libtmp_rte_mempool_ring.a.p/mempool_ring_rte_mempool_ring.c.o 00:02:59.661 [600/740] Linking static target drivers/libtmp_rte_mempool_ring.a 00:02:59.661 [601/740] Generating symbol file drivers/librte_bus_pci.so.23.0.p/librte_bus_pci.so.23.0.symbols 00:02:59.920 [602/740] Generating drivers/rte_mempool_ring.pmd.c with a custom command 00:02:59.920 [603/740] Compiling C object drivers/librte_mempool_ring.a.p/meson-generated_.._rte_mempool_ring.pmd.c.o 00:02:59.920 [604/740] Linking static target drivers/librte_mempool_ring.a 00:02:59.920 [605/740] Compiling C object drivers/librte_mempool_ring.so.23.0.p/meson-generated_.._rte_mempool_ring.pmd.c.o 00:02:59.920 [606/740] Linking target drivers/librte_mempool_ring.so.23.0 00:02:59.920 [607/740] Compiling C object drivers/net/i40e/base/libi40e_base.a.p/i40e_hmc.c.o 00:03:00.179 [608/740] Compiling C object drivers/net/i40e/base/libi40e_base.a.p/i40e_lan_hmc.c.o 00:03:00.179 [609/740] Compiling C object drivers/net/i40e/base/libi40e_base.a.p/i40e_nvm.c.o 00:03:00.747 [610/740] Compiling C object drivers/net/i40e/base/libi40e_base.a.p/i40e_common.c.o 00:03:00.747 [611/740] Linking static target drivers/net/i40e/base/libi40e_base.a 00:03:00.747 [612/740] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_i40e_pf.c.o 00:03:01.005 [613/740] Compiling C object drivers/net/i40e/libi40e_avx512_lib.a.p/i40e_rxtx_vec_avx512.c.o 00:03:01.005 [614/740] Linking static target drivers/net/i40e/libi40e_avx512_lib.a 00:03:01.005 [615/740] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_i40e_fdir.c.o 00:03:01.264 [616/740] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_i40e_tm.c.o 00:03:01.264 [617/740] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_i40e_flow.c.o 00:03:01.523 [618/740] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_i40e_hash.c.o 00:03:01.523 [619/740] Generating drivers/rte_net_i40e_def with a custom command 00:03:01.523 [620/740] Generating drivers/rte_net_i40e_mingw with a custom command 00:03:01.523 [621/740] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_i40e_vf_representor.c.o 00:03:01.783 [622/740] Compiling C object lib/librte_pipeline.a.p/pipeline_rte_swx_pipeline.c.o 00:03:02.042 [623/740] Compiling C object app/dpdk-dumpcap.p/dumpcap_main.c.o 00:03:02.302 [624/740] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_rte_pmd_i40e.c.o 00:03:02.561 [625/740] Compiling C object app/dpdk-test-bbdev.p/test-bbdev_main.c.o 00:03:02.561 [626/740] Compiling C object app/dpdk-proc-info.p/proc-info_main.c.o 00:03:02.561 [627/740] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_i40e_rxtx_vec_sse.c.o 00:03:02.561 [628/740] Compiling C object app/dpdk-pdump.p/pdump_main.c.o 00:03:02.561 [629/740] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_i40e_rxtx.c.o 00:03:02.561 [630/740] Compiling C object app/dpdk-test-cmdline.p/test-cmdline_commands.c.o 00:03:02.561 [631/740] Compiling C object app/dpdk-test-cmdline.p/test-cmdline_cmdline_test.c.o 00:03:02.820 [632/740] Compiling C object app/dpdk-test-acl.p/test-acl_main.c.o 00:03:02.820 [633/740] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_i40e_rxtx_vec_avx2.c.o 00:03:03.078 [634/740] Compiling C object app/dpdk-test-compress-perf.p/test-compress-perf_comp_perf_options_parse.c.o 00:03:03.337 [635/740] Compiling C object app/dpdk-test-bbdev.p/test-bbdev_test_bbdev.c.o 00:03:03.337 [636/740] Compiling C object app/dpdk-test-bbdev.p/test-bbdev_test_bbdev_vector.c.o 00:03:03.337 [637/740] Compiling C object app/dpdk-test-compress-perf.p/test-compress-perf_comp_perf_test_common.c.o 00:03:03.337 [638/740] Compiling C object app/dpdk-test-compress-perf.p/test-compress-perf_comp_perf_test_throughput.c.o 00:03:03.596 [639/740] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_i40e_ethdev.c.o 00:03:03.596 [640/740] Linking static target drivers/libtmp_rte_net_i40e.a 00:03:03.596 [641/740] Compiling C object app/dpdk-test-compress-perf.p/test-compress-perf_comp_perf_test_cyclecount.c.o 00:03:03.855 [642/740] Compiling C object app/dpdk-test-compress-perf.p/test-compress-perf_main.c.o 00:03:03.855 [643/740] Generating drivers/rte_net_i40e.pmd.c with a custom command 00:03:03.855 [644/740] Compiling C object drivers/librte_net_i40e.a.p/meson-generated_.._rte_net_i40e.pmd.c.o 00:03:03.855 [645/740] Compiling C object app/dpdk-test-compress-perf.p/test-compress-perf_comp_perf_test_verify.c.o 00:03:03.855 [646/740] Compiling C object drivers/librte_net_i40e.so.23.0.p/meson-generated_.._rte_net_i40e.pmd.c.o 00:03:03.855 [647/740] Compiling C object app/dpdk-test-crypto-perf.p/test-crypto-perf_cperf_test_common.c.o 00:03:03.855 [648/740] Linking static target drivers/librte_net_i40e.a 00:03:04.114 [649/740] Compiling C object app/dpdk-test-crypto-perf.p/test-crypto-perf_cperf_ops.c.o 00:03:04.114 [650/740] Compiling C object app/dpdk-test-crypto-perf.p/test-crypto-perf_cperf_options_parsing.c.o 00:03:04.373 [651/740] Generating drivers/rte_net_i40e.sym_chk with a custom command (wrapped by meson to capture output) 00:03:04.373 [652/740] Compiling C object app/dpdk-test-crypto-perf.p/test-crypto-perf_cperf_test_vector_parsing.c.o 00:03:04.373 [653/740] Linking target drivers/librte_net_i40e.so.23.0 00:03:04.373 [654/740] Compiling C object app/dpdk-test-crypto-perf.p/test-crypto-perf_cperf_test_vectors.c.o 00:03:04.373 [655/740] Compiling C object app/dpdk-test-crypto-perf.p/test-crypto-perf_cperf_test_latency.c.o 00:03:04.373 [656/740] Compiling C object app/dpdk-test-crypto-perf.p/test-crypto-perf_cperf_test_throughput.c.o 00:03:04.373 [657/740] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_evt_test.c.o 00:03:04.373 [658/740] Compiling C object app/dpdk-test-crypto-perf.p/test-crypto-perf_cperf_test_pmd_cyclecount.c.o 00:03:04.632 [659/740] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_parser.c.o 00:03:04.891 [660/740] Compiling C object app/dpdk-test-crypto-perf.p/test-crypto-perf_cperf_test_verify.c.o 00:03:04.891 [661/740] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_evt_main.c.o 00:03:04.891 [662/740] Compiling C object app/dpdk-test-crypto-perf.p/test-crypto-perf_main.c.o 00:03:04.891 [663/740] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_evt_options.c.o 00:03:04.891 [664/740] Compiling C object lib/librte_vhost.a.p/vhost_virtio_net.c.o 00:03:05.149 [665/740] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_test_order_common.c.o 00:03:05.149 [666/740] Linking static target lib/librte_vhost.a 00:03:05.149 [667/740] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_test_order_atq.c.o 00:03:05.149 [668/740] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_test_order_queue.c.o 00:03:05.718 [669/740] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_test_perf_atq.c.o 00:03:05.718 [670/740] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_test_perf_queue.c.o 00:03:05.718 [671/740] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_test_pipeline_atq.c.o 00:03:05.977 [672/740] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_test_pipeline_common.c.o 00:03:05.977 [673/740] Generating lib/vhost.sym_chk with a custom command (wrapped by meson to capture output) 00:03:05.977 [674/740] Linking target lib/librte_vhost.so.23.0 00:03:05.977 [675/740] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_test_pipeline_queue.c.o 00:03:05.977 [676/740] Compiling C object app/dpdk-test-flow-perf.p/test-flow-perf_flow_gen.c.o 00:03:05.977 [677/740] Compiling C object app/dpdk-test-flow-perf.p/test-flow-perf_items_gen.c.o 00:03:05.977 [678/740] Compiling C object app/dpdk-test-fib.p/test-fib_main.c.o 00:03:06.235 [679/740] Compiling C object app/dpdk-test-flow-perf.p/test-flow-perf_actions_gen.c.o 00:03:06.235 [680/740] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_test_perf_common.c.o 00:03:06.494 [681/740] Compiling C object app/dpdk-test-gpudev.p/test-gpudev_main.c.o 00:03:06.494 [682/740] Compiling C object app/dpdk-test-pipeline.p/test-pipeline_config.c.o 00:03:06.494 [683/740] Compiling C object app/dpdk-test-pipeline.p/test-pipeline_main.c.o 00:03:06.494 [684/740] Compiling C object app/dpdk-test-pipeline.p/test-pipeline_init.c.o 00:03:06.494 [685/740] Compiling C object app/dpdk-test-pipeline.p/test-pipeline_pipeline_acl.c.o 00:03:06.752 [686/740] Compiling C object app/dpdk-test-pipeline.p/test-pipeline_pipeline_hash.c.o 00:03:06.752 [687/740] Compiling C object app/dpdk-test-pipeline.p/test-pipeline_pipeline_lpm.c.o 00:03:06.752 [688/740] Compiling C object app/dpdk-test-pipeline.p/test-pipeline_pipeline_lpm_ipv6.c.o 00:03:06.752 [689/740] Compiling C object app/dpdk-test-pipeline.p/test-pipeline_pipeline_stub.c.o 00:03:06.752 [690/740] Compiling C object app/dpdk-test-flow-perf.p/test-flow-perf_main.c.o 00:03:07.011 [691/740] Compiling C object app/dpdk-test-bbdev.p/test-bbdev_test_bbdev_perf.c.o 00:03:07.011 [692/740] Compiling C object app/dpdk-testpmd.p/test-pmd_5tswap.c.o 00:03:07.011 [693/740] Compiling C object app/dpdk-test-pipeline.p/test-pipeline_runtime.c.o 00:03:07.275 [694/740] Compiling C object app/dpdk-testpmd.p/test-pmd_cmd_flex_item.c.o 00:03:07.275 [695/740] Compiling C object app/dpdk-testpmd.p/test-pmd_cmdline_mtr.c.o 00:03:07.543 [696/740] Compiling C object app/dpdk-testpmd.p/test-pmd_cmdline_tm.c.o 00:03:07.543 [697/740] Compiling C object app/dpdk-testpmd.p/test-pmd_flowgen.c.o 00:03:07.543 [698/740] Compiling C object app/dpdk-testpmd.p/test-pmd_icmpecho.c.o 00:03:07.802 [699/740] Compiling C object app/dpdk-testpmd.p/test-pmd_ieee1588fwd.c.o 00:03:07.802 [700/740] Compiling C object app/dpdk-testpmd.p/test-pmd_iofwd.c.o 00:03:08.061 [701/740] Compiling C object app/dpdk-testpmd.p/test-pmd_macfwd.c.o 00:03:08.061 [702/740] Compiling C object app/dpdk-testpmd.p/test-pmd_csumonly.c.o 00:03:08.061 [703/740] Compiling C object app/dpdk-testpmd.p/test-pmd_macswap.c.o 00:03:08.320 [704/740] Compiling C object app/dpdk-testpmd.p/test-pmd_rxonly.c.o 00:03:08.320 [705/740] Compiling C object app/dpdk-testpmd.p/test-pmd_cmdline.c.o 00:03:08.578 [706/740] Compiling C object app/dpdk-testpmd.p/test-pmd_parameters.c.o 00:03:08.578 [707/740] Compiling C object app/dpdk-testpmd.p/test-pmd_shared_rxq_fwd.c.o 00:03:08.578 [708/740] Compiling C object app/dpdk-testpmd.p/test-pmd_bpf_cmd.c.o 00:03:08.837 [709/740] Compiling C object app/dpdk-testpmd.p/test-pmd_util.c.o 00:03:09.096 [710/740] Compiling C object app/dpdk-testpmd.p/test-pmd_txonly.c.o 00:03:09.096 [711/740] Compiling C object app/dpdk-testpmd.p/.._drivers_net_i40e_i40e_testpmd.c.o 00:03:09.096 [712/740] Compiling C object app/dpdk-testpmd.p/test-pmd_config.c.o 00:03:09.096 [713/740] Compiling C object app/dpdk-test-sad.p/test-sad_main.c.o 00:03:09.356 [714/740] Compiling C object app/dpdk-test-security-perf.p/test-security-perf_test_security_perf.c.o 00:03:09.356 [715/740] Compiling C object app/dpdk-test-regex.p/test-regex_main.c.o 00:03:09.356 [716/740] Compiling C object app/dpdk-testpmd.p/test-pmd_cmdline_flow.c.o 00:03:09.356 [717/740] Compiling C object app/dpdk-testpmd.p/test-pmd_noisy_vnf.c.o 00:03:09.617 [718/740] Compiling C object app/dpdk-testpmd.p/test-pmd_testpmd.c.o 00:03:09.877 [719/740] Compiling C object app/dpdk-test-security-perf.p/test_test_cryptodev_security_ipsec.c.o 00:03:10.816 [720/740] Compiling C object lib/librte_pipeline.a.p/pipeline_rte_table_action.c.o 00:03:10.816 [721/740] Linking static target lib/librte_pipeline.a 00:03:11.075 [722/740] Linking target app/dpdk-test-acl 00:03:11.075 [723/740] Linking target app/dpdk-test-compress-perf 00:03:11.333 [724/740] Linking target app/dpdk-test-eventdev 00:03:11.333 [725/740] Linking target app/dpdk-dumpcap 00:03:11.333 [726/740] Linking target app/dpdk-test-cmdline 00:03:11.333 [727/740] Linking target app/dpdk-proc-info 00:03:11.333 [728/740] Linking target app/dpdk-pdump 00:03:11.333 [729/740] Linking target app/dpdk-test-crypto-perf 00:03:11.333 [730/740] Linking target app/dpdk-test-bbdev 00:03:11.593 [731/740] Linking target app/dpdk-test-flow-perf 00:03:11.593 [732/740] Linking target app/dpdk-test-fib 00:03:11.593 [733/740] Linking target app/dpdk-test-gpudev 00:03:11.593 [734/740] Linking target app/dpdk-test-security-perf 00:03:11.593 [735/740] Linking target app/dpdk-test-sad 00:03:11.593 [736/740] Linking target app/dpdk-test-pipeline 00:03:11.593 [737/740] Linking target app/dpdk-test-regex 00:03:11.593 [738/740] Linking target app/dpdk-testpmd 00:03:15.785 [739/740] Generating lib/pipeline.sym_chk with a custom command (wrapped by meson to capture output) 00:03:15.785 [740/740] Linking target lib/librte_pipeline.so.23.0 00:03:15.785 17:24:46 build_native_dpdk -- common/autobuild_common.sh@194 -- $ uname -s 00:03:15.785 17:24:46 build_native_dpdk -- common/autobuild_common.sh@194 -- $ [[ Linux == \F\r\e\e\B\S\D ]] 00:03:15.785 17:24:46 build_native_dpdk -- common/autobuild_common.sh@207 -- $ ninja -C /home/vagrant/spdk_repo/dpdk/build-tmp -j10 install 00:03:15.785 ninja: Entering directory `/home/vagrant/spdk_repo/dpdk/build-tmp' 00:03:15.785 [0/1] Installing files. 00:03:16.043 Installing subdir /home/vagrant/spdk_repo/dpdk/examples to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples 00:03:16.043 Installing /home/vagrant/spdk_repo/dpdk/examples/bbdev_app/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/bbdev_app 00:03:16.043 Installing /home/vagrant/spdk_repo/dpdk/examples/bbdev_app/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/bbdev_app 00:03:16.043 Installing /home/vagrant/spdk_repo/dpdk/examples/bond/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/bond 00:03:16.043 Installing /home/vagrant/spdk_repo/dpdk/examples/bond/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/bond 00:03:16.043 Installing /home/vagrant/spdk_repo/dpdk/examples/bond/main.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/bond 00:03:16.043 Installing /home/vagrant/spdk_repo/dpdk/examples/bpf/README to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/bpf 00:03:16.043 Installing /home/vagrant/spdk_repo/dpdk/examples/bpf/dummy.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/bpf 00:03:16.043 Installing /home/vagrant/spdk_repo/dpdk/examples/bpf/t1.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/bpf 00:03:16.043 Installing /home/vagrant/spdk_repo/dpdk/examples/bpf/t2.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/bpf 00:03:16.043 Installing /home/vagrant/spdk_repo/dpdk/examples/bpf/t3.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/bpf 00:03:16.043 Installing /home/vagrant/spdk_repo/dpdk/examples/cmdline/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/cmdline 00:03:16.043 Installing /home/vagrant/spdk_repo/dpdk/examples/cmdline/commands.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/cmdline 00:03:16.043 Installing /home/vagrant/spdk_repo/dpdk/examples/cmdline/commands.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/cmdline 00:03:16.043 Installing /home/vagrant/spdk_repo/dpdk/examples/cmdline/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/cmdline 00:03:16.043 Installing /home/vagrant/spdk_repo/dpdk/examples/cmdline/parse_obj_list.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/cmdline 00:03:16.043 Installing /home/vagrant/spdk_repo/dpdk/examples/cmdline/parse_obj_list.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/cmdline 00:03:16.043 Installing /home/vagrant/spdk_repo/dpdk/examples/common/pkt_group.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/common 00:03:16.044 Installing /home/vagrant/spdk_repo/dpdk/examples/common/altivec/port_group.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/common/altivec 00:03:16.044 Installing /home/vagrant/spdk_repo/dpdk/examples/common/neon/port_group.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/common/neon 00:03:16.044 Installing /home/vagrant/spdk_repo/dpdk/examples/common/sse/port_group.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/common/sse 00:03:16.044 Installing /home/vagrant/spdk_repo/dpdk/examples/distributor/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/distributor 00:03:16.044 Installing /home/vagrant/spdk_repo/dpdk/examples/distributor/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/distributor 00:03:16.044 Installing /home/vagrant/spdk_repo/dpdk/examples/dma/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/dma 00:03:16.044 Installing /home/vagrant/spdk_repo/dpdk/examples/dma/dmafwd.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/dma 00:03:16.044 Installing /home/vagrant/spdk_repo/dpdk/examples/ethtool/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ethtool 00:03:16.044 Installing /home/vagrant/spdk_repo/dpdk/examples/ethtool/ethtool-app/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ethtool/ethtool-app 00:03:16.044 Installing /home/vagrant/spdk_repo/dpdk/examples/ethtool/ethtool-app/ethapp.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ethtool/ethtool-app 00:03:16.044 Installing /home/vagrant/spdk_repo/dpdk/examples/ethtool/ethtool-app/ethapp.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ethtool/ethtool-app 00:03:16.044 Installing /home/vagrant/spdk_repo/dpdk/examples/ethtool/ethtool-app/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ethtool/ethtool-app 00:03:16.044 Installing /home/vagrant/spdk_repo/dpdk/examples/ethtool/lib/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ethtool/lib 00:03:16.044 Installing /home/vagrant/spdk_repo/dpdk/examples/ethtool/lib/rte_ethtool.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ethtool/lib 00:03:16.044 Installing /home/vagrant/spdk_repo/dpdk/examples/ethtool/lib/rte_ethtool.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ethtool/lib 00:03:16.044 Installing /home/vagrant/spdk_repo/dpdk/examples/eventdev_pipeline/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/eventdev_pipeline 00:03:16.044 Installing /home/vagrant/spdk_repo/dpdk/examples/eventdev_pipeline/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/eventdev_pipeline 00:03:16.044 Installing /home/vagrant/spdk_repo/dpdk/examples/eventdev_pipeline/pipeline_common.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/eventdev_pipeline 00:03:16.044 Installing /home/vagrant/spdk_repo/dpdk/examples/eventdev_pipeline/pipeline_worker_generic.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/eventdev_pipeline 00:03:16.044 Installing /home/vagrant/spdk_repo/dpdk/examples/eventdev_pipeline/pipeline_worker_tx.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/eventdev_pipeline 00:03:16.044 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:03:16.044 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_dev_self_test.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:03:16.044 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_dev_self_test.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:03:16.044 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:03:16.044 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:03:16.044 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation_aes.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:03:16.044 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation_ccm.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:03:16.044 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation_cmac.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:03:16.044 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation_ecdsa.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:03:16.044 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation_gcm.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:03:16.044 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation_hmac.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:03:16.044 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation_rsa.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:03:16.044 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation_sha.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:03:16.044 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation_tdes.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:03:16.044 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/fips_validation_xts.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:03:16.044 Installing /home/vagrant/spdk_repo/dpdk/examples/fips_validation/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/fips_validation 00:03:16.044 Installing /home/vagrant/spdk_repo/dpdk/examples/flow_classify/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/flow_classify 00:03:16.044 Installing /home/vagrant/spdk_repo/dpdk/examples/flow_classify/flow_classify.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/flow_classify 00:03:16.044 Installing /home/vagrant/spdk_repo/dpdk/examples/flow_classify/ipv4_rules_file.txt to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/flow_classify 00:03:16.044 Installing /home/vagrant/spdk_repo/dpdk/examples/flow_filtering/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/flow_filtering 00:03:16.044 Installing /home/vagrant/spdk_repo/dpdk/examples/flow_filtering/flow_blocks.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/flow_filtering 00:03:16.044 Installing /home/vagrant/spdk_repo/dpdk/examples/flow_filtering/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/flow_filtering 00:03:16.044 Installing /home/vagrant/spdk_repo/dpdk/examples/helloworld/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/helloworld 00:03:16.044 Installing /home/vagrant/spdk_repo/dpdk/examples/helloworld/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/helloworld 00:03:16.044 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_fragmentation/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_fragmentation 00:03:16.044 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_fragmentation/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_fragmentation 00:03:16.044 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:16.044 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/action.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:16.044 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/action.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:16.044 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/cli.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:16.044 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/cli.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:16.044 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/common.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:16.044 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/conn.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:16.044 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/conn.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:16.044 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/cryptodev.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:16.044 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/cryptodev.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:16.044 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/kni.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:16.044 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/kni.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:16.044 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/link.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:16.044 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/link.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:16.044 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:16.044 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/mempool.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:16.044 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/mempool.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:16.044 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/parser.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:16.044 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/parser.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:16.044 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/pipeline.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:16.044 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/pipeline.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:16.045 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/swq.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:16.045 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/swq.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:16.045 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/tap.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:16.045 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/tap.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:16.045 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/thread.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:16.045 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/thread.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:16.045 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/tmgr.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:16.045 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/tmgr.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline 00:03:16.045 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/examples/firewall.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline/examples 00:03:16.045 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/examples/flow.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline/examples 00:03:16.045 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/examples/flow_crypto.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline/examples 00:03:16.045 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/examples/kni.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline/examples 00:03:16.045 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/examples/l2fwd.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline/examples 00:03:16.045 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/examples/route.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline/examples 00:03:16.045 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/examples/route_ecmp.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline/examples 00:03:16.045 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/examples/rss.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline/examples 00:03:16.045 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_pipeline/examples/tap.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_pipeline/examples 00:03:16.045 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_reassembly/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_reassembly 00:03:16.045 Installing /home/vagrant/spdk_repo/dpdk/examples/ip_reassembly/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ip_reassembly 00:03:16.045 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:16.045 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ep0.cfg to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:16.045 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ep1.cfg to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:16.045 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/esp.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:16.045 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/esp.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:16.045 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/event_helper.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:16.045 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/event_helper.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:16.045 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/flow.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:16.045 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/flow.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:16.045 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ipip.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:16.045 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ipsec-secgw.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:16.045 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ipsec-secgw.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:16.045 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ipsec.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:16.045 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ipsec.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:16.045 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ipsec_lpm_neon.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:16.045 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ipsec_neon.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:16.045 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ipsec_process.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:16.045 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ipsec_worker.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:16.045 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/ipsec_worker.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:16.045 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/parser.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:16.045 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/parser.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:16.045 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/rt.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:16.045 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/sa.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:16.045 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/sad.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:16.045 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/sad.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:16.045 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/sp4.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:16.045 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/sp6.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw 00:03:16.045 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/bypass_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:16.045 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/common_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:16.045 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/common_defs_secgw.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:16.045 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/data_rxtx.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:16.045 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/linux_test.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:16.045 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/load_env.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:16.045 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/pkttest.py to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:16.045 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/pkttest.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:16.046 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/run_test.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:16.046 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/trs_3descbc_sha1_common_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:16.046 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/trs_3descbc_sha1_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:16.046 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/trs_aescbc_sha1_common_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:16.046 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/trs_aescbc_sha1_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:16.046 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/trs_aesctr_sha1_common_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:16.046 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/trs_aesctr_sha1_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:16.046 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/trs_aesgcm_common_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:16.046 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/trs_aesgcm_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:16.046 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/trs_ipv6opts.py to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:16.046 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/tun_3descbc_sha1_common_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:16.046 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/tun_3descbc_sha1_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:16.046 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/tun_aescbc_sha1_common_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:16.046 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/tun_aescbc_sha1_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:16.046 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/tun_aesctr_sha1_common_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:16.046 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/tun_aesctr_sha1_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:16.046 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/tun_aesgcm_common_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:16.046 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/tun_aesgcm_defs.sh to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:16.046 Installing /home/vagrant/spdk_repo/dpdk/examples/ipsec-secgw/test/tun_null_header_reconstruct.py to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:03:16.046 Installing /home/vagrant/spdk_repo/dpdk/examples/ipv4_multicast/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipv4_multicast 00:03:16.046 Installing /home/vagrant/spdk_repo/dpdk/examples/ipv4_multicast/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ipv4_multicast 00:03:16.046 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-cat/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-cat 00:03:16.046 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-cat/cat.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-cat 00:03:16.046 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-cat/cat.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-cat 00:03:16.046 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-cat/l2fwd-cat.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-cat 00:03:16.046 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-crypto/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-crypto 00:03:16.046 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-crypto/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-crypto 00:03:16.046 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-event/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-event 00:03:16.307 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-event/l2fwd_common.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-event 00:03:16.307 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-event/l2fwd_common.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-event 00:03:16.307 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-event/l2fwd_event.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-event 00:03:16.307 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-event/l2fwd_event.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-event 00:03:16.307 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-event/l2fwd_event_generic.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-event 00:03:16.307 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-event/l2fwd_event_internal_port.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-event 00:03:16.307 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-event/l2fwd_poll.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-event 00:03:16.307 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-event/l2fwd_poll.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-event 00:03:16.307 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-event/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-event 00:03:16.307 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-jobstats/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-jobstats 00:03:16.307 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-jobstats/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-jobstats 00:03:16.307 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-keepalive/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-keepalive 00:03:16.307 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-keepalive/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-keepalive 00:03:16.307 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-keepalive/shm.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-keepalive 00:03:16.307 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-keepalive/shm.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-keepalive 00:03:16.307 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-keepalive/ka-agent/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-keepalive/ka-agent 00:03:16.307 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd-keepalive/ka-agent/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd-keepalive/ka-agent 00:03:16.307 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd 00:03:16.307 Installing /home/vagrant/spdk_repo/dpdk/examples/l2fwd/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l2fwd 00:03:16.307 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd-graph/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd-graph 00:03:16.307 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd-graph/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd-graph 00:03:16.308 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd-power/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd-power 00:03:16.308 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd-power/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd-power 00:03:16.308 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd-power/main.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd-power 00:03:16.308 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd-power/perf_core.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd-power 00:03:16.308 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd-power/perf_core.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd-power 00:03:16.308 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:16.308 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/em_default_v4.cfg to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:16.308 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/em_default_v6.cfg to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:16.308 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/em_route_parse.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:16.308 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:16.308 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_acl.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:16.308 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_acl.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:16.308 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_acl_scalar.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:16.308 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_altivec.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:16.308 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_common.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:16.308 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_em.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:16.308 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_em.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:16.308 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_em_hlm.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:16.308 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_em_hlm_neon.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:16.308 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_em_hlm_sse.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:16.308 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_em_sequential.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:16.308 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_event.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:16.308 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_event.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:16.308 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_event_generic.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:16.308 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_event_internal_port.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:16.308 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_fib.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:16.308 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_lpm.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:16.308 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_lpm.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:16.308 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_lpm_altivec.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:16.308 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_lpm_neon.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:16.308 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_lpm_sse.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:16.308 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_neon.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:16.308 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_route.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:16.308 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/l3fwd_sse.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:16.308 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/lpm_default_v4.cfg to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:16.308 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/lpm_default_v6.cfg to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:16.308 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/lpm_route_parse.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:16.308 Installing /home/vagrant/spdk_repo/dpdk/examples/l3fwd/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/l3fwd 00:03:16.308 Installing /home/vagrant/spdk_repo/dpdk/examples/link_status_interrupt/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/link_status_interrupt 00:03:16.308 Installing /home/vagrant/spdk_repo/dpdk/examples/link_status_interrupt/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/link_status_interrupt 00:03:16.308 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process 00:03:16.308 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/client_server_mp/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/client_server_mp 00:03:16.308 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/client_server_mp/mp_client/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/client_server_mp/mp_client 00:03:16.308 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/client_server_mp/mp_client/client.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/client_server_mp/mp_client 00:03:16.308 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/client_server_mp/mp_server/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/client_server_mp/mp_server 00:03:16.308 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/client_server_mp/mp_server/args.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/client_server_mp/mp_server 00:03:16.308 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/client_server_mp/mp_server/args.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/client_server_mp/mp_server 00:03:16.308 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/client_server_mp/mp_server/init.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/client_server_mp/mp_server 00:03:16.308 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/client_server_mp/mp_server/init.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/client_server_mp/mp_server 00:03:16.308 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/client_server_mp/mp_server/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/client_server_mp/mp_server 00:03:16.308 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/client_server_mp/shared/common.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/client_server_mp/shared 00:03:16.308 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/hotplug_mp/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/hotplug_mp 00:03:16.308 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/hotplug_mp/commands.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/hotplug_mp 00:03:16.308 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/hotplug_mp/commands.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/hotplug_mp 00:03:16.308 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/hotplug_mp/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/hotplug_mp 00:03:16.308 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/simple_mp/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/simple_mp 00:03:16.308 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/simple_mp/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/simple_mp 00:03:16.308 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/simple_mp/mp_commands.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/simple_mp 00:03:16.308 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/simple_mp/mp_commands.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/simple_mp 00:03:16.308 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/symmetric_mp/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/symmetric_mp 00:03:16.308 Installing /home/vagrant/spdk_repo/dpdk/examples/multi_process/symmetric_mp/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/multi_process/symmetric_mp 00:03:16.308 Installing /home/vagrant/spdk_repo/dpdk/examples/ntb/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ntb 00:03:16.308 Installing /home/vagrant/spdk_repo/dpdk/examples/ntb/ntb_fwd.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ntb 00:03:16.308 Installing /home/vagrant/spdk_repo/dpdk/examples/packet_ordering/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/packet_ordering 00:03:16.308 Installing /home/vagrant/spdk_repo/dpdk/examples/packet_ordering/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/packet_ordering 00:03:16.308 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline 00:03:16.308 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/cli.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline 00:03:16.308 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/cli.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline 00:03:16.308 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/conn.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline 00:03:16.308 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/conn.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline 00:03:16.308 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline 00:03:16.308 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/obj.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline 00:03:16.308 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/obj.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline 00:03:16.308 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/thread.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline 00:03:16.308 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/thread.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline 00:03:16.308 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/ethdev.io to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:16.308 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/fib.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:16.308 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/fib.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:16.308 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/fib_nexthop_group_table.txt to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:16.308 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/fib_nexthop_table.txt to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:16.308 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/fib_routing_table.txt to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:16.308 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/hash_func.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:16.308 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/hash_func.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:16.308 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/l2fwd.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:16.308 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/l2fwd.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:16.309 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/l2fwd_macswp.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:16.309 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/l2fwd_macswp.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:16.309 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/l2fwd_macswp_pcap.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:16.309 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/l2fwd_pcap.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:16.309 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/learner.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:16.309 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/learner.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:16.309 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/meter.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:16.309 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/meter.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:16.309 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/mirroring.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:16.309 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/mirroring.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:16.309 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/packet.txt to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:16.309 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/pcap.io to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:16.309 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/recirculation.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:16.309 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/recirculation.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:16.309 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/registers.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:16.309 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/registers.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:16.309 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/selector.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:16.309 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/selector.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:16.309 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/selector.txt to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:16.309 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/varbit.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:16.309 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/varbit.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:16.309 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/vxlan.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:16.309 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/vxlan.spec to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:16.309 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/vxlan_pcap.cli to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:16.309 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/vxlan_table.py to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:16.309 Installing /home/vagrant/spdk_repo/dpdk/examples/pipeline/examples/vxlan_table.txt to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/pipeline/examples 00:03:16.309 Installing /home/vagrant/spdk_repo/dpdk/examples/ptpclient/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ptpclient 00:03:16.309 Installing /home/vagrant/spdk_repo/dpdk/examples/ptpclient/ptpclient.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/ptpclient 00:03:16.309 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_meter/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_meter 00:03:16.309 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_meter/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_meter 00:03:16.309 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_meter/main.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_meter 00:03:16.309 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_meter/rte_policer.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_meter 00:03:16.309 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_meter/rte_policer.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_meter 00:03:16.309 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:03:16.309 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/app_thread.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:03:16.309 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/args.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:03:16.309 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/cfg_file.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:03:16.309 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/cfg_file.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:03:16.309 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/cmdline.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:03:16.309 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/init.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:03:16.309 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:03:16.309 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/main.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:03:16.309 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/profile.cfg to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:03:16.309 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/profile_ov.cfg to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:03:16.309 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/profile_pie.cfg to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:03:16.309 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/profile_red.cfg to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:03:16.309 Installing /home/vagrant/spdk_repo/dpdk/examples/qos_sched/stats.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/qos_sched 00:03:16.309 Installing /home/vagrant/spdk_repo/dpdk/examples/rxtx_callbacks/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/rxtx_callbacks 00:03:16.309 Installing /home/vagrant/spdk_repo/dpdk/examples/rxtx_callbacks/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/rxtx_callbacks 00:03:16.309 Installing /home/vagrant/spdk_repo/dpdk/examples/server_node_efd/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/server_node_efd 00:03:16.309 Installing /home/vagrant/spdk_repo/dpdk/examples/server_node_efd/node/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/server_node_efd/node 00:03:16.309 Installing /home/vagrant/spdk_repo/dpdk/examples/server_node_efd/node/node.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/server_node_efd/node 00:03:16.309 Installing /home/vagrant/spdk_repo/dpdk/examples/server_node_efd/server/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/server_node_efd/server 00:03:16.309 Installing /home/vagrant/spdk_repo/dpdk/examples/server_node_efd/server/args.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/server_node_efd/server 00:03:16.309 Installing /home/vagrant/spdk_repo/dpdk/examples/server_node_efd/server/args.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/server_node_efd/server 00:03:16.309 Installing /home/vagrant/spdk_repo/dpdk/examples/server_node_efd/server/init.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/server_node_efd/server 00:03:16.309 Installing /home/vagrant/spdk_repo/dpdk/examples/server_node_efd/server/init.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/server_node_efd/server 00:03:16.309 Installing /home/vagrant/spdk_repo/dpdk/examples/server_node_efd/server/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/server_node_efd/server 00:03:16.309 Installing /home/vagrant/spdk_repo/dpdk/examples/server_node_efd/shared/common.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/server_node_efd/shared 00:03:16.309 Installing /home/vagrant/spdk_repo/dpdk/examples/service_cores/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/service_cores 00:03:16.309 Installing /home/vagrant/spdk_repo/dpdk/examples/service_cores/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/service_cores 00:03:16.309 Installing /home/vagrant/spdk_repo/dpdk/examples/skeleton/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/skeleton 00:03:16.309 Installing /home/vagrant/spdk_repo/dpdk/examples/skeleton/basicfwd.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/skeleton 00:03:16.309 Installing /home/vagrant/spdk_repo/dpdk/examples/timer/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/timer 00:03:16.309 Installing /home/vagrant/spdk_repo/dpdk/examples/timer/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/timer 00:03:16.309 Installing /home/vagrant/spdk_repo/dpdk/examples/vdpa/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vdpa 00:03:16.309 Installing /home/vagrant/spdk_repo/dpdk/examples/vdpa/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vdpa 00:03:16.309 Installing /home/vagrant/spdk_repo/dpdk/examples/vdpa/vdpa_blk_compact.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vdpa 00:03:16.309 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost 00:03:16.309 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost 00:03:16.309 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost/main.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost 00:03:16.309 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost/virtio_net.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost 00:03:16.309 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost_blk/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost_blk 00:03:16.309 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost_blk/blk.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost_blk 00:03:16.309 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost_blk/blk_spec.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost_blk 00:03:16.309 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost_blk/vhost_blk.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost_blk 00:03:16.309 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost_blk/vhost_blk.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost_blk 00:03:16.309 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost_blk/vhost_blk_compat.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost_blk 00:03:16.309 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost_crypto/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost_crypto 00:03:16.309 Installing /home/vagrant/spdk_repo/dpdk/examples/vhost_crypto/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vhost_crypto 00:03:16.309 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:03:16.309 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/channel_manager.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:03:16.309 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/channel_manager.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:03:16.309 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/channel_monitor.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:03:16.309 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/channel_monitor.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:03:16.309 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:03:16.309 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/oob_monitor.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:03:16.310 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/oob_monitor_nop.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:03:16.310 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/oob_monitor_x86.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:03:16.310 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/parse.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:03:16.310 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/parse.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:03:16.310 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/power_manager.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:03:16.310 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/power_manager.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:03:16.310 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/vm_power_cli.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:03:16.310 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/vm_power_cli.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager 00:03:16.310 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/guest_cli/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager/guest_cli 00:03:16.310 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/guest_cli/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager/guest_cli 00:03:16.310 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/guest_cli/parse.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager/guest_cli 00:03:16.310 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/guest_cli/parse.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager/guest_cli 00:03:16.310 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/guest_cli/vm_power_cli_guest.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager/guest_cli 00:03:16.310 Installing /home/vagrant/spdk_repo/dpdk/examples/vm_power_manager/guest_cli/vm_power_cli_guest.h to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vm_power_manager/guest_cli 00:03:16.310 Installing /home/vagrant/spdk_repo/dpdk/examples/vmdq/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vmdq 00:03:16.310 Installing /home/vagrant/spdk_repo/dpdk/examples/vmdq/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vmdq 00:03:16.310 Installing /home/vagrant/spdk_repo/dpdk/examples/vmdq_dcb/Makefile to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vmdq_dcb 00:03:16.310 Installing /home/vagrant/spdk_repo/dpdk/examples/vmdq_dcb/main.c to /home/vagrant/spdk_repo/dpdk/build/share/dpdk/examples/vmdq_dcb 00:03:16.310 Installing lib/librte_kvargs.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.310 Installing lib/librte_kvargs.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.310 Installing lib/librte_telemetry.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.310 Installing lib/librte_telemetry.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.310 Installing lib/librte_eal.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.310 Installing lib/librte_eal.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.310 Installing lib/librte_ring.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.310 Installing lib/librte_ring.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.310 Installing lib/librte_rcu.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.310 Installing lib/librte_rcu.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.310 Installing lib/librte_mempool.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.310 Installing lib/librte_mempool.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.310 Installing lib/librte_mbuf.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.310 Installing lib/librte_mbuf.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.310 Installing lib/librte_net.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.570 Installing lib/librte_net.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.570 Installing lib/librte_meter.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.570 Installing lib/librte_meter.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.570 Installing lib/librte_ethdev.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.570 Installing lib/librte_ethdev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.570 Installing lib/librte_pci.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.570 Installing lib/librte_pci.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.570 Installing lib/librte_cmdline.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.570 Installing lib/librte_cmdline.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.570 Installing lib/librte_metrics.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.570 Installing lib/librte_metrics.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.570 Installing lib/librte_hash.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.570 Installing lib/librte_hash.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.570 Installing lib/librte_timer.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.570 Installing lib/librte_timer.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.570 Installing lib/librte_acl.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.570 Installing lib/librte_acl.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.570 Installing lib/librte_bbdev.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.570 Installing lib/librte_bbdev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.570 Installing lib/librte_bitratestats.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.570 Installing lib/librte_bitratestats.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.570 Installing lib/librte_bpf.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.570 Installing lib/librte_bpf.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.570 Installing lib/librte_cfgfile.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.570 Installing lib/librte_cfgfile.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.570 Installing lib/librte_compressdev.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.570 Installing lib/librte_compressdev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.570 Installing lib/librte_cryptodev.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.570 Installing lib/librte_cryptodev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.570 Installing lib/librte_distributor.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.570 Installing lib/librte_distributor.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.570 Installing lib/librte_efd.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.570 Installing lib/librte_efd.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.570 Installing lib/librte_eventdev.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.570 Installing lib/librte_eventdev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.570 Installing lib/librte_gpudev.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.570 Installing lib/librte_gpudev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.570 Installing lib/librte_gro.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.570 Installing lib/librte_gro.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.570 Installing lib/librte_gso.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.570 Installing lib/librte_gso.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.570 Installing lib/librte_ip_frag.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.570 Installing lib/librte_ip_frag.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.570 Installing lib/librte_jobstats.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.570 Installing lib/librte_jobstats.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.570 Installing lib/librte_latencystats.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.570 Installing lib/librte_latencystats.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.570 Installing lib/librte_lpm.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.570 Installing lib/librte_lpm.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.570 Installing lib/librte_member.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.570 Installing lib/librte_member.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.570 Installing lib/librte_pcapng.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.570 Installing lib/librte_pcapng.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.570 Installing lib/librte_power.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.570 Installing lib/librte_power.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.570 Installing lib/librte_rawdev.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.570 Installing lib/librte_rawdev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.570 Installing lib/librte_regexdev.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.570 Installing lib/librte_regexdev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.570 Installing lib/librte_dmadev.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.570 Installing lib/librte_dmadev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.570 Installing lib/librte_rib.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.570 Installing lib/librte_rib.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.570 Installing lib/librte_reorder.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.570 Installing lib/librte_reorder.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.570 Installing lib/librte_sched.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.570 Installing lib/librte_sched.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.570 Installing lib/librte_security.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.570 Installing lib/librte_security.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.570 Installing lib/librte_stack.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.570 Installing lib/librte_stack.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.570 Installing lib/librte_vhost.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.570 Installing lib/librte_vhost.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.570 Installing lib/librte_ipsec.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.570 Installing lib/librte_ipsec.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.570 Installing lib/librte_fib.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.570 Installing lib/librte_fib.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.570 Installing lib/librte_port.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.570 Installing lib/librte_port.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.570 Installing lib/librte_pdump.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.570 Installing lib/librte_pdump.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.570 Installing lib/librte_table.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.571 Installing lib/librte_table.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.571 Installing lib/librte_pipeline.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.571 Installing lib/librte_pipeline.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.571 Installing lib/librte_graph.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.571 Installing lib/librte_graph.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.571 Installing lib/librte_node.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.571 Installing lib/librte_node.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.571 Installing drivers/librte_bus_pci.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.571 Installing drivers/librte_bus_pci.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0 00:03:16.571 Installing drivers/librte_bus_vdev.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.571 Installing drivers/librte_bus_vdev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0 00:03:16.571 Installing drivers/librte_mempool_ring.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.571 Installing drivers/librte_mempool_ring.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0 00:03:16.571 Installing drivers/librte_net_i40e.a to /home/vagrant/spdk_repo/dpdk/build/lib 00:03:16.571 Installing drivers/librte_net_i40e.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0 00:03:16.571 Installing app/dpdk-dumpcap to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:16.571 Installing app/dpdk-pdump to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:16.571 Installing app/dpdk-proc-info to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:16.571 Installing app/dpdk-test-acl to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:16.571 Installing app/dpdk-test-bbdev to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:16.571 Installing app/dpdk-test-cmdline to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:16.571 Installing app/dpdk-test-compress-perf to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:16.571 Installing app/dpdk-test-crypto-perf to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:16.571 Installing app/dpdk-test-eventdev to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:16.571 Installing app/dpdk-test-fib to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:16.571 Installing app/dpdk-test-flow-perf to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:16.571 Installing app/dpdk-test-gpudev to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:16.571 Installing app/dpdk-test-pipeline to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:16.571 Installing app/dpdk-testpmd to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:16.571 Installing app/dpdk-test-regex to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:16.571 Installing app/dpdk-test-sad to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:16.571 Installing app/dpdk-test-security-perf to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:16.571 Installing /home/vagrant/spdk_repo/dpdk/config/rte_config.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.571 Installing /home/vagrant/spdk_repo/dpdk/lib/kvargs/rte_kvargs.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.571 Installing /home/vagrant/spdk_repo/dpdk/lib/telemetry/rte_telemetry.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.571 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_atomic.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:03:16.571 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_byteorder.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:03:16.571 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_cpuflags.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:03:16.571 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_cycles.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:03:16.571 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_io.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:03:16.571 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_memcpy.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:03:16.571 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_pause.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:03:16.571 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_power_intrinsics.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:03:16.571 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_prefetch.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:03:16.571 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_rwlock.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:03:16.571 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_spinlock.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:03:16.571 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/generic/rte_vect.h to /home/vagrant/spdk_repo/dpdk/build/include/generic 00:03:16.571 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_atomic.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.571 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_byteorder.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.571 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_cpuflags.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.571 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_cycles.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.571 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_io.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.571 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_memcpy.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.571 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_pause.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.571 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_power_intrinsics.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.571 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_prefetch.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.571 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_rtm.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.571 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_rwlock.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.571 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_spinlock.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.571 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_vect.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.571 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_atomic_32.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.571 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_atomic_64.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.571 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_byteorder_32.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.571 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/x86/include/rte_byteorder_64.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.571 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_alarm.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.571 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_bitmap.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.571 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_bitops.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.571 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_branch_prediction.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.571 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_bus.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.571 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_class.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.571 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_common.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.571 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_compat.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.571 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_debug.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.571 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_dev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.571 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_devargs.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.571 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_eal.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.571 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_eal_memconfig.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.571 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_eal_trace.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.571 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_errno.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.571 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_epoll.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.571 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_fbarray.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.571 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_hexdump.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.571 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_hypervisor.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.571 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_interrupts.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.571 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_keepalive.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.571 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_launch.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.571 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_lcore.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.571 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_log.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.571 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_malloc.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.571 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_mcslock.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.571 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_memory.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.571 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_memzone.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.571 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_pci_dev_feature_defs.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.571 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_pci_dev_features.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.571 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_per_lcore.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.571 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_pflock.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.571 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_random.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.571 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_reciprocal.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.571 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_seqcount.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.571 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_seqlock.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.571 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_service.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.571 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_service_component.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.571 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_string_fns.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.571 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_tailq.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.571 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_thread.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.571 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_ticketlock.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.571 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_time.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.571 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_trace.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.571 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_trace_point.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.572 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_trace_point_register.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.572 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_uuid.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.572 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_version.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.572 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/include/rte_vfio.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.572 Installing /home/vagrant/spdk_repo/dpdk/lib/eal/linux/include/rte_os.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.572 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.572 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_core.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.572 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_elem.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.572 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_elem_pvt.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.572 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_c11_pvt.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.572 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_generic_pvt.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.572 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_hts.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.572 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_hts_elem_pvt.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.572 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_peek.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.572 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_peek_elem_pvt.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.572 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_peek_zc.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.572 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_rts.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.572 Installing /home/vagrant/spdk_repo/dpdk/lib/ring/rte_ring_rts_elem_pvt.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.572 Installing /home/vagrant/spdk_repo/dpdk/lib/rcu/rte_rcu_qsbr.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.572 Installing /home/vagrant/spdk_repo/dpdk/lib/mempool/rte_mempool.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.572 Installing /home/vagrant/spdk_repo/dpdk/lib/mempool/rte_mempool_trace.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.572 Installing /home/vagrant/spdk_repo/dpdk/lib/mempool/rte_mempool_trace_fp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.572 Installing /home/vagrant/spdk_repo/dpdk/lib/mbuf/rte_mbuf.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.572 Installing /home/vagrant/spdk_repo/dpdk/lib/mbuf/rte_mbuf_core.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.572 Installing /home/vagrant/spdk_repo/dpdk/lib/mbuf/rte_mbuf_ptype.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.572 Installing /home/vagrant/spdk_repo/dpdk/lib/mbuf/rte_mbuf_pool_ops.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.572 Installing /home/vagrant/spdk_repo/dpdk/lib/mbuf/rte_mbuf_dyn.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.572 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_ip.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.572 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_tcp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.572 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_udp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.572 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_esp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.572 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_sctp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.572 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_icmp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.572 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_arp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.572 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_ether.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.572 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_macsec.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.572 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_vxlan.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.572 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_gre.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.572 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_gtp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.572 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_net.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.572 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_net_crc.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.572 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_mpls.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.572 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_higig.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.572 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_ecpri.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.572 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_geneve.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.572 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_l2tpv2.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.572 Installing /home/vagrant/spdk_repo/dpdk/lib/net/rte_ppp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.572 Installing /home/vagrant/spdk_repo/dpdk/lib/meter/rte_meter.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.572 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_cman.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.572 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_ethdev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.572 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_ethdev_trace.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.572 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_ethdev_trace_fp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.572 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_dev_info.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.572 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_flow.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.572 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_flow_driver.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.572 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_mtr.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.572 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_mtr_driver.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.572 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_tm.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.572 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_tm_driver.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.572 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_ethdev_core.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.572 Installing /home/vagrant/spdk_repo/dpdk/lib/ethdev/rte_eth_ctrl.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.572 Installing /home/vagrant/spdk_repo/dpdk/lib/pci/rte_pci.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.572 Installing /home/vagrant/spdk_repo/dpdk/lib/cmdline/cmdline.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.572 Installing /home/vagrant/spdk_repo/dpdk/lib/cmdline/cmdline_parse.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.572 Installing /home/vagrant/spdk_repo/dpdk/lib/cmdline/cmdline_parse_num.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.572 Installing /home/vagrant/spdk_repo/dpdk/lib/cmdline/cmdline_parse_ipaddr.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.572 Installing /home/vagrant/spdk_repo/dpdk/lib/cmdline/cmdline_parse_etheraddr.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.572 Installing /home/vagrant/spdk_repo/dpdk/lib/cmdline/cmdline_parse_string.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.572 Installing /home/vagrant/spdk_repo/dpdk/lib/cmdline/cmdline_rdline.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.572 Installing /home/vagrant/spdk_repo/dpdk/lib/cmdline/cmdline_vt100.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.572 Installing /home/vagrant/spdk_repo/dpdk/lib/cmdline/cmdline_socket.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.572 Installing /home/vagrant/spdk_repo/dpdk/lib/cmdline/cmdline_cirbuf.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.572 Installing /home/vagrant/spdk_repo/dpdk/lib/cmdline/cmdline_parse_portlist.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.572 Installing /home/vagrant/spdk_repo/dpdk/lib/metrics/rte_metrics.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.572 Installing /home/vagrant/spdk_repo/dpdk/lib/metrics/rte_metrics_telemetry.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.572 Installing /home/vagrant/spdk_repo/dpdk/lib/hash/rte_fbk_hash.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.572 Installing /home/vagrant/spdk_repo/dpdk/lib/hash/rte_hash_crc.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.572 Installing /home/vagrant/spdk_repo/dpdk/lib/hash/rte_hash.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.572 Installing /home/vagrant/spdk_repo/dpdk/lib/hash/rte_jhash.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.572 Installing /home/vagrant/spdk_repo/dpdk/lib/hash/rte_thash.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.572 Installing /home/vagrant/spdk_repo/dpdk/lib/hash/rte_thash_gfni.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.572 Installing /home/vagrant/spdk_repo/dpdk/lib/hash/rte_crc_arm64.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.572 Installing /home/vagrant/spdk_repo/dpdk/lib/hash/rte_crc_generic.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.572 Installing /home/vagrant/spdk_repo/dpdk/lib/hash/rte_crc_sw.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.572 Installing /home/vagrant/spdk_repo/dpdk/lib/hash/rte_crc_x86.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.572 Installing /home/vagrant/spdk_repo/dpdk/lib/hash/rte_thash_x86_gfni.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.572 Installing /home/vagrant/spdk_repo/dpdk/lib/timer/rte_timer.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.572 Installing /home/vagrant/spdk_repo/dpdk/lib/acl/rte_acl.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.572 Installing /home/vagrant/spdk_repo/dpdk/lib/acl/rte_acl_osdep.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.572 Installing /home/vagrant/spdk_repo/dpdk/lib/bbdev/rte_bbdev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.572 Installing /home/vagrant/spdk_repo/dpdk/lib/bbdev/rte_bbdev_pmd.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.572 Installing /home/vagrant/spdk_repo/dpdk/lib/bbdev/rte_bbdev_op.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.572 Installing /home/vagrant/spdk_repo/dpdk/lib/bitratestats/rte_bitrate.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.572 Installing /home/vagrant/spdk_repo/dpdk/lib/bpf/bpf_def.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.572 Installing /home/vagrant/spdk_repo/dpdk/lib/bpf/rte_bpf.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.572 Installing /home/vagrant/spdk_repo/dpdk/lib/bpf/rte_bpf_ethdev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.572 Installing /home/vagrant/spdk_repo/dpdk/lib/cfgfile/rte_cfgfile.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.572 Installing /home/vagrant/spdk_repo/dpdk/lib/compressdev/rte_compressdev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.572 Installing /home/vagrant/spdk_repo/dpdk/lib/compressdev/rte_comp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.572 Installing /home/vagrant/spdk_repo/dpdk/lib/cryptodev/rte_cryptodev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.572 Installing /home/vagrant/spdk_repo/dpdk/lib/cryptodev/rte_cryptodev_trace.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.572 Installing /home/vagrant/spdk_repo/dpdk/lib/cryptodev/rte_cryptodev_trace_fp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.572 Installing /home/vagrant/spdk_repo/dpdk/lib/cryptodev/rte_crypto.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.572 Installing /home/vagrant/spdk_repo/dpdk/lib/cryptodev/rte_crypto_sym.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.572 Installing /home/vagrant/spdk_repo/dpdk/lib/cryptodev/rte_crypto_asym.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.572 Installing /home/vagrant/spdk_repo/dpdk/lib/cryptodev/rte_cryptodev_core.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.573 Installing /home/vagrant/spdk_repo/dpdk/lib/distributor/rte_distributor.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.573 Installing /home/vagrant/spdk_repo/dpdk/lib/efd/rte_efd.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.573 Installing /home/vagrant/spdk_repo/dpdk/lib/eventdev/rte_event_crypto_adapter.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.573 Installing /home/vagrant/spdk_repo/dpdk/lib/eventdev/rte_event_eth_rx_adapter.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.573 Installing /home/vagrant/spdk_repo/dpdk/lib/eventdev/rte_event_eth_tx_adapter.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.573 Installing /home/vagrant/spdk_repo/dpdk/lib/eventdev/rte_event_ring.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.573 Installing /home/vagrant/spdk_repo/dpdk/lib/eventdev/rte_event_timer_adapter.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.573 Installing /home/vagrant/spdk_repo/dpdk/lib/eventdev/rte_eventdev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.573 Installing /home/vagrant/spdk_repo/dpdk/lib/eventdev/rte_eventdev_trace_fp.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.573 Installing /home/vagrant/spdk_repo/dpdk/lib/eventdev/rte_eventdev_core.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.573 Installing /home/vagrant/spdk_repo/dpdk/lib/gpudev/rte_gpudev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.573 Installing /home/vagrant/spdk_repo/dpdk/lib/gro/rte_gro.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.573 Installing /home/vagrant/spdk_repo/dpdk/lib/gso/rte_gso.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.573 Installing /home/vagrant/spdk_repo/dpdk/lib/ip_frag/rte_ip_frag.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.573 Installing /home/vagrant/spdk_repo/dpdk/lib/jobstats/rte_jobstats.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.573 Installing /home/vagrant/spdk_repo/dpdk/lib/latencystats/rte_latencystats.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.573 Installing /home/vagrant/spdk_repo/dpdk/lib/lpm/rte_lpm.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.573 Installing /home/vagrant/spdk_repo/dpdk/lib/lpm/rte_lpm6.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.573 Installing /home/vagrant/spdk_repo/dpdk/lib/lpm/rte_lpm_altivec.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.573 Installing /home/vagrant/spdk_repo/dpdk/lib/lpm/rte_lpm_neon.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.573 Installing /home/vagrant/spdk_repo/dpdk/lib/lpm/rte_lpm_scalar.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.573 Installing /home/vagrant/spdk_repo/dpdk/lib/lpm/rte_lpm_sse.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.573 Installing /home/vagrant/spdk_repo/dpdk/lib/lpm/rte_lpm_sve.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.573 Installing /home/vagrant/spdk_repo/dpdk/lib/member/rte_member.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.573 Installing /home/vagrant/spdk_repo/dpdk/lib/pcapng/rte_pcapng.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.573 Installing /home/vagrant/spdk_repo/dpdk/lib/power/rte_power.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.573 Installing /home/vagrant/spdk_repo/dpdk/lib/power/rte_power_empty_poll.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.573 Installing /home/vagrant/spdk_repo/dpdk/lib/power/rte_power_intel_uncore.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.573 Installing /home/vagrant/spdk_repo/dpdk/lib/power/rte_power_pmd_mgmt.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.573 Installing /home/vagrant/spdk_repo/dpdk/lib/power/rte_power_guest_channel.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.573 Installing /home/vagrant/spdk_repo/dpdk/lib/rawdev/rte_rawdev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.573 Installing /home/vagrant/spdk_repo/dpdk/lib/rawdev/rte_rawdev_pmd.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.573 Installing /home/vagrant/spdk_repo/dpdk/lib/regexdev/rte_regexdev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.573 Installing /home/vagrant/spdk_repo/dpdk/lib/regexdev/rte_regexdev_driver.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.573 Installing /home/vagrant/spdk_repo/dpdk/lib/regexdev/rte_regexdev_core.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.573 Installing /home/vagrant/spdk_repo/dpdk/lib/dmadev/rte_dmadev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.573 Installing /home/vagrant/spdk_repo/dpdk/lib/dmadev/rte_dmadev_core.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.573 Installing /home/vagrant/spdk_repo/dpdk/lib/rib/rte_rib.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.573 Installing /home/vagrant/spdk_repo/dpdk/lib/rib/rte_rib6.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.573 Installing /home/vagrant/spdk_repo/dpdk/lib/reorder/rte_reorder.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.573 Installing /home/vagrant/spdk_repo/dpdk/lib/sched/rte_approx.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.573 Installing /home/vagrant/spdk_repo/dpdk/lib/sched/rte_red.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.573 Installing /home/vagrant/spdk_repo/dpdk/lib/sched/rte_sched.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.573 Installing /home/vagrant/spdk_repo/dpdk/lib/sched/rte_sched_common.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.573 Installing /home/vagrant/spdk_repo/dpdk/lib/sched/rte_pie.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.573 Installing /home/vagrant/spdk_repo/dpdk/lib/security/rte_security.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.573 Installing /home/vagrant/spdk_repo/dpdk/lib/security/rte_security_driver.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.573 Installing /home/vagrant/spdk_repo/dpdk/lib/stack/rte_stack.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.573 Installing /home/vagrant/spdk_repo/dpdk/lib/stack/rte_stack_std.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.573 Installing /home/vagrant/spdk_repo/dpdk/lib/stack/rte_stack_lf.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.573 Installing /home/vagrant/spdk_repo/dpdk/lib/stack/rte_stack_lf_generic.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.573 Installing /home/vagrant/spdk_repo/dpdk/lib/stack/rte_stack_lf_c11.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.573 Installing /home/vagrant/spdk_repo/dpdk/lib/stack/rte_stack_lf_stubs.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.573 Installing /home/vagrant/spdk_repo/dpdk/lib/vhost/rte_vdpa.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.573 Installing /home/vagrant/spdk_repo/dpdk/lib/vhost/rte_vhost.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.573 Installing /home/vagrant/spdk_repo/dpdk/lib/vhost/rte_vhost_async.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.573 Installing /home/vagrant/spdk_repo/dpdk/lib/vhost/rte_vhost_crypto.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.573 Installing /home/vagrant/spdk_repo/dpdk/lib/ipsec/rte_ipsec.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.573 Installing /home/vagrant/spdk_repo/dpdk/lib/ipsec/rte_ipsec_sa.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.573 Installing /home/vagrant/spdk_repo/dpdk/lib/ipsec/rte_ipsec_sad.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.573 Installing /home/vagrant/spdk_repo/dpdk/lib/ipsec/rte_ipsec_group.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.573 Installing /home/vagrant/spdk_repo/dpdk/lib/fib/rte_fib.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.573 Installing /home/vagrant/spdk_repo/dpdk/lib/fib/rte_fib6.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.573 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_port_ethdev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.573 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_port_fd.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.573 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_port_frag.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.573 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_port_ras.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.573 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_port.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.573 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_port_ring.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.573 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_port_sched.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.573 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_port_source_sink.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.573 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_port_sym_crypto.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.573 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_port_eventdev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.573 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_swx_port.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.573 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_swx_port_ethdev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.573 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_swx_port_fd.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.573 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_swx_port_ring.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.573 Installing /home/vagrant/spdk_repo/dpdk/lib/port/rte_swx_port_source_sink.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.573 Installing /home/vagrant/spdk_repo/dpdk/lib/pdump/rte_pdump.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.573 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_lru.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.573 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_swx_hash_func.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.573 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_swx_table.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.573 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_swx_table_em.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.573 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_swx_table_learner.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.573 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_swx_table_selector.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.573 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_swx_table_wm.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.573 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_table.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.573 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_table_acl.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.573 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_table_array.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.573 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_table_hash.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.573 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_table_hash_cuckoo.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.573 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_table_hash_func.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.573 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_table_lpm.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.573 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_table_lpm_ipv6.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.573 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_table_stub.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.573 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_lru_arm64.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.573 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_lru_x86.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.573 Installing /home/vagrant/spdk_repo/dpdk/lib/table/rte_table_hash_func_arm64.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.573 Installing /home/vagrant/spdk_repo/dpdk/lib/pipeline/rte_pipeline.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.573 Installing /home/vagrant/spdk_repo/dpdk/lib/pipeline/rte_port_in_action.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.573 Installing /home/vagrant/spdk_repo/dpdk/lib/pipeline/rte_table_action.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.573 Installing /home/vagrant/spdk_repo/dpdk/lib/pipeline/rte_swx_pipeline.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.573 Installing /home/vagrant/spdk_repo/dpdk/lib/pipeline/rte_swx_extern.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.573 Installing /home/vagrant/spdk_repo/dpdk/lib/pipeline/rte_swx_ctl.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.573 Installing /home/vagrant/spdk_repo/dpdk/lib/graph/rte_graph.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.573 Installing /home/vagrant/spdk_repo/dpdk/lib/graph/rte_graph_worker.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.574 Installing /home/vagrant/spdk_repo/dpdk/lib/node/rte_node_ip4_api.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.574 Installing /home/vagrant/spdk_repo/dpdk/lib/node/rte_node_eth_api.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.574 Installing /home/vagrant/spdk_repo/dpdk/drivers/bus/pci/rte_bus_pci.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.574 Installing /home/vagrant/spdk_repo/dpdk/drivers/bus/vdev/rte_bus_vdev.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.574 Installing /home/vagrant/spdk_repo/dpdk/drivers/net/i40e/rte_pmd_i40e.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.574 Installing /home/vagrant/spdk_repo/dpdk/usertools/dpdk-devbind.py to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:16.574 Installing /home/vagrant/spdk_repo/dpdk/usertools/dpdk-pmdinfo.py to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:16.574 Installing /home/vagrant/spdk_repo/dpdk/usertools/dpdk-telemetry.py to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:16.574 Installing /home/vagrant/spdk_repo/dpdk/usertools/dpdk-hugepages.py to /home/vagrant/spdk_repo/dpdk/build/bin 00:03:16.574 Installing /home/vagrant/spdk_repo/dpdk/build-tmp/rte_build_config.h to /home/vagrant/spdk_repo/dpdk/build/include 00:03:16.574 Installing /home/vagrant/spdk_repo/dpdk/build-tmp/meson-private/libdpdk-libs.pc to /home/vagrant/spdk_repo/dpdk/build/lib/pkgconfig 00:03:16.574 Installing /home/vagrant/spdk_repo/dpdk/build-tmp/meson-private/libdpdk.pc to /home/vagrant/spdk_repo/dpdk/build/lib/pkgconfig 00:03:16.574 Installing symlink pointing to librte_kvargs.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_kvargs.so.23 00:03:16.574 Installing symlink pointing to librte_kvargs.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_kvargs.so 00:03:16.574 Installing symlink pointing to librte_telemetry.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_telemetry.so.23 00:03:16.574 Installing symlink pointing to librte_telemetry.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_telemetry.so 00:03:16.574 Installing symlink pointing to librte_eal.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_eal.so.23 00:03:16.574 Installing symlink pointing to librte_eal.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_eal.so 00:03:16.574 Installing symlink pointing to librte_ring.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_ring.so.23 00:03:16.574 Installing symlink pointing to librte_ring.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_ring.so 00:03:16.574 Installing symlink pointing to librte_rcu.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_rcu.so.23 00:03:16.574 Installing symlink pointing to librte_rcu.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_rcu.so 00:03:16.574 Installing symlink pointing to librte_mempool.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_mempool.so.23 00:03:16.574 Installing symlink pointing to librte_mempool.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_mempool.so 00:03:16.574 Installing symlink pointing to librte_mbuf.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_mbuf.so.23 00:03:16.574 Installing symlink pointing to librte_mbuf.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_mbuf.so 00:03:16.574 Installing symlink pointing to librte_net.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_net.so.23 00:03:16.574 Installing symlink pointing to librte_net.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_net.so 00:03:16.574 Installing symlink pointing to librte_meter.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_meter.so.23 00:03:16.574 Installing symlink pointing to librte_meter.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_meter.so 00:03:16.574 Installing symlink pointing to librte_ethdev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_ethdev.so.23 00:03:16.574 Installing symlink pointing to librte_ethdev.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_ethdev.so 00:03:16.574 Installing symlink pointing to librte_pci.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_pci.so.23 00:03:16.574 Installing symlink pointing to librte_pci.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_pci.so 00:03:16.574 Installing symlink pointing to librte_cmdline.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_cmdline.so.23 00:03:16.574 Installing symlink pointing to librte_cmdline.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_cmdline.so 00:03:16.574 Installing symlink pointing to librte_metrics.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_metrics.so.23 00:03:16.574 Installing symlink pointing to librte_metrics.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_metrics.so 00:03:16.574 Installing symlink pointing to librte_hash.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_hash.so.23 00:03:16.574 Installing symlink pointing to librte_hash.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_hash.so 00:03:16.574 Installing symlink pointing to librte_timer.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_timer.so.23 00:03:16.574 Installing symlink pointing to librte_timer.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_timer.so 00:03:16.574 Installing symlink pointing to librte_acl.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_acl.so.23 00:03:16.574 Installing symlink pointing to librte_acl.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_acl.so 00:03:16.574 Installing symlink pointing to librte_bbdev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_bbdev.so.23 00:03:16.574 Installing symlink pointing to librte_bbdev.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_bbdev.so 00:03:16.574 Installing symlink pointing to librte_bitratestats.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_bitratestats.so.23 00:03:16.574 Installing symlink pointing to librte_bitratestats.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_bitratestats.so 00:03:16.574 Installing symlink pointing to librte_bpf.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_bpf.so.23 00:03:16.574 Installing symlink pointing to librte_bpf.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_bpf.so 00:03:16.574 Installing symlink pointing to librte_cfgfile.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_cfgfile.so.23 00:03:16.574 Installing symlink pointing to librte_cfgfile.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_cfgfile.so 00:03:16.574 Installing symlink pointing to librte_compressdev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_compressdev.so.23 00:03:16.574 Installing symlink pointing to librte_compressdev.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_compressdev.so 00:03:16.574 Installing symlink pointing to librte_cryptodev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_cryptodev.so.23 00:03:16.574 Installing symlink pointing to librte_cryptodev.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_cryptodev.so 00:03:16.574 Installing symlink pointing to librte_distributor.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_distributor.so.23 00:03:16.574 Installing symlink pointing to librte_distributor.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_distributor.so 00:03:16.574 Installing symlink pointing to librte_efd.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_efd.so.23 00:03:16.574 Installing symlink pointing to librte_efd.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_efd.so 00:03:16.574 Installing symlink pointing to librte_eventdev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_eventdev.so.23 00:03:16.574 Installing symlink pointing to librte_eventdev.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_eventdev.so 00:03:16.574 Installing symlink pointing to librte_gpudev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_gpudev.so.23 00:03:16.574 './librte_bus_pci.so' -> 'dpdk/pmds-23.0/librte_bus_pci.so' 00:03:16.574 './librte_bus_pci.so.23' -> 'dpdk/pmds-23.0/librte_bus_pci.so.23' 00:03:16.574 './librte_bus_pci.so.23.0' -> 'dpdk/pmds-23.0/librte_bus_pci.so.23.0' 00:03:16.574 './librte_bus_vdev.so' -> 'dpdk/pmds-23.0/librte_bus_vdev.so' 00:03:16.574 './librte_bus_vdev.so.23' -> 'dpdk/pmds-23.0/librte_bus_vdev.so.23' 00:03:16.574 './librte_bus_vdev.so.23.0' -> 'dpdk/pmds-23.0/librte_bus_vdev.so.23.0' 00:03:16.574 './librte_mempool_ring.so' -> 'dpdk/pmds-23.0/librte_mempool_ring.so' 00:03:16.574 './librte_mempool_ring.so.23' -> 'dpdk/pmds-23.0/librte_mempool_ring.so.23' 00:03:16.574 './librte_mempool_ring.so.23.0' -> 'dpdk/pmds-23.0/librte_mempool_ring.so.23.0' 00:03:16.574 './librte_net_i40e.so' -> 'dpdk/pmds-23.0/librte_net_i40e.so' 00:03:16.574 './librte_net_i40e.so.23' -> 'dpdk/pmds-23.0/librte_net_i40e.so.23' 00:03:16.574 './librte_net_i40e.so.23.0' -> 'dpdk/pmds-23.0/librte_net_i40e.so.23.0' 00:03:16.574 Installing symlink pointing to librte_gpudev.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_gpudev.so 00:03:16.574 Installing symlink pointing to librte_gro.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_gro.so.23 00:03:16.574 Installing symlink pointing to librte_gro.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_gro.so 00:03:16.574 Installing symlink pointing to librte_gso.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_gso.so.23 00:03:16.574 Installing symlink pointing to librte_gso.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_gso.so 00:03:16.574 Installing symlink pointing to librte_ip_frag.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_ip_frag.so.23 00:03:16.574 Installing symlink pointing to librte_ip_frag.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_ip_frag.so 00:03:16.574 Installing symlink pointing to librte_jobstats.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_jobstats.so.23 00:03:16.574 Installing symlink pointing to librte_jobstats.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_jobstats.so 00:03:16.574 Installing symlink pointing to librte_latencystats.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_latencystats.so.23 00:03:16.574 Installing symlink pointing to librte_latencystats.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_latencystats.so 00:03:16.574 Installing symlink pointing to librte_lpm.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_lpm.so.23 00:03:16.574 Installing symlink pointing to librte_lpm.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_lpm.so 00:03:16.574 Installing symlink pointing to librte_member.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_member.so.23 00:03:16.574 Installing symlink pointing to librte_member.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_member.so 00:03:16.574 Installing symlink pointing to librte_pcapng.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_pcapng.so.23 00:03:16.574 Installing symlink pointing to librte_pcapng.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_pcapng.so 00:03:16.574 Installing symlink pointing to librte_power.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_power.so.23 00:03:16.574 Installing symlink pointing to librte_power.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_power.so 00:03:16.574 Installing symlink pointing to librte_rawdev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_rawdev.so.23 00:03:16.574 Installing symlink pointing to librte_rawdev.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_rawdev.so 00:03:16.575 Installing symlink pointing to librte_regexdev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_regexdev.so.23 00:03:16.575 Installing symlink pointing to librte_regexdev.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_regexdev.so 00:03:16.575 Installing symlink pointing to librte_dmadev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_dmadev.so.23 00:03:16.575 Installing symlink pointing to librte_dmadev.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_dmadev.so 00:03:16.575 Installing symlink pointing to librte_rib.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_rib.so.23 00:03:16.575 Installing symlink pointing to librte_rib.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_rib.so 00:03:16.575 Installing symlink pointing to librte_reorder.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_reorder.so.23 00:03:16.575 Installing symlink pointing to librte_reorder.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_reorder.so 00:03:16.575 Installing symlink pointing to librte_sched.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_sched.so.23 00:03:16.575 Installing symlink pointing to librte_sched.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_sched.so 00:03:16.575 Installing symlink pointing to librte_security.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_security.so.23 00:03:16.575 Installing symlink pointing to librte_security.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_security.so 00:03:16.575 Installing symlink pointing to librte_stack.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_stack.so.23 00:03:16.575 Installing symlink pointing to librte_stack.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_stack.so 00:03:16.575 Installing symlink pointing to librte_vhost.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_vhost.so.23 00:03:16.575 Installing symlink pointing to librte_vhost.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_vhost.so 00:03:16.575 Installing symlink pointing to librte_ipsec.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_ipsec.so.23 00:03:16.575 Installing symlink pointing to librte_ipsec.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_ipsec.so 00:03:16.575 Installing symlink pointing to librte_fib.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_fib.so.23 00:03:16.575 Installing symlink pointing to librte_fib.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_fib.so 00:03:16.575 Installing symlink pointing to librte_port.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_port.so.23 00:03:16.575 Installing symlink pointing to librte_port.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_port.so 00:03:16.575 Installing symlink pointing to librte_pdump.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_pdump.so.23 00:03:16.575 Installing symlink pointing to librte_pdump.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_pdump.so 00:03:16.575 Installing symlink pointing to librte_table.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_table.so.23 00:03:16.575 Installing symlink pointing to librte_table.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_table.so 00:03:16.575 Installing symlink pointing to librte_pipeline.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_pipeline.so.23 00:03:16.575 Installing symlink pointing to librte_pipeline.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_pipeline.so 00:03:16.575 Installing symlink pointing to librte_graph.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_graph.so.23 00:03:16.575 Installing symlink pointing to librte_graph.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_graph.so 00:03:16.575 Installing symlink pointing to librte_node.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_node.so.23 00:03:16.575 Installing symlink pointing to librte_node.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/librte_node.so 00:03:16.575 Installing symlink pointing to librte_bus_pci.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0/librte_bus_pci.so.23 00:03:16.575 Installing symlink pointing to librte_bus_pci.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0/librte_bus_pci.so 00:03:16.575 Installing symlink pointing to librte_bus_vdev.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0/librte_bus_vdev.so.23 00:03:16.575 Installing symlink pointing to librte_bus_vdev.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0/librte_bus_vdev.so 00:03:16.575 Installing symlink pointing to librte_mempool_ring.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0/librte_mempool_ring.so.23 00:03:16.575 Installing symlink pointing to librte_mempool_ring.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0/librte_mempool_ring.so 00:03:16.575 Installing symlink pointing to librte_net_i40e.so.23.0 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0/librte_net_i40e.so.23 00:03:16.575 Installing symlink pointing to librte_net_i40e.so.23 to /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0/librte_net_i40e.so 00:03:16.575 Running custom install script '/bin/sh /home/vagrant/spdk_repo/dpdk/config/../buildtools/symlink-drivers-solibs.sh lib dpdk/pmds-23.0' 00:03:16.834 17:24:47 build_native_dpdk -- common/autobuild_common.sh@213 -- $ cat 00:03:16.834 17:24:47 build_native_dpdk -- common/autobuild_common.sh@218 -- $ cd /home/vagrant/spdk_repo/spdk 00:03:16.834 00:03:16.834 real 0m43.078s 00:03:16.834 user 4m7.086s 00:03:16.834 sys 0m50.174s 00:03:16.834 17:24:47 build_native_dpdk -- common/autotest_common.sh@1126 -- $ xtrace_disable 00:03:16.834 17:24:47 build_native_dpdk -- common/autotest_common.sh@10 -- $ set +x 00:03:16.834 ************************************ 00:03:16.834 END TEST build_native_dpdk 00:03:16.834 ************************************ 00:03:16.834 17:24:47 -- spdk/autobuild.sh@31 -- $ case "$SPDK_TEST_AUTOBUILD" in 00:03:16.834 17:24:47 -- spdk/autobuild.sh@47 -- $ [[ 0 -eq 1 ]] 00:03:16.834 17:24:47 -- spdk/autobuild.sh@51 -- $ [[ 0 -eq 1 ]] 00:03:16.834 17:24:47 -- spdk/autobuild.sh@55 -- $ [[ -n '' ]] 00:03:16.834 17:24:47 -- spdk/autobuild.sh@57 -- $ [[ 0 -eq 1 ]] 00:03:16.834 17:24:47 -- spdk/autobuild.sh@59 -- $ [[ 0 -eq 1 ]] 00:03:16.834 17:24:47 -- spdk/autobuild.sh@62 -- $ [[ 0 -eq 1 ]] 00:03:16.834 17:24:47 -- spdk/autobuild.sh@67 -- $ /home/vagrant/spdk_repo/spdk/configure --enable-debug --enable-werror --with-rdma --with-idxd --with-fio=/usr/src/fio --with-iscsi-initiator --disable-unit-tests --enable-ubsan --enable-asan --enable-coverage --with-ublk --with-raid5f --with-dpdk=/home/vagrant/spdk_repo/dpdk/build --with-shared 00:03:16.834 Using /home/vagrant/spdk_repo/dpdk/build/lib/pkgconfig for additional libs... 00:03:17.092 DPDK libraries: /home/vagrant/spdk_repo/dpdk/build/lib 00:03:17.092 DPDK includes: //home/vagrant/spdk_repo/dpdk/build/include 00:03:17.092 Using default SPDK env in /home/vagrant/spdk_repo/spdk/lib/env_dpdk 00:03:17.351 Using 'verbs' RDMA provider 00:03:33.611 Configuring ISA-L (logfile: /home/vagrant/spdk_repo/spdk/.spdk-isal.log)...done. 00:03:51.703 Configuring ISA-L-crypto (logfile: /home/vagrant/spdk_repo/spdk/.spdk-isal-crypto.log)...done. 00:03:51.703 Creating mk/config.mk...done. 00:03:51.703 Creating mk/cc.flags.mk...done. 00:03:51.703 Type 'make' to build. 00:03:51.703 17:25:21 -- spdk/autobuild.sh@70 -- $ run_test make make -j10 00:03:51.703 17:25:21 -- common/autotest_common.sh@1101 -- $ '[' 3 -le 1 ']' 00:03:51.703 17:25:21 -- common/autotest_common.sh@1107 -- $ xtrace_disable 00:03:51.703 17:25:21 -- common/autotest_common.sh@10 -- $ set +x 00:03:51.703 ************************************ 00:03:51.703 START TEST make 00:03:51.703 ************************************ 00:03:51.703 17:25:21 make -- common/autotest_common.sh@1125 -- $ make -j10 00:03:51.704 make[1]: Nothing to be done for 'all'. 00:04:30.431 CC lib/log/log_flags.o 00:04:30.431 CC lib/log/log.o 00:04:30.431 CC lib/ut_mock/mock.o 00:04:30.431 CC lib/log/log_deprecated.o 00:04:30.431 CC lib/ut/ut.o 00:04:30.431 LIB libspdk_log.a 00:04:30.431 LIB libspdk_ut_mock.a 00:04:30.431 LIB libspdk_ut.a 00:04:30.431 SO libspdk_log.so.7.0 00:04:30.431 SO libspdk_ut_mock.so.6.0 00:04:30.431 SO libspdk_ut.so.2.0 00:04:30.431 SYMLINK libspdk_log.so 00:04:30.431 SYMLINK libspdk_ut_mock.so 00:04:30.431 SYMLINK libspdk_ut.so 00:04:30.689 CC lib/ioat/ioat.o 00:04:30.689 CC lib/dma/dma.o 00:04:30.689 CXX lib/trace_parser/trace.o 00:04:30.689 CC lib/util/base64.o 00:04:30.689 CC lib/util/bit_array.o 00:04:30.689 CC lib/util/crc32.o 00:04:30.689 CC lib/util/cpuset.o 00:04:30.689 CC lib/util/crc16.o 00:04:30.689 CC lib/util/crc32c.o 00:04:30.948 CC lib/vfio_user/host/vfio_user_pci.o 00:04:30.948 CC lib/util/crc32_ieee.o 00:04:30.948 CC lib/util/crc64.o 00:04:30.948 CC lib/vfio_user/host/vfio_user.o 00:04:30.948 LIB libspdk_dma.a 00:04:30.948 SO libspdk_dma.so.5.0 00:04:30.948 CC lib/util/dif.o 00:04:30.948 CC lib/util/fd.o 00:04:30.948 LIB libspdk_ioat.a 00:04:30.948 SYMLINK libspdk_dma.so 00:04:30.948 CC lib/util/fd_group.o 00:04:30.948 CC lib/util/file.o 00:04:30.948 SO libspdk_ioat.so.7.0 00:04:30.948 CC lib/util/hexlify.o 00:04:30.948 CC lib/util/iov.o 00:04:30.948 SYMLINK libspdk_ioat.so 00:04:30.948 CC lib/util/math.o 00:04:30.948 CC lib/util/net.o 00:04:31.207 CC lib/util/pipe.o 00:04:31.207 LIB libspdk_vfio_user.a 00:04:31.207 CC lib/util/strerror_tls.o 00:04:31.207 SO libspdk_vfio_user.so.5.0 00:04:31.207 CC lib/util/string.o 00:04:31.207 CC lib/util/uuid.o 00:04:31.207 SYMLINK libspdk_vfio_user.so 00:04:31.207 CC lib/util/xor.o 00:04:31.207 CC lib/util/zipf.o 00:04:31.207 CC lib/util/md5.o 00:04:31.466 LIB libspdk_util.a 00:04:31.725 LIB libspdk_trace_parser.a 00:04:31.725 SO libspdk_util.so.10.0 00:04:31.725 SO libspdk_trace_parser.so.6.0 00:04:31.725 SYMLINK libspdk_util.so 00:04:31.725 SYMLINK libspdk_trace_parser.so 00:04:31.984 CC lib/conf/conf.o 00:04:31.984 CC lib/rdma_utils/rdma_utils.o 00:04:31.984 CC lib/json/json_parse.o 00:04:31.984 CC lib/json/json_util.o 00:04:31.984 CC lib/json/json_write.o 00:04:31.984 CC lib/idxd/idxd.o 00:04:31.984 CC lib/idxd/idxd_user.o 00:04:31.984 CC lib/vmd/vmd.o 00:04:31.984 CC lib/rdma_provider/common.o 00:04:31.984 CC lib/env_dpdk/env.o 00:04:31.984 CC lib/rdma_provider/rdma_provider_verbs.o 00:04:31.985 LIB libspdk_conf.a 00:04:32.243 SO libspdk_conf.so.6.0 00:04:32.243 CC lib/idxd/idxd_kernel.o 00:04:32.243 CC lib/env_dpdk/memory.o 00:04:32.243 CC lib/env_dpdk/pci.o 00:04:32.243 LIB libspdk_rdma_utils.a 00:04:32.243 SYMLINK libspdk_conf.so 00:04:32.243 LIB libspdk_json.a 00:04:32.243 CC lib/env_dpdk/init.o 00:04:32.243 SO libspdk_rdma_utils.so.1.0 00:04:32.243 SO libspdk_json.so.6.0 00:04:32.243 LIB libspdk_rdma_provider.a 00:04:32.243 SYMLINK libspdk_rdma_utils.so 00:04:32.243 SYMLINK libspdk_json.so 00:04:32.243 CC lib/vmd/led.o 00:04:32.243 CC lib/env_dpdk/threads.o 00:04:32.243 SO libspdk_rdma_provider.so.6.0 00:04:32.243 SYMLINK libspdk_rdma_provider.so 00:04:32.243 CC lib/env_dpdk/pci_ioat.o 00:04:32.502 CC lib/env_dpdk/pci_virtio.o 00:04:32.502 CC lib/jsonrpc/jsonrpc_server.o 00:04:32.502 CC lib/jsonrpc/jsonrpc_server_tcp.o 00:04:32.502 CC lib/env_dpdk/pci_vmd.o 00:04:32.502 CC lib/env_dpdk/pci_idxd.o 00:04:32.502 CC lib/env_dpdk/pci_event.o 00:04:32.502 CC lib/env_dpdk/sigbus_handler.o 00:04:32.502 CC lib/env_dpdk/pci_dpdk.o 00:04:32.502 LIB libspdk_idxd.a 00:04:32.502 CC lib/env_dpdk/pci_dpdk_2207.o 00:04:32.502 SO libspdk_idxd.so.12.1 00:04:32.502 LIB libspdk_vmd.a 00:04:32.502 CC lib/env_dpdk/pci_dpdk_2211.o 00:04:32.502 SO libspdk_vmd.so.6.0 00:04:32.761 CC lib/jsonrpc/jsonrpc_client.o 00:04:32.761 CC lib/jsonrpc/jsonrpc_client_tcp.o 00:04:32.761 SYMLINK libspdk_idxd.so 00:04:32.761 SYMLINK libspdk_vmd.so 00:04:32.761 LIB libspdk_jsonrpc.a 00:04:33.020 SO libspdk_jsonrpc.so.6.0 00:04:33.021 SYMLINK libspdk_jsonrpc.so 00:04:33.279 LIB libspdk_env_dpdk.a 00:04:33.279 CC lib/rpc/rpc.o 00:04:33.537 SO libspdk_env_dpdk.so.15.0 00:04:33.537 SYMLINK libspdk_env_dpdk.so 00:04:33.537 LIB libspdk_rpc.a 00:04:33.796 SO libspdk_rpc.so.6.0 00:04:33.796 SYMLINK libspdk_rpc.so 00:04:34.055 CC lib/notify/notify.o 00:04:34.055 CC lib/notify/notify_rpc.o 00:04:34.055 CC lib/trace/trace.o 00:04:34.055 CC lib/trace/trace_rpc.o 00:04:34.055 CC lib/trace/trace_flags.o 00:04:34.055 CC lib/keyring/keyring.o 00:04:34.055 CC lib/keyring/keyring_rpc.o 00:04:34.313 LIB libspdk_notify.a 00:04:34.313 SO libspdk_notify.so.6.0 00:04:34.313 SYMLINK libspdk_notify.so 00:04:34.313 LIB libspdk_keyring.a 00:04:34.313 LIB libspdk_trace.a 00:04:34.313 SO libspdk_keyring.so.2.0 00:04:34.313 SO libspdk_trace.so.11.0 00:04:34.572 SYMLINK libspdk_keyring.so 00:04:34.572 SYMLINK libspdk_trace.so 00:04:34.830 CC lib/sock/sock.o 00:04:34.830 CC lib/sock/sock_rpc.o 00:04:34.830 CC lib/thread/thread.o 00:04:34.830 CC lib/thread/iobuf.o 00:04:35.396 LIB libspdk_sock.a 00:04:35.396 SO libspdk_sock.so.10.0 00:04:35.396 SYMLINK libspdk_sock.so 00:04:35.655 CC lib/nvme/nvme_ctrlr_cmd.o 00:04:35.655 CC lib/nvme/nvme_ctrlr.o 00:04:35.655 CC lib/nvme/nvme_fabric.o 00:04:35.655 CC lib/nvme/nvme_ns.o 00:04:35.655 CC lib/nvme/nvme_ns_cmd.o 00:04:35.655 CC lib/nvme/nvme_pcie_common.o 00:04:35.655 CC lib/nvme/nvme_pcie.o 00:04:35.655 CC lib/nvme/nvme_qpair.o 00:04:35.655 CC lib/nvme/nvme.o 00:04:36.223 LIB libspdk_thread.a 00:04:36.481 CC lib/nvme/nvme_quirks.o 00:04:36.481 SO libspdk_thread.so.10.1 00:04:36.481 CC lib/nvme/nvme_transport.o 00:04:36.481 CC lib/nvme/nvme_discovery.o 00:04:36.481 SYMLINK libspdk_thread.so 00:04:36.481 CC lib/nvme/nvme_ctrlr_ocssd_cmd.o 00:04:36.481 CC lib/nvme/nvme_ns_ocssd_cmd.o 00:04:36.481 CC lib/nvme/nvme_tcp.o 00:04:36.740 CC lib/nvme/nvme_opal.o 00:04:36.740 CC lib/accel/accel.o 00:04:36.740 CC lib/nvme/nvme_io_msg.o 00:04:36.740 CC lib/nvme/nvme_poll_group.o 00:04:36.999 CC lib/accel/accel_rpc.o 00:04:36.999 CC lib/accel/accel_sw.o 00:04:36.999 CC lib/nvme/nvme_zns.o 00:04:37.257 CC lib/nvme/nvme_stubs.o 00:04:37.257 CC lib/nvme/nvme_auth.o 00:04:37.257 CC lib/blob/blobstore.o 00:04:37.257 CC lib/nvme/nvme_cuse.o 00:04:37.516 CC lib/nvme/nvme_rdma.o 00:04:37.516 CC lib/init/json_config.o 00:04:37.516 CC lib/init/subsystem.o 00:04:37.516 CC lib/init/subsystem_rpc.o 00:04:37.775 CC lib/init/rpc.o 00:04:37.775 LIB libspdk_accel.a 00:04:37.775 CC lib/blob/request.o 00:04:37.775 CC lib/blob/zeroes.o 00:04:37.776 SO libspdk_accel.so.16.0 00:04:37.776 LIB libspdk_init.a 00:04:37.776 SO libspdk_init.so.6.0 00:04:37.776 SYMLINK libspdk_accel.so 00:04:37.776 CC lib/blob/blob_bs_dev.o 00:04:37.776 SYMLINK libspdk_init.so 00:04:38.034 CC lib/virtio/virtio.o 00:04:38.034 CC lib/fsdev/fsdev.o 00:04:38.034 CC lib/fsdev/fsdev_io.o 00:04:38.034 CC lib/virtio/virtio_vhost_user.o 00:04:38.034 CC lib/fsdev/fsdev_rpc.o 00:04:38.034 CC lib/bdev/bdev.o 00:04:38.034 CC lib/event/app.o 00:04:38.293 CC lib/bdev/bdev_rpc.o 00:04:38.293 CC lib/bdev/bdev_zone.o 00:04:38.293 CC lib/virtio/virtio_vfio_user.o 00:04:38.293 CC lib/event/reactor.o 00:04:38.293 CC lib/event/log_rpc.o 00:04:38.293 CC lib/bdev/part.o 00:04:38.552 CC lib/bdev/scsi_nvme.o 00:04:38.552 CC lib/virtio/virtio_pci.o 00:04:38.552 CC lib/event/app_rpc.o 00:04:38.552 CC lib/event/scheduler_static.o 00:04:38.552 LIB libspdk_fsdev.a 00:04:38.812 SO libspdk_fsdev.so.1.0 00:04:38.812 LIB libspdk_nvme.a 00:04:38.812 SYMLINK libspdk_fsdev.so 00:04:38.812 LIB libspdk_virtio.a 00:04:38.812 LIB libspdk_event.a 00:04:38.812 SO libspdk_virtio.so.7.0 00:04:38.812 SO libspdk_event.so.14.0 00:04:38.812 SYMLINK libspdk_virtio.so 00:04:38.812 SO libspdk_nvme.so.14.0 00:04:38.812 SYMLINK libspdk_event.so 00:04:39.072 CC lib/fuse_dispatcher/fuse_dispatcher.o 00:04:39.072 SYMLINK libspdk_nvme.so 00:04:39.641 LIB libspdk_fuse_dispatcher.a 00:04:39.900 SO libspdk_fuse_dispatcher.so.1.0 00:04:39.900 SYMLINK libspdk_fuse_dispatcher.so 00:04:40.837 LIB libspdk_blob.a 00:04:40.837 SO libspdk_blob.so.11.0 00:04:40.837 SYMLINK libspdk_blob.so 00:04:40.837 LIB libspdk_bdev.a 00:04:41.096 SO libspdk_bdev.so.16.0 00:04:41.096 SYMLINK libspdk_bdev.so 00:04:41.096 CC lib/lvol/lvol.o 00:04:41.096 CC lib/blobfs/blobfs.o 00:04:41.096 CC lib/blobfs/tree.o 00:04:41.355 CC lib/nbd/nbd.o 00:04:41.355 CC lib/scsi/dev.o 00:04:41.355 CC lib/ftl/ftl_core.o 00:04:41.355 CC lib/nbd/nbd_rpc.o 00:04:41.355 CC lib/scsi/lun.o 00:04:41.355 CC lib/ublk/ublk.o 00:04:41.355 CC lib/nvmf/ctrlr.o 00:04:41.355 CC lib/nvmf/ctrlr_discovery.o 00:04:41.355 CC lib/ublk/ublk_rpc.o 00:04:41.355 CC lib/scsi/port.o 00:04:41.614 CC lib/scsi/scsi.o 00:04:41.614 CC lib/scsi/scsi_bdev.o 00:04:41.614 CC lib/scsi/scsi_pr.o 00:04:41.614 LIB libspdk_nbd.a 00:04:41.614 CC lib/ftl/ftl_init.o 00:04:41.614 SO libspdk_nbd.so.7.0 00:04:41.614 CC lib/scsi/scsi_rpc.o 00:04:41.873 SYMLINK libspdk_nbd.so 00:04:41.873 CC lib/scsi/task.o 00:04:41.873 CC lib/nvmf/ctrlr_bdev.o 00:04:41.873 CC lib/nvmf/subsystem.o 00:04:41.873 CC lib/ftl/ftl_layout.o 00:04:41.873 LIB libspdk_ublk.a 00:04:41.873 SO libspdk_ublk.so.3.0 00:04:41.873 CC lib/ftl/ftl_debug.o 00:04:41.873 CC lib/ftl/ftl_io.o 00:04:41.873 LIB libspdk_blobfs.a 00:04:41.873 SYMLINK libspdk_ublk.so 00:04:41.873 CC lib/nvmf/nvmf.o 00:04:41.873 SO libspdk_blobfs.so.10.0 00:04:42.132 LIB libspdk_scsi.a 00:04:42.132 SYMLINK libspdk_blobfs.so 00:04:42.132 CC lib/nvmf/nvmf_rpc.o 00:04:42.132 LIB libspdk_lvol.a 00:04:42.132 SO libspdk_scsi.so.9.0 00:04:42.132 SO libspdk_lvol.so.10.0 00:04:42.132 SYMLINK libspdk_scsi.so 00:04:42.132 CC lib/ftl/ftl_sb.o 00:04:42.132 CC lib/ftl/ftl_l2p.o 00:04:42.132 CC lib/nvmf/transport.o 00:04:42.132 SYMLINK libspdk_lvol.so 00:04:42.132 CC lib/nvmf/tcp.o 00:04:42.391 CC lib/iscsi/conn.o 00:04:42.391 CC lib/ftl/ftl_l2p_flat.o 00:04:42.391 CC lib/nvmf/stubs.o 00:04:42.391 CC lib/vhost/vhost.o 00:04:42.650 CC lib/ftl/ftl_nv_cache.o 00:04:42.909 CC lib/nvmf/mdns_server.o 00:04:42.909 CC lib/nvmf/rdma.o 00:04:42.909 CC lib/nvmf/auth.o 00:04:42.909 CC lib/vhost/vhost_rpc.o 00:04:42.909 CC lib/iscsi/init_grp.o 00:04:43.167 CC lib/ftl/ftl_band.o 00:04:43.167 CC lib/vhost/vhost_scsi.o 00:04:43.167 CC lib/vhost/vhost_blk.o 00:04:43.167 CC lib/iscsi/iscsi.o 00:04:43.167 CC lib/ftl/ftl_band_ops.o 00:04:43.426 CC lib/ftl/ftl_writer.o 00:04:43.426 CC lib/vhost/rte_vhost_user.o 00:04:43.685 CC lib/ftl/ftl_rq.o 00:04:43.685 CC lib/iscsi/param.o 00:04:43.685 CC lib/iscsi/portal_grp.o 00:04:43.685 CC lib/ftl/ftl_reloc.o 00:04:43.685 CC lib/ftl/ftl_l2p_cache.o 00:04:43.945 CC lib/iscsi/tgt_node.o 00:04:43.945 CC lib/iscsi/iscsi_subsystem.o 00:04:43.945 CC lib/iscsi/iscsi_rpc.o 00:04:43.945 CC lib/iscsi/task.o 00:04:43.945 CC lib/ftl/ftl_p2l.o 00:04:43.945 CC lib/ftl/ftl_p2l_log.o 00:04:44.203 CC lib/ftl/mngt/ftl_mngt.o 00:04:44.203 CC lib/ftl/mngt/ftl_mngt_bdev.o 00:04:44.203 CC lib/ftl/mngt/ftl_mngt_shutdown.o 00:04:44.203 CC lib/ftl/mngt/ftl_mngt_startup.o 00:04:44.203 CC lib/ftl/mngt/ftl_mngt_md.o 00:04:44.462 CC lib/ftl/mngt/ftl_mngt_misc.o 00:04:44.462 LIB libspdk_vhost.a 00:04:44.462 CC lib/ftl/mngt/ftl_mngt_ioch.o 00:04:44.462 CC lib/ftl/mngt/ftl_mngt_l2p.o 00:04:44.462 SO libspdk_vhost.so.8.0 00:04:44.462 CC lib/ftl/mngt/ftl_mngt_band.o 00:04:44.462 CC lib/ftl/mngt/ftl_mngt_self_test.o 00:04:44.462 CC lib/ftl/mngt/ftl_mngt_p2l.o 00:04:44.462 SYMLINK libspdk_vhost.so 00:04:44.462 CC lib/ftl/mngt/ftl_mngt_recovery.o 00:04:44.462 CC lib/ftl/mngt/ftl_mngt_upgrade.o 00:04:44.462 CC lib/ftl/utils/ftl_conf.o 00:04:44.721 CC lib/ftl/utils/ftl_md.o 00:04:44.721 CC lib/ftl/utils/ftl_mempool.o 00:04:44.721 CC lib/ftl/utils/ftl_bitmap.o 00:04:44.721 CC lib/ftl/utils/ftl_property.o 00:04:44.721 LIB libspdk_iscsi.a 00:04:44.721 CC lib/ftl/utils/ftl_layout_tracker_bdev.o 00:04:44.721 CC lib/ftl/upgrade/ftl_layout_upgrade.o 00:04:44.721 SO libspdk_iscsi.so.8.0 00:04:44.721 CC lib/ftl/upgrade/ftl_sb_upgrade.o 00:04:44.721 CC lib/ftl/upgrade/ftl_p2l_upgrade.o 00:04:44.721 CC lib/ftl/upgrade/ftl_band_upgrade.o 00:04:44.979 CC lib/ftl/upgrade/ftl_chunk_upgrade.o 00:04:44.979 CC lib/ftl/upgrade/ftl_trim_upgrade.o 00:04:44.979 SYMLINK libspdk_iscsi.so 00:04:44.979 CC lib/ftl/upgrade/ftl_sb_v3.o 00:04:44.979 CC lib/ftl/upgrade/ftl_sb_v5.o 00:04:44.979 CC lib/ftl/nvc/ftl_nvc_dev.o 00:04:44.979 CC lib/ftl/nvc/ftl_nvc_bdev_vss.o 00:04:44.979 CC lib/ftl/nvc/ftl_nvc_bdev_non_vss.o 00:04:44.979 CC lib/ftl/nvc/ftl_nvc_bdev_common.o 00:04:44.979 LIB libspdk_nvmf.a 00:04:44.979 CC lib/ftl/base/ftl_base_dev.o 00:04:44.979 CC lib/ftl/base/ftl_base_bdev.o 00:04:44.979 CC lib/ftl/ftl_trace.o 00:04:45.237 SO libspdk_nvmf.so.19.0 00:04:45.237 LIB libspdk_ftl.a 00:04:45.237 SYMLINK libspdk_nvmf.so 00:04:45.496 SO libspdk_ftl.so.9.0 00:04:45.755 SYMLINK libspdk_ftl.so 00:04:46.014 CC module/env_dpdk/env_dpdk_rpc.o 00:04:46.272 CC module/keyring/linux/keyring.o 00:04:46.272 CC module/keyring/file/keyring.o 00:04:46.272 CC module/accel/ioat/accel_ioat.o 00:04:46.272 CC module/accel/error/accel_error.o 00:04:46.272 CC module/blob/bdev/blob_bdev.o 00:04:46.272 CC module/sock/posix/posix.o 00:04:46.272 CC module/fsdev/aio/fsdev_aio.o 00:04:46.272 CC module/accel/dsa/accel_dsa.o 00:04:46.272 CC module/scheduler/dynamic/scheduler_dynamic.o 00:04:46.272 LIB libspdk_env_dpdk_rpc.a 00:04:46.272 SO libspdk_env_dpdk_rpc.so.6.0 00:04:46.272 CC module/keyring/linux/keyring_rpc.o 00:04:46.272 CC module/keyring/file/keyring_rpc.o 00:04:46.272 SYMLINK libspdk_env_dpdk_rpc.so 00:04:46.272 CC module/accel/dsa/accel_dsa_rpc.o 00:04:46.272 CC module/accel/error/accel_error_rpc.o 00:04:46.272 CC module/accel/ioat/accel_ioat_rpc.o 00:04:46.272 LIB libspdk_scheduler_dynamic.a 00:04:46.272 SO libspdk_scheduler_dynamic.so.4.0 00:04:46.531 LIB libspdk_keyring_linux.a 00:04:46.531 SYMLINK libspdk_scheduler_dynamic.so 00:04:46.531 LIB libspdk_accel_dsa.a 00:04:46.531 LIB libspdk_blob_bdev.a 00:04:46.531 LIB libspdk_keyring_file.a 00:04:46.531 LIB libspdk_accel_error.a 00:04:46.531 SO libspdk_keyring_linux.so.1.0 00:04:46.531 LIB libspdk_accel_ioat.a 00:04:46.531 SO libspdk_accel_dsa.so.5.0 00:04:46.531 SO libspdk_blob_bdev.so.11.0 00:04:46.531 SO libspdk_accel_error.so.2.0 00:04:46.531 SO libspdk_keyring_file.so.2.0 00:04:46.531 SO libspdk_accel_ioat.so.6.0 00:04:46.531 SYMLINK libspdk_keyring_linux.so 00:04:46.531 SYMLINK libspdk_accel_dsa.so 00:04:46.531 SYMLINK libspdk_blob_bdev.so 00:04:46.531 CC module/fsdev/aio/fsdev_aio_rpc.o 00:04:46.531 SYMLINK libspdk_accel_error.so 00:04:46.531 CC module/fsdev/aio/linux_aio_mgr.o 00:04:46.531 SYMLINK libspdk_keyring_file.so 00:04:46.531 SYMLINK libspdk_accel_ioat.so 00:04:46.531 CC module/scheduler/dpdk_governor/dpdk_governor.o 00:04:46.531 CC module/accel/iaa/accel_iaa.o 00:04:46.531 CC module/accel/iaa/accel_iaa_rpc.o 00:04:46.790 CC module/scheduler/gscheduler/gscheduler.o 00:04:46.790 LIB libspdk_scheduler_dpdk_governor.a 00:04:46.790 SO libspdk_scheduler_dpdk_governor.so.4.0 00:04:46.790 CC module/bdev/delay/vbdev_delay.o 00:04:46.790 CC module/blobfs/bdev/blobfs_bdev.o 00:04:46.790 LIB libspdk_accel_iaa.a 00:04:46.790 LIB libspdk_fsdev_aio.a 00:04:46.790 CC module/bdev/error/vbdev_error.o 00:04:46.790 SO libspdk_accel_iaa.so.3.0 00:04:46.790 SYMLINK libspdk_scheduler_dpdk_governor.so 00:04:46.790 LIB libspdk_scheduler_gscheduler.a 00:04:46.790 CC module/bdev/error/vbdev_error_rpc.o 00:04:46.790 SO libspdk_fsdev_aio.so.1.0 00:04:46.790 SO libspdk_scheduler_gscheduler.so.4.0 00:04:46.790 CC module/bdev/gpt/gpt.o 00:04:46.790 CC module/bdev/lvol/vbdev_lvol.o 00:04:46.790 SYMLINK libspdk_accel_iaa.so 00:04:46.790 CC module/bdev/lvol/vbdev_lvol_rpc.o 00:04:47.048 SYMLINK libspdk_fsdev_aio.so 00:04:47.048 CC module/bdev/gpt/vbdev_gpt.o 00:04:47.048 SYMLINK libspdk_scheduler_gscheduler.so 00:04:47.048 LIB libspdk_sock_posix.a 00:04:47.048 CC module/blobfs/bdev/blobfs_bdev_rpc.o 00:04:47.048 SO libspdk_sock_posix.so.6.0 00:04:47.048 SYMLINK libspdk_sock_posix.so 00:04:47.048 CC module/bdev/delay/vbdev_delay_rpc.o 00:04:47.048 CC module/bdev/malloc/bdev_malloc.o 00:04:47.048 LIB libspdk_bdev_error.a 00:04:47.048 SO libspdk_bdev_error.so.6.0 00:04:47.048 LIB libspdk_blobfs_bdev.a 00:04:47.048 CC module/bdev/null/bdev_null.o 00:04:47.048 CC module/bdev/null/bdev_null_rpc.o 00:04:47.048 SO libspdk_blobfs_bdev.so.6.0 00:04:47.310 SYMLINK libspdk_bdev_error.so 00:04:47.310 LIB libspdk_bdev_gpt.a 00:04:47.310 SO libspdk_bdev_gpt.so.6.0 00:04:47.310 CC module/bdev/nvme/bdev_nvme.o 00:04:47.310 LIB libspdk_bdev_delay.a 00:04:47.310 SYMLINK libspdk_blobfs_bdev.so 00:04:47.310 CC module/bdev/nvme/bdev_nvme_rpc.o 00:04:47.310 SO libspdk_bdev_delay.so.6.0 00:04:47.310 SYMLINK libspdk_bdev_gpt.so 00:04:47.310 CC module/bdev/malloc/bdev_malloc_rpc.o 00:04:47.310 SYMLINK libspdk_bdev_delay.so 00:04:47.310 CC module/bdev/nvme/nvme_rpc.o 00:04:47.310 CC module/bdev/passthru/vbdev_passthru.o 00:04:47.310 LIB libspdk_bdev_lvol.a 00:04:47.310 CC module/bdev/raid/bdev_raid.o 00:04:47.310 LIB libspdk_bdev_null.a 00:04:47.310 CC module/bdev/nvme/bdev_mdns_client.o 00:04:47.310 SO libspdk_bdev_lvol.so.6.0 00:04:47.310 LIB libspdk_bdev_malloc.a 00:04:47.576 CC module/bdev/split/vbdev_split.o 00:04:47.576 SO libspdk_bdev_null.so.6.0 00:04:47.576 SO libspdk_bdev_malloc.so.6.0 00:04:47.576 SYMLINK libspdk_bdev_lvol.so 00:04:47.576 CC module/bdev/split/vbdev_split_rpc.o 00:04:47.576 SYMLINK libspdk_bdev_malloc.so 00:04:47.576 CC module/bdev/raid/bdev_raid_rpc.o 00:04:47.576 SYMLINK libspdk_bdev_null.so 00:04:47.576 CC module/bdev/nvme/vbdev_opal.o 00:04:47.576 CC module/bdev/passthru/vbdev_passthru_rpc.o 00:04:47.576 CC module/bdev/nvme/vbdev_opal_rpc.o 00:04:47.576 LIB libspdk_bdev_split.a 00:04:47.846 CC module/bdev/zone_block/vbdev_zone_block.o 00:04:47.846 SO libspdk_bdev_split.so.6.0 00:04:47.846 CC module/bdev/aio/bdev_aio.o 00:04:47.846 CC module/bdev/nvme/bdev_nvme_cuse_rpc.o 00:04:47.846 LIB libspdk_bdev_passthru.a 00:04:47.846 SYMLINK libspdk_bdev_split.so 00:04:47.846 CC module/bdev/raid/bdev_raid_sb.o 00:04:47.846 SO libspdk_bdev_passthru.so.6.0 00:04:47.846 SYMLINK libspdk_bdev_passthru.so 00:04:47.846 CC module/bdev/aio/bdev_aio_rpc.o 00:04:47.846 CC module/bdev/raid/raid0.o 00:04:48.114 CC module/bdev/virtio/bdev_virtio_scsi.o 00:04:48.114 CC module/bdev/iscsi/bdev_iscsi.o 00:04:48.114 CC module/bdev/ftl/bdev_ftl.o 00:04:48.114 CC module/bdev/zone_block/vbdev_zone_block_rpc.o 00:04:48.114 CC module/bdev/virtio/bdev_virtio_blk.o 00:04:48.114 LIB libspdk_bdev_aio.a 00:04:48.114 CC module/bdev/raid/raid1.o 00:04:48.114 SO libspdk_bdev_aio.so.6.0 00:04:48.114 CC module/bdev/raid/concat.o 00:04:48.114 SYMLINK libspdk_bdev_aio.so 00:04:48.114 CC module/bdev/ftl/bdev_ftl_rpc.o 00:04:48.114 LIB libspdk_bdev_zone_block.a 00:04:48.114 SO libspdk_bdev_zone_block.so.6.0 00:04:48.371 SYMLINK libspdk_bdev_zone_block.so 00:04:48.371 CC module/bdev/raid/raid5f.o 00:04:48.371 CC module/bdev/iscsi/bdev_iscsi_rpc.o 00:04:48.371 CC module/bdev/virtio/bdev_virtio_rpc.o 00:04:48.371 LIB libspdk_bdev_ftl.a 00:04:48.371 SO libspdk_bdev_ftl.so.6.0 00:04:48.371 LIB libspdk_bdev_iscsi.a 00:04:48.371 SYMLINK libspdk_bdev_ftl.so 00:04:48.371 SO libspdk_bdev_iscsi.so.6.0 00:04:48.629 LIB libspdk_bdev_virtio.a 00:04:48.629 SYMLINK libspdk_bdev_iscsi.so 00:04:48.629 SO libspdk_bdev_virtio.so.6.0 00:04:48.629 SYMLINK libspdk_bdev_virtio.so 00:04:48.629 LIB libspdk_bdev_raid.a 00:04:48.888 SO libspdk_bdev_raid.so.6.0 00:04:48.888 SYMLINK libspdk_bdev_raid.so 00:04:49.826 LIB libspdk_bdev_nvme.a 00:04:49.826 SO libspdk_bdev_nvme.so.7.0 00:04:49.826 SYMLINK libspdk_bdev_nvme.so 00:04:50.393 CC module/event/subsystems/sock/sock.o 00:04:50.393 CC module/event/subsystems/fsdev/fsdev.o 00:04:50.393 CC module/event/subsystems/vmd/vmd.o 00:04:50.393 CC module/event/subsystems/scheduler/scheduler.o 00:04:50.393 CC module/event/subsystems/vmd/vmd_rpc.o 00:04:50.393 CC module/event/subsystems/iobuf/iobuf.o 00:04:50.393 CC module/event/subsystems/vhost_blk/vhost_blk.o 00:04:50.393 CC module/event/subsystems/iobuf/iobuf_rpc.o 00:04:50.393 CC module/event/subsystems/keyring/keyring.o 00:04:50.651 LIB libspdk_event_fsdev.a 00:04:50.651 LIB libspdk_event_scheduler.a 00:04:50.651 LIB libspdk_event_keyring.a 00:04:50.651 LIB libspdk_event_vhost_blk.a 00:04:50.651 LIB libspdk_event_vmd.a 00:04:50.651 LIB libspdk_event_iobuf.a 00:04:50.651 LIB libspdk_event_sock.a 00:04:50.651 SO libspdk_event_scheduler.so.4.0 00:04:50.651 SO libspdk_event_fsdev.so.1.0 00:04:50.651 SO libspdk_event_keyring.so.1.0 00:04:50.651 SO libspdk_event_vhost_blk.so.3.0 00:04:50.651 SO libspdk_event_vmd.so.6.0 00:04:50.651 SO libspdk_event_sock.so.5.0 00:04:50.651 SO libspdk_event_iobuf.so.3.0 00:04:50.651 SYMLINK libspdk_event_scheduler.so 00:04:50.651 SYMLINK libspdk_event_keyring.so 00:04:50.651 SYMLINK libspdk_event_vhost_blk.so 00:04:50.651 SYMLINK libspdk_event_fsdev.so 00:04:50.651 SYMLINK libspdk_event_sock.so 00:04:50.651 SYMLINK libspdk_event_vmd.so 00:04:50.651 SYMLINK libspdk_event_iobuf.so 00:04:50.909 CC module/event/subsystems/accel/accel.o 00:04:51.172 LIB libspdk_event_accel.a 00:04:51.172 SO libspdk_event_accel.so.6.0 00:04:51.172 SYMLINK libspdk_event_accel.so 00:04:51.740 CC module/event/subsystems/bdev/bdev.o 00:04:51.740 LIB libspdk_event_bdev.a 00:04:51.998 SO libspdk_event_bdev.so.6.0 00:04:51.998 SYMLINK libspdk_event_bdev.so 00:04:52.256 CC module/event/subsystems/scsi/scsi.o 00:04:52.256 CC module/event/subsystems/nvmf/nvmf_rpc.o 00:04:52.256 CC module/event/subsystems/nvmf/nvmf_tgt.o 00:04:52.256 CC module/event/subsystems/nbd/nbd.o 00:04:52.256 CC module/event/subsystems/ublk/ublk.o 00:04:52.515 LIB libspdk_event_scsi.a 00:04:52.515 LIB libspdk_event_ublk.a 00:04:52.515 LIB libspdk_event_nbd.a 00:04:52.515 SO libspdk_event_scsi.so.6.0 00:04:52.515 SO libspdk_event_nbd.so.6.0 00:04:52.515 SO libspdk_event_ublk.so.3.0 00:04:52.515 LIB libspdk_event_nvmf.a 00:04:52.515 SYMLINK libspdk_event_nbd.so 00:04:52.515 SYMLINK libspdk_event_ublk.so 00:04:52.515 SYMLINK libspdk_event_scsi.so 00:04:52.515 SO libspdk_event_nvmf.so.6.0 00:04:52.515 SYMLINK libspdk_event_nvmf.so 00:04:52.774 CC module/event/subsystems/vhost_scsi/vhost_scsi.o 00:04:52.774 CC module/event/subsystems/iscsi/iscsi.o 00:04:53.033 LIB libspdk_event_vhost_scsi.a 00:04:53.033 LIB libspdk_event_iscsi.a 00:04:53.033 SO libspdk_event_vhost_scsi.so.3.0 00:04:53.033 SO libspdk_event_iscsi.so.6.0 00:04:53.033 SYMLINK libspdk_event_vhost_scsi.so 00:04:53.290 SYMLINK libspdk_event_iscsi.so 00:04:53.290 SO libspdk.so.6.0 00:04:53.290 SYMLINK libspdk.so 00:04:53.548 TEST_HEADER include/spdk/accel.h 00:04:53.808 TEST_HEADER include/spdk/accel_module.h 00:04:53.808 TEST_HEADER include/spdk/assert.h 00:04:53.808 TEST_HEADER include/spdk/barrier.h 00:04:53.808 TEST_HEADER include/spdk/base64.h 00:04:53.808 CXX app/trace/trace.o 00:04:53.808 TEST_HEADER include/spdk/bdev.h 00:04:53.808 TEST_HEADER include/spdk/bdev_module.h 00:04:53.808 TEST_HEADER include/spdk/bdev_zone.h 00:04:53.808 CC test/rpc_client/rpc_client_test.o 00:04:53.808 TEST_HEADER include/spdk/bit_array.h 00:04:53.808 TEST_HEADER include/spdk/bit_pool.h 00:04:53.808 TEST_HEADER include/spdk/blob_bdev.h 00:04:53.808 TEST_HEADER include/spdk/blobfs_bdev.h 00:04:53.808 TEST_HEADER include/spdk/blobfs.h 00:04:53.808 TEST_HEADER include/spdk/blob.h 00:04:53.808 TEST_HEADER include/spdk/conf.h 00:04:53.808 TEST_HEADER include/spdk/config.h 00:04:53.808 CC examples/interrupt_tgt/interrupt_tgt.o 00:04:53.808 TEST_HEADER include/spdk/cpuset.h 00:04:53.808 TEST_HEADER include/spdk/crc16.h 00:04:53.808 TEST_HEADER include/spdk/crc32.h 00:04:53.808 TEST_HEADER include/spdk/crc64.h 00:04:53.808 TEST_HEADER include/spdk/dif.h 00:04:53.808 TEST_HEADER include/spdk/dma.h 00:04:53.808 TEST_HEADER include/spdk/endian.h 00:04:53.808 TEST_HEADER include/spdk/env_dpdk.h 00:04:53.808 TEST_HEADER include/spdk/env.h 00:04:53.808 TEST_HEADER include/spdk/event.h 00:04:53.808 TEST_HEADER include/spdk/fd_group.h 00:04:53.808 TEST_HEADER include/spdk/fd.h 00:04:53.808 TEST_HEADER include/spdk/file.h 00:04:53.808 TEST_HEADER include/spdk/fsdev.h 00:04:53.808 TEST_HEADER include/spdk/fsdev_module.h 00:04:53.808 TEST_HEADER include/spdk/ftl.h 00:04:53.808 TEST_HEADER include/spdk/fuse_dispatcher.h 00:04:53.808 TEST_HEADER include/spdk/gpt_spec.h 00:04:53.808 TEST_HEADER include/spdk/hexlify.h 00:04:53.808 TEST_HEADER include/spdk/histogram_data.h 00:04:53.808 TEST_HEADER include/spdk/idxd.h 00:04:53.808 CC examples/ioat/perf/perf.o 00:04:53.808 TEST_HEADER include/spdk/idxd_spec.h 00:04:53.808 CC test/thread/poller_perf/poller_perf.o 00:04:53.808 TEST_HEADER include/spdk/init.h 00:04:53.808 TEST_HEADER include/spdk/ioat.h 00:04:53.808 TEST_HEADER include/spdk/ioat_spec.h 00:04:53.808 TEST_HEADER include/spdk/iscsi_spec.h 00:04:53.808 TEST_HEADER include/spdk/json.h 00:04:53.808 TEST_HEADER include/spdk/jsonrpc.h 00:04:53.808 TEST_HEADER include/spdk/keyring.h 00:04:53.808 CC examples/util/zipf/zipf.o 00:04:53.808 TEST_HEADER include/spdk/keyring_module.h 00:04:53.808 TEST_HEADER include/spdk/likely.h 00:04:53.808 TEST_HEADER include/spdk/log.h 00:04:53.808 TEST_HEADER include/spdk/lvol.h 00:04:53.808 TEST_HEADER include/spdk/md5.h 00:04:53.808 TEST_HEADER include/spdk/memory.h 00:04:53.808 TEST_HEADER include/spdk/mmio.h 00:04:53.808 TEST_HEADER include/spdk/nbd.h 00:04:53.808 TEST_HEADER include/spdk/net.h 00:04:53.808 TEST_HEADER include/spdk/notify.h 00:04:53.808 TEST_HEADER include/spdk/nvme.h 00:04:53.808 TEST_HEADER include/spdk/nvme_intel.h 00:04:53.808 TEST_HEADER include/spdk/nvme_ocssd.h 00:04:53.808 CC test/dma/test_dma/test_dma.o 00:04:53.808 TEST_HEADER include/spdk/nvme_ocssd_spec.h 00:04:53.808 TEST_HEADER include/spdk/nvme_spec.h 00:04:53.808 TEST_HEADER include/spdk/nvme_zns.h 00:04:53.808 CC test/app/bdev_svc/bdev_svc.o 00:04:53.808 TEST_HEADER include/spdk/nvmf_cmd.h 00:04:53.808 TEST_HEADER include/spdk/nvmf_fc_spec.h 00:04:53.808 TEST_HEADER include/spdk/nvmf.h 00:04:53.808 TEST_HEADER include/spdk/nvmf_spec.h 00:04:53.808 TEST_HEADER include/spdk/nvmf_transport.h 00:04:53.808 TEST_HEADER include/spdk/opal.h 00:04:53.808 TEST_HEADER include/spdk/opal_spec.h 00:04:53.808 TEST_HEADER include/spdk/pci_ids.h 00:04:53.808 TEST_HEADER include/spdk/pipe.h 00:04:53.808 TEST_HEADER include/spdk/queue.h 00:04:53.808 TEST_HEADER include/spdk/reduce.h 00:04:53.808 TEST_HEADER include/spdk/rpc.h 00:04:53.808 TEST_HEADER include/spdk/scheduler.h 00:04:53.808 TEST_HEADER include/spdk/scsi.h 00:04:53.808 TEST_HEADER include/spdk/scsi_spec.h 00:04:53.808 TEST_HEADER include/spdk/sock.h 00:04:53.808 TEST_HEADER include/spdk/stdinc.h 00:04:53.808 TEST_HEADER include/spdk/string.h 00:04:53.808 TEST_HEADER include/spdk/thread.h 00:04:53.808 TEST_HEADER include/spdk/trace.h 00:04:53.808 TEST_HEADER include/spdk/trace_parser.h 00:04:53.808 TEST_HEADER include/spdk/tree.h 00:04:53.808 TEST_HEADER include/spdk/ublk.h 00:04:53.808 TEST_HEADER include/spdk/util.h 00:04:53.808 TEST_HEADER include/spdk/uuid.h 00:04:53.808 CC test/env/mem_callbacks/mem_callbacks.o 00:04:53.808 TEST_HEADER include/spdk/version.h 00:04:53.808 TEST_HEADER include/spdk/vfio_user_pci.h 00:04:53.808 TEST_HEADER include/spdk/vfio_user_spec.h 00:04:53.808 TEST_HEADER include/spdk/vhost.h 00:04:53.808 TEST_HEADER include/spdk/vmd.h 00:04:53.808 TEST_HEADER include/spdk/xor.h 00:04:53.808 TEST_HEADER include/spdk/zipf.h 00:04:53.808 CXX test/cpp_headers/accel.o 00:04:53.808 LINK rpc_client_test 00:04:53.808 LINK interrupt_tgt 00:04:53.808 LINK poller_perf 00:04:53.808 LINK zipf 00:04:54.068 LINK bdev_svc 00:04:54.068 LINK ioat_perf 00:04:54.068 CXX test/cpp_headers/accel_module.o 00:04:54.068 LINK spdk_trace 00:04:54.068 LINK mem_callbacks 00:04:54.068 CC app/trace_record/trace_record.o 00:04:54.068 CC examples/ioat/verify/verify.o 00:04:54.068 CC app/nvmf_tgt/nvmf_main.o 00:04:54.068 CXX test/cpp_headers/assert.o 00:04:54.068 CC test/env/vtophys/vtophys.o 00:04:54.328 CC test/app/fuzz/nvme_fuzz/nvme_fuzz.o 00:04:54.328 CC test/app/histogram_perf/histogram_perf.o 00:04:54.328 LINK nvmf_tgt 00:04:54.328 LINK test_dma 00:04:54.328 CC examples/thread/thread/thread_ex.o 00:04:54.328 CXX test/cpp_headers/barrier.o 00:04:54.328 CC test/app/fuzz/iscsi_fuzz/iscsi_fuzz.o 00:04:54.328 LINK verify 00:04:54.328 LINK spdk_trace_record 00:04:54.328 LINK vtophys 00:04:54.328 LINK histogram_perf 00:04:54.328 CXX test/cpp_headers/base64.o 00:04:54.587 CC test/app/fuzz/vhost_fuzz/vhost_fuzz_rpc.o 00:04:54.587 LINK thread 00:04:54.587 CC test/env/env_dpdk_post_init/env_dpdk_post_init.o 00:04:54.587 CXX test/cpp_headers/bdev.o 00:04:54.587 CC app/iscsi_tgt/iscsi_tgt.o 00:04:54.587 CC examples/sock/hello_world/hello_sock.o 00:04:54.587 CC examples/vmd/lsvmd/lsvmd.o 00:04:54.587 CC test/app/fuzz/vhost_fuzz/vhost_fuzz.o 00:04:54.587 CC examples/idxd/perf/perf.o 00:04:54.587 LINK nvme_fuzz 00:04:54.587 LINK env_dpdk_post_init 00:04:54.587 CXX test/cpp_headers/bdev_module.o 00:04:54.846 LINK lsvmd 00:04:54.846 LINK iscsi_tgt 00:04:54.846 LINK hello_sock 00:04:54.846 CXX test/cpp_headers/bdev_zone.o 00:04:54.846 CC examples/accel/perf/accel_perf.o 00:04:54.846 CC test/env/memory/memory_ut.o 00:04:54.846 CC examples/vmd/led/led.o 00:04:55.105 LINK idxd_perf 00:04:55.105 CC examples/blob/hello_world/hello_blob.o 00:04:55.105 LINK vhost_fuzz 00:04:55.105 CC app/spdk_tgt/spdk_tgt.o 00:04:55.105 CXX test/cpp_headers/bit_array.o 00:04:55.105 LINK led 00:04:55.105 CC examples/fsdev/hello_world/hello_fsdev.o 00:04:55.105 CXX test/cpp_headers/bit_pool.o 00:04:55.105 LINK hello_blob 00:04:55.105 LINK spdk_tgt 00:04:55.105 CC test/event/event_perf/event_perf.o 00:04:55.365 CC test/env/pci/pci_ut.o 00:04:55.365 CC test/app/jsoncat/jsoncat.o 00:04:55.365 CXX test/cpp_headers/blob_bdev.o 00:04:55.365 LINK event_perf 00:04:55.365 LINK accel_perf 00:04:55.365 LINK jsoncat 00:04:55.365 LINK hello_fsdev 00:04:55.365 CXX test/cpp_headers/blobfs_bdev.o 00:04:55.365 CC examples/blob/cli/blobcli.o 00:04:55.625 CC app/spdk_lspci/spdk_lspci.o 00:04:55.625 CC test/event/reactor/reactor.o 00:04:55.625 LINK memory_ut 00:04:55.625 CC app/spdk_nvme_perf/perf.o 00:04:55.625 CXX test/cpp_headers/blobfs.o 00:04:55.625 LINK spdk_lspci 00:04:55.625 CC test/event/reactor_perf/reactor_perf.o 00:04:55.625 LINK pci_ut 00:04:55.625 CXX test/cpp_headers/blob.o 00:04:55.625 LINK reactor 00:04:55.884 LINK reactor_perf 00:04:55.884 CXX test/cpp_headers/conf.o 00:04:55.884 CC test/app/stub/stub.o 00:04:55.884 LINK iscsi_fuzz 00:04:55.884 LINK blobcli 00:04:55.884 CXX test/cpp_headers/config.o 00:04:55.884 CC test/nvme/aer/aer.o 00:04:55.884 CXX test/cpp_headers/cpuset.o 00:04:55.884 CC test/event/app_repeat/app_repeat.o 00:04:55.884 CC test/accel/dif/dif.o 00:04:56.143 LINK stub 00:04:56.143 CC test/event/scheduler/scheduler.o 00:04:56.143 CC test/blobfs/mkfs/mkfs.o 00:04:56.143 CXX test/cpp_headers/crc16.o 00:04:56.143 LINK app_repeat 00:04:56.143 CXX test/cpp_headers/crc32.o 00:04:56.143 LINK mkfs 00:04:56.143 LINK aer 00:04:56.143 LINK scheduler 00:04:56.402 CC examples/nvme/hello_world/hello_world.o 00:04:56.402 CXX test/cpp_headers/crc64.o 00:04:56.402 CC examples/nvme/reconnect/reconnect.o 00:04:56.402 CC examples/bdev/hello_world/hello_bdev.o 00:04:56.402 LINK spdk_nvme_perf 00:04:56.402 CC test/nvme/reset/reset.o 00:04:56.402 CXX test/cpp_headers/dif.o 00:04:56.403 CC app/spdk_nvme_identify/identify.o 00:04:56.403 CC app/spdk_nvme_discover/discovery_aer.o 00:04:56.403 CC test/lvol/esnap/esnap.o 00:04:56.403 LINK hello_world 00:04:56.661 LINK hello_bdev 00:04:56.661 CXX test/cpp_headers/dma.o 00:04:56.661 LINK spdk_nvme_discover 00:04:56.661 CC examples/nvme/nvme_manage/nvme_manage.o 00:04:56.661 LINK reset 00:04:56.661 LINK dif 00:04:56.661 LINK reconnect 00:04:56.661 CXX test/cpp_headers/endian.o 00:04:56.921 CC app/spdk_top/spdk_top.o 00:04:56.921 CC examples/bdev/bdevperf/bdevperf.o 00:04:56.921 CXX test/cpp_headers/env_dpdk.o 00:04:56.921 CC test/nvme/sgl/sgl.o 00:04:56.921 CXX test/cpp_headers/env.o 00:04:56.921 CC app/vhost/vhost.o 00:04:56.921 CC examples/nvme/arbitration/arbitration.o 00:04:57.181 CXX test/cpp_headers/event.o 00:04:57.181 LINK vhost 00:04:57.181 LINK nvme_manage 00:04:57.181 LINK sgl 00:04:57.181 CXX test/cpp_headers/fd_group.o 00:04:57.181 CC test/bdev/bdevio/bdevio.o 00:04:57.181 LINK arbitration 00:04:57.181 CXX test/cpp_headers/fd.o 00:04:57.440 CXX test/cpp_headers/file.o 00:04:57.440 LINK spdk_nvme_identify 00:04:57.440 CC examples/nvme/hotplug/hotplug.o 00:04:57.440 CC test/nvme/e2edp/nvme_dp.o 00:04:57.440 CC test/nvme/overhead/overhead.o 00:04:57.440 CC test/nvme/err_injection/err_injection.o 00:04:57.440 CXX test/cpp_headers/fsdev.o 00:04:57.700 CC app/spdk_dd/spdk_dd.o 00:04:57.700 LINK bdevio 00:04:57.700 LINK err_injection 00:04:57.700 LINK hotplug 00:04:57.700 CXX test/cpp_headers/fsdev_module.o 00:04:57.700 LINK spdk_top 00:04:57.700 LINK nvme_dp 00:04:57.700 LINK overhead 00:04:57.700 LINK bdevperf 00:04:57.700 CXX test/cpp_headers/ftl.o 00:04:57.700 CXX test/cpp_headers/fuse_dispatcher.o 00:04:57.700 CXX test/cpp_headers/gpt_spec.o 00:04:57.962 CXX test/cpp_headers/hexlify.o 00:04:57.962 CC examples/nvme/cmb_copy/cmb_copy.o 00:04:57.962 CC test/nvme/startup/startup.o 00:04:57.962 CXX test/cpp_headers/histogram_data.o 00:04:57.962 LINK spdk_dd 00:04:57.962 CXX test/cpp_headers/idxd.o 00:04:57.962 LINK cmb_copy 00:04:57.962 CC test/nvme/simple_copy/simple_copy.o 00:04:57.962 CC test/nvme/reserve/reserve.o 00:04:57.962 LINK startup 00:04:57.962 CC test/nvme/connect_stress/connect_stress.o 00:04:58.219 CC app/fio/nvme/fio_plugin.o 00:04:58.219 CXX test/cpp_headers/idxd_spec.o 00:04:58.219 CC test/nvme/boot_partition/boot_partition.o 00:04:58.219 CC app/fio/bdev/fio_plugin.o 00:04:58.219 LINK connect_stress 00:04:58.219 LINK reserve 00:04:58.219 CC test/nvme/compliance/nvme_compliance.o 00:04:58.219 LINK simple_copy 00:04:58.219 CC examples/nvme/abort/abort.o 00:04:58.219 CXX test/cpp_headers/init.o 00:04:58.219 LINK boot_partition 00:04:58.478 CXX test/cpp_headers/ioat.o 00:04:58.478 CC test/nvme/fused_ordering/fused_ordering.o 00:04:58.478 CC test/nvme/doorbell_aers/doorbell_aers.o 00:04:58.478 CC examples/nvme/pmr_persistence/pmr_persistence.o 00:04:58.478 CC test/nvme/fdp/fdp.o 00:04:58.478 LINK nvme_compliance 00:04:58.478 CXX test/cpp_headers/ioat_spec.o 00:04:58.478 LINK abort 00:04:58.737 LINK spdk_nvme 00:04:58.737 LINK pmr_persistence 00:04:58.737 LINK doorbell_aers 00:04:58.737 LINK fused_ordering 00:04:58.737 LINK spdk_bdev 00:04:58.737 CXX test/cpp_headers/iscsi_spec.o 00:04:58.737 CXX test/cpp_headers/json.o 00:04:58.737 CXX test/cpp_headers/jsonrpc.o 00:04:58.737 CXX test/cpp_headers/keyring.o 00:04:58.737 CC test/nvme/cuse/cuse.o 00:04:58.737 CXX test/cpp_headers/keyring_module.o 00:04:58.737 CXX test/cpp_headers/likely.o 00:04:58.737 LINK fdp 00:04:58.737 CXX test/cpp_headers/log.o 00:04:58.737 CXX test/cpp_headers/lvol.o 00:04:58.995 CXX test/cpp_headers/md5.o 00:04:58.995 CXX test/cpp_headers/memory.o 00:04:58.995 CXX test/cpp_headers/mmio.o 00:04:58.995 CXX test/cpp_headers/nbd.o 00:04:58.995 CXX test/cpp_headers/net.o 00:04:58.995 CXX test/cpp_headers/notify.o 00:04:58.995 CC examples/nvmf/nvmf/nvmf.o 00:04:58.995 CXX test/cpp_headers/nvme.o 00:04:58.995 CXX test/cpp_headers/nvme_intel.o 00:04:58.995 CXX test/cpp_headers/nvme_ocssd.o 00:04:58.995 CXX test/cpp_headers/nvme_ocssd_spec.o 00:04:58.995 CXX test/cpp_headers/nvme_spec.o 00:04:58.995 CXX test/cpp_headers/nvme_zns.o 00:04:58.995 CXX test/cpp_headers/nvmf_cmd.o 00:04:58.995 CXX test/cpp_headers/nvmf_fc_spec.o 00:04:59.255 CXX test/cpp_headers/nvmf.o 00:04:59.255 CXX test/cpp_headers/nvmf_spec.o 00:04:59.255 CXX test/cpp_headers/nvmf_transport.o 00:04:59.255 CXX test/cpp_headers/opal.o 00:04:59.255 CXX test/cpp_headers/opal_spec.o 00:04:59.255 LINK nvmf 00:04:59.255 CXX test/cpp_headers/pci_ids.o 00:04:59.255 CXX test/cpp_headers/pipe.o 00:04:59.255 CXX test/cpp_headers/queue.o 00:04:59.255 CXX test/cpp_headers/reduce.o 00:04:59.255 CXX test/cpp_headers/rpc.o 00:04:59.255 CXX test/cpp_headers/scheduler.o 00:04:59.255 CXX test/cpp_headers/scsi.o 00:04:59.255 CXX test/cpp_headers/scsi_spec.o 00:04:59.514 CXX test/cpp_headers/sock.o 00:04:59.514 CXX test/cpp_headers/stdinc.o 00:04:59.514 CXX test/cpp_headers/string.o 00:04:59.514 CXX test/cpp_headers/thread.o 00:04:59.514 CXX test/cpp_headers/trace.o 00:04:59.514 CXX test/cpp_headers/trace_parser.o 00:04:59.514 CXX test/cpp_headers/tree.o 00:04:59.514 CXX test/cpp_headers/ublk.o 00:04:59.514 CXX test/cpp_headers/util.o 00:04:59.514 CXX test/cpp_headers/uuid.o 00:04:59.514 CXX test/cpp_headers/version.o 00:04:59.514 CXX test/cpp_headers/vfio_user_pci.o 00:04:59.514 CXX test/cpp_headers/vfio_user_spec.o 00:04:59.514 CXX test/cpp_headers/vhost.o 00:04:59.514 CXX test/cpp_headers/vmd.o 00:04:59.514 CXX test/cpp_headers/xor.o 00:04:59.774 CXX test/cpp_headers/zipf.o 00:05:00.033 LINK cuse 00:05:01.411 LINK esnap 00:05:01.980 00:05:01.980 real 1m11.669s 00:05:01.980 user 5m33.666s 00:05:01.980 sys 1m9.431s 00:05:01.980 17:26:32 make -- common/autotest_common.sh@1126 -- $ xtrace_disable 00:05:01.980 17:26:32 make -- common/autotest_common.sh@10 -- $ set +x 00:05:01.980 ************************************ 00:05:01.980 END TEST make 00:05:01.980 ************************************ 00:05:01.980 17:26:32 -- spdk/autobuild.sh@1 -- $ stop_monitor_resources 00:05:01.980 17:26:32 -- pm/common@29 -- $ signal_monitor_resources TERM 00:05:01.980 17:26:32 -- pm/common@40 -- $ local monitor pid pids signal=TERM 00:05:01.980 17:26:32 -- pm/common@42 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:05:01.980 17:26:32 -- pm/common@43 -- $ [[ -e /home/vagrant/spdk_repo/spdk/../output/power/collect-cpu-load.pid ]] 00:05:01.980 17:26:32 -- pm/common@44 -- $ pid=6198 00:05:01.980 17:26:32 -- pm/common@50 -- $ kill -TERM 6198 00:05:01.980 17:26:32 -- pm/common@42 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:05:01.980 17:26:32 -- pm/common@43 -- $ [[ -e /home/vagrant/spdk_repo/spdk/../output/power/collect-vmstat.pid ]] 00:05:01.980 17:26:32 -- pm/common@44 -- $ pid=6200 00:05:01.980 17:26:32 -- pm/common@50 -- $ kill -TERM 6200 00:05:01.980 17:26:33 -- common/autotest_common.sh@1680 -- # [[ y == y ]] 00:05:01.980 17:26:33 -- common/autotest_common.sh@1681 -- # lcov --version 00:05:01.980 17:26:33 -- common/autotest_common.sh@1681 -- # awk '{print $NF}' 00:05:01.980 17:26:33 -- common/autotest_common.sh@1681 -- # lt 1.15 2 00:05:01.980 17:26:33 -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:05:01.980 17:26:33 -- scripts/common.sh@333 -- # local ver1 ver1_l 00:05:01.980 17:26:33 -- scripts/common.sh@334 -- # local ver2 ver2_l 00:05:01.980 17:26:33 -- scripts/common.sh@336 -- # IFS=.-: 00:05:01.980 17:26:33 -- scripts/common.sh@336 -- # read -ra ver1 00:05:01.980 17:26:33 -- scripts/common.sh@337 -- # IFS=.-: 00:05:01.980 17:26:33 -- scripts/common.sh@337 -- # read -ra ver2 00:05:01.980 17:26:33 -- scripts/common.sh@338 -- # local 'op=<' 00:05:01.980 17:26:33 -- scripts/common.sh@340 -- # ver1_l=2 00:05:01.980 17:26:33 -- scripts/common.sh@341 -- # ver2_l=1 00:05:01.980 17:26:33 -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:05:01.980 17:26:33 -- scripts/common.sh@344 -- # case "$op" in 00:05:01.980 17:26:33 -- scripts/common.sh@345 -- # : 1 00:05:01.980 17:26:33 -- scripts/common.sh@364 -- # (( v = 0 )) 00:05:01.980 17:26:33 -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:05:01.980 17:26:33 -- scripts/common.sh@365 -- # decimal 1 00:05:01.980 17:26:33 -- scripts/common.sh@353 -- # local d=1 00:05:01.980 17:26:33 -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:05:01.980 17:26:33 -- scripts/common.sh@355 -- # echo 1 00:05:01.980 17:26:33 -- scripts/common.sh@365 -- # ver1[v]=1 00:05:01.980 17:26:33 -- scripts/common.sh@366 -- # decimal 2 00:05:01.980 17:26:33 -- scripts/common.sh@353 -- # local d=2 00:05:01.980 17:26:33 -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:05:01.980 17:26:33 -- scripts/common.sh@355 -- # echo 2 00:05:01.980 17:26:33 -- scripts/common.sh@366 -- # ver2[v]=2 00:05:01.980 17:26:33 -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:05:01.980 17:26:33 -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:05:01.980 17:26:33 -- scripts/common.sh@368 -- # return 0 00:05:01.980 17:26:33 -- common/autotest_common.sh@1682 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:05:01.980 17:26:33 -- common/autotest_common.sh@1694 -- # export 'LCOV_OPTS= 00:05:01.980 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:01.980 --rc genhtml_branch_coverage=1 00:05:01.980 --rc genhtml_function_coverage=1 00:05:01.980 --rc genhtml_legend=1 00:05:01.980 --rc geninfo_all_blocks=1 00:05:01.980 --rc geninfo_unexecuted_blocks=1 00:05:01.980 00:05:01.980 ' 00:05:01.980 17:26:33 -- common/autotest_common.sh@1694 -- # LCOV_OPTS=' 00:05:01.980 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:01.980 --rc genhtml_branch_coverage=1 00:05:01.980 --rc genhtml_function_coverage=1 00:05:01.980 --rc genhtml_legend=1 00:05:01.980 --rc geninfo_all_blocks=1 00:05:01.980 --rc geninfo_unexecuted_blocks=1 00:05:01.980 00:05:01.980 ' 00:05:01.980 17:26:33 -- common/autotest_common.sh@1695 -- # export 'LCOV=lcov 00:05:01.980 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:01.980 --rc genhtml_branch_coverage=1 00:05:01.980 --rc genhtml_function_coverage=1 00:05:01.980 --rc genhtml_legend=1 00:05:01.980 --rc geninfo_all_blocks=1 00:05:01.980 --rc geninfo_unexecuted_blocks=1 00:05:01.980 00:05:01.980 ' 00:05:01.980 17:26:33 -- common/autotest_common.sh@1695 -- # LCOV='lcov 00:05:01.980 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:01.980 --rc genhtml_branch_coverage=1 00:05:01.980 --rc genhtml_function_coverage=1 00:05:01.980 --rc genhtml_legend=1 00:05:01.980 --rc geninfo_all_blocks=1 00:05:01.980 --rc geninfo_unexecuted_blocks=1 00:05:01.980 00:05:01.980 ' 00:05:01.980 17:26:33 -- spdk/autotest.sh@25 -- # source /home/vagrant/spdk_repo/spdk/test/nvmf/common.sh 00:05:01.980 17:26:33 -- nvmf/common.sh@7 -- # uname -s 00:05:02.240 17:26:33 -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:05:02.240 17:26:33 -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:05:02.240 17:26:33 -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:05:02.240 17:26:33 -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:05:02.240 17:26:33 -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:05:02.240 17:26:33 -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:05:02.240 17:26:33 -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:05:02.240 17:26:33 -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:05:02.240 17:26:33 -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:05:02.240 17:26:33 -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:05:02.240 17:26:33 -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:8044f8d2-4aeb-4cab-84e2-f73e0a4751e5 00:05:02.240 17:26:33 -- nvmf/common.sh@18 -- # NVME_HOSTID=8044f8d2-4aeb-4cab-84e2-f73e0a4751e5 00:05:02.240 17:26:33 -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:05:02.240 17:26:33 -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:05:02.240 17:26:33 -- nvmf/common.sh@21 -- # NET_TYPE=phy-fallback 00:05:02.240 17:26:33 -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:05:02.240 17:26:33 -- nvmf/common.sh@49 -- # source /home/vagrant/spdk_repo/spdk/scripts/common.sh 00:05:02.240 17:26:33 -- scripts/common.sh@15 -- # shopt -s extglob 00:05:02.240 17:26:33 -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:05:02.240 17:26:33 -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:05:02.240 17:26:33 -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:05:02.240 17:26:33 -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:05:02.240 17:26:33 -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:05:02.240 17:26:33 -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:05:02.240 17:26:33 -- paths/export.sh@5 -- # export PATH 00:05:02.240 17:26:33 -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:05:02.240 17:26:33 -- nvmf/common.sh@51 -- # : 0 00:05:02.240 17:26:33 -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:05:02.240 17:26:33 -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:05:02.240 17:26:33 -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:05:02.240 17:26:33 -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:05:02.240 17:26:33 -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:05:02.240 17:26:33 -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:05:02.240 /home/vagrant/spdk_repo/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:05:02.240 17:26:33 -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:05:02.240 17:26:33 -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:05:02.240 17:26:33 -- nvmf/common.sh@55 -- # have_pci_nics=0 00:05:02.240 17:26:33 -- spdk/autotest.sh@27 -- # '[' 0 -ne 0 ']' 00:05:02.240 17:26:33 -- spdk/autotest.sh@32 -- # uname -s 00:05:02.240 17:26:33 -- spdk/autotest.sh@32 -- # '[' Linux = Linux ']' 00:05:02.240 17:26:33 -- spdk/autotest.sh@33 -- # old_core_pattern='|/usr/lib/systemd/systemd-coredump %P %u %g %s %t %c %h' 00:05:02.240 17:26:33 -- spdk/autotest.sh@34 -- # mkdir -p /home/vagrant/spdk_repo/spdk/../output/coredumps 00:05:02.240 17:26:33 -- spdk/autotest.sh@39 -- # echo '|/home/vagrant/spdk_repo/spdk/scripts/core-collector.sh %P %s %t' 00:05:02.240 17:26:33 -- spdk/autotest.sh@40 -- # echo /home/vagrant/spdk_repo/spdk/../output/coredumps 00:05:02.240 17:26:33 -- spdk/autotest.sh@44 -- # modprobe nbd 00:05:02.240 17:26:33 -- spdk/autotest.sh@46 -- # type -P udevadm 00:05:02.240 17:26:33 -- spdk/autotest.sh@46 -- # udevadm=/usr/sbin/udevadm 00:05:02.240 17:26:33 -- spdk/autotest.sh@48 -- # udevadm_pid=66423 00:05:02.240 17:26:33 -- spdk/autotest.sh@53 -- # start_monitor_resources 00:05:02.240 17:26:33 -- pm/common@17 -- # local monitor 00:05:02.240 17:26:33 -- pm/common@19 -- # for monitor in "${MONITOR_RESOURCES[@]}" 00:05:02.240 17:26:33 -- pm/common@19 -- # for monitor in "${MONITOR_RESOURCES[@]}" 00:05:02.240 17:26:33 -- spdk/autotest.sh@47 -- # /usr/sbin/udevadm monitor --property 00:05:02.240 17:26:33 -- pm/common@25 -- # sleep 1 00:05:02.240 17:26:33 -- pm/common@21 -- # date +%s 00:05:02.240 17:26:33 -- pm/common@21 -- # /home/vagrant/spdk_repo/spdk/scripts/perf/pm/collect-vmstat -d /home/vagrant/spdk_repo/spdk/../output/power -l -p monitor.autotest.sh.1732728393 00:05:02.240 17:26:33 -- pm/common@21 -- # date +%s 00:05:02.240 17:26:33 -- pm/common@21 -- # /home/vagrant/spdk_repo/spdk/scripts/perf/pm/collect-cpu-load -d /home/vagrant/spdk_repo/spdk/../output/power -l -p monitor.autotest.sh.1732728393 00:05:02.240 Redirecting to /home/vagrant/spdk_repo/spdk/../output/power/monitor.autotest.sh.1732728393_collect-cpu-load.pm.log 00:05:02.240 Redirecting to /home/vagrant/spdk_repo/spdk/../output/power/monitor.autotest.sh.1732728393_collect-vmstat.pm.log 00:05:03.179 17:26:34 -- spdk/autotest.sh@55 -- # trap 'autotest_cleanup || :; exit 1' SIGINT SIGTERM EXIT 00:05:03.179 17:26:34 -- spdk/autotest.sh@57 -- # timing_enter autotest 00:05:03.179 17:26:34 -- common/autotest_common.sh@724 -- # xtrace_disable 00:05:03.179 17:26:34 -- common/autotest_common.sh@10 -- # set +x 00:05:03.179 17:26:34 -- spdk/autotest.sh@59 -- # create_test_list 00:05:03.179 17:26:34 -- common/autotest_common.sh@748 -- # xtrace_disable 00:05:03.179 17:26:34 -- common/autotest_common.sh@10 -- # set +x 00:05:03.179 17:26:34 -- spdk/autotest.sh@61 -- # dirname /home/vagrant/spdk_repo/spdk/autotest.sh 00:05:03.179 17:26:34 -- spdk/autotest.sh@61 -- # readlink -f /home/vagrant/spdk_repo/spdk 00:05:03.179 17:26:34 -- spdk/autotest.sh@61 -- # src=/home/vagrant/spdk_repo/spdk 00:05:03.179 17:26:34 -- spdk/autotest.sh@62 -- # out=/home/vagrant/spdk_repo/spdk/../output 00:05:03.179 17:26:34 -- spdk/autotest.sh@63 -- # cd /home/vagrant/spdk_repo/spdk 00:05:03.179 17:26:34 -- spdk/autotest.sh@65 -- # freebsd_update_contigmem_mod 00:05:03.179 17:26:34 -- common/autotest_common.sh@1455 -- # uname 00:05:03.179 17:26:34 -- common/autotest_common.sh@1455 -- # '[' Linux = FreeBSD ']' 00:05:03.179 17:26:34 -- spdk/autotest.sh@66 -- # freebsd_set_maxsock_buf 00:05:03.179 17:26:34 -- common/autotest_common.sh@1475 -- # uname 00:05:03.438 17:26:34 -- common/autotest_common.sh@1475 -- # [[ Linux = FreeBSD ]] 00:05:03.438 17:26:34 -- spdk/autotest.sh@68 -- # [[ y == y ]] 00:05:03.438 17:26:34 -- spdk/autotest.sh@70 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 --version 00:05:03.438 lcov: LCOV version 1.15 00:05:03.438 17:26:34 -- spdk/autotest.sh@72 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -c --no-external -i -t Baseline -d /home/vagrant/spdk_repo/spdk -o /home/vagrant/spdk_repo/spdk/../output/cov_base.info 00:05:18.329 /home/vagrant/spdk_repo/spdk/lib/nvme/nvme_stubs.gcno:no functions found 00:05:18.329 geninfo: WARNING: GCOV did not produce any data for /home/vagrant/spdk_repo/spdk/lib/nvme/nvme_stubs.gcno 00:05:33.217 17:27:03 -- spdk/autotest.sh@76 -- # timing_enter pre_cleanup 00:05:33.217 17:27:03 -- common/autotest_common.sh@724 -- # xtrace_disable 00:05:33.217 17:27:03 -- common/autotest_common.sh@10 -- # set +x 00:05:33.217 17:27:03 -- spdk/autotest.sh@78 -- # rm -f 00:05:33.217 17:27:03 -- spdk/autotest.sh@81 -- # /home/vagrant/spdk_repo/spdk/scripts/setup.sh reset 00:05:33.477 0000:00:03.0 (1af4 1001): Active devices: mount@vda:vda2,mount@vda:vda3,mount@vda:vda5, so not binding PCI dev 00:05:33.477 0000:00:11.0 (1b36 0010): Already using the nvme driver 00:05:33.477 0000:00:10.0 (1b36 0010): Already using the nvme driver 00:05:33.477 17:27:04 -- spdk/autotest.sh@83 -- # get_zoned_devs 00:05:33.477 17:27:04 -- common/autotest_common.sh@1655 -- # zoned_devs=() 00:05:33.477 17:27:04 -- common/autotest_common.sh@1655 -- # local -gA zoned_devs 00:05:33.477 17:27:04 -- common/autotest_common.sh@1656 -- # local nvme bdf 00:05:33.477 17:27:04 -- common/autotest_common.sh@1658 -- # for nvme in /sys/block/nvme* 00:05:33.477 17:27:04 -- common/autotest_common.sh@1659 -- # is_block_zoned nvme0n1 00:05:33.477 17:27:04 -- common/autotest_common.sh@1648 -- # local device=nvme0n1 00:05:33.477 17:27:04 -- common/autotest_common.sh@1650 -- # [[ -e /sys/block/nvme0n1/queue/zoned ]] 00:05:33.477 17:27:04 -- common/autotest_common.sh@1651 -- # [[ none != none ]] 00:05:33.477 17:27:04 -- common/autotest_common.sh@1658 -- # for nvme in /sys/block/nvme* 00:05:33.477 17:27:04 -- common/autotest_common.sh@1659 -- # is_block_zoned nvme1n1 00:05:33.477 17:27:04 -- common/autotest_common.sh@1648 -- # local device=nvme1n1 00:05:33.477 17:27:04 -- common/autotest_common.sh@1650 -- # [[ -e /sys/block/nvme1n1/queue/zoned ]] 00:05:33.477 17:27:04 -- common/autotest_common.sh@1651 -- # [[ none != none ]] 00:05:33.477 17:27:04 -- common/autotest_common.sh@1658 -- # for nvme in /sys/block/nvme* 00:05:33.477 17:27:04 -- common/autotest_common.sh@1659 -- # is_block_zoned nvme1n2 00:05:33.477 17:27:04 -- common/autotest_common.sh@1648 -- # local device=nvme1n2 00:05:33.477 17:27:04 -- common/autotest_common.sh@1650 -- # [[ -e /sys/block/nvme1n2/queue/zoned ]] 00:05:33.477 17:27:04 -- common/autotest_common.sh@1651 -- # [[ none != none ]] 00:05:33.477 17:27:04 -- common/autotest_common.sh@1658 -- # for nvme in /sys/block/nvme* 00:05:33.477 17:27:04 -- common/autotest_common.sh@1659 -- # is_block_zoned nvme1n3 00:05:33.477 17:27:04 -- common/autotest_common.sh@1648 -- # local device=nvme1n3 00:05:33.477 17:27:04 -- common/autotest_common.sh@1650 -- # [[ -e /sys/block/nvme1n3/queue/zoned ]] 00:05:33.477 17:27:04 -- common/autotest_common.sh@1651 -- # [[ none != none ]] 00:05:33.477 17:27:04 -- spdk/autotest.sh@85 -- # (( 0 > 0 )) 00:05:33.477 17:27:04 -- spdk/autotest.sh@97 -- # for dev in /dev/nvme*n!(*p*) 00:05:33.477 17:27:04 -- spdk/autotest.sh@99 -- # [[ -z '' ]] 00:05:33.477 17:27:04 -- spdk/autotest.sh@100 -- # block_in_use /dev/nvme0n1 00:05:33.477 17:27:04 -- scripts/common.sh@381 -- # local block=/dev/nvme0n1 pt 00:05:33.477 17:27:04 -- scripts/common.sh@390 -- # /home/vagrant/spdk_repo/spdk/scripts/spdk-gpt.py /dev/nvme0n1 00:05:33.737 No valid GPT data, bailing 00:05:33.737 17:27:04 -- scripts/common.sh@394 -- # blkid -s PTTYPE -o value /dev/nvme0n1 00:05:33.737 17:27:04 -- scripts/common.sh@394 -- # pt= 00:05:33.737 17:27:04 -- scripts/common.sh@395 -- # return 1 00:05:33.737 17:27:04 -- spdk/autotest.sh@101 -- # dd if=/dev/zero of=/dev/nvme0n1 bs=1M count=1 00:05:33.737 1+0 records in 00:05:33.737 1+0 records out 00:05:33.737 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.00430437 s, 244 MB/s 00:05:33.737 17:27:04 -- spdk/autotest.sh@97 -- # for dev in /dev/nvme*n!(*p*) 00:05:33.737 17:27:04 -- spdk/autotest.sh@99 -- # [[ -z '' ]] 00:05:33.737 17:27:04 -- spdk/autotest.sh@100 -- # block_in_use /dev/nvme1n1 00:05:33.737 17:27:04 -- scripts/common.sh@381 -- # local block=/dev/nvme1n1 pt 00:05:33.737 17:27:04 -- scripts/common.sh@390 -- # /home/vagrant/spdk_repo/spdk/scripts/spdk-gpt.py /dev/nvme1n1 00:05:33.737 No valid GPT data, bailing 00:05:33.737 17:27:04 -- scripts/common.sh@394 -- # blkid -s PTTYPE -o value /dev/nvme1n1 00:05:33.737 17:27:04 -- scripts/common.sh@394 -- # pt= 00:05:33.737 17:27:04 -- scripts/common.sh@395 -- # return 1 00:05:33.737 17:27:04 -- spdk/autotest.sh@101 -- # dd if=/dev/zero of=/dev/nvme1n1 bs=1M count=1 00:05:33.737 1+0 records in 00:05:33.737 1+0 records out 00:05:33.737 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0040681 s, 258 MB/s 00:05:33.737 17:27:04 -- spdk/autotest.sh@97 -- # for dev in /dev/nvme*n!(*p*) 00:05:33.737 17:27:04 -- spdk/autotest.sh@99 -- # [[ -z '' ]] 00:05:33.737 17:27:04 -- spdk/autotest.sh@100 -- # block_in_use /dev/nvme1n2 00:05:33.737 17:27:04 -- scripts/common.sh@381 -- # local block=/dev/nvme1n2 pt 00:05:33.737 17:27:04 -- scripts/common.sh@390 -- # /home/vagrant/spdk_repo/spdk/scripts/spdk-gpt.py /dev/nvme1n2 00:05:33.737 No valid GPT data, bailing 00:05:33.737 17:27:04 -- scripts/common.sh@394 -- # blkid -s PTTYPE -o value /dev/nvme1n2 00:05:33.737 17:27:04 -- scripts/common.sh@394 -- # pt= 00:05:33.737 17:27:04 -- scripts/common.sh@395 -- # return 1 00:05:33.737 17:27:04 -- spdk/autotest.sh@101 -- # dd if=/dev/zero of=/dev/nvme1n2 bs=1M count=1 00:05:33.737 1+0 records in 00:05:33.737 1+0 records out 00:05:33.737 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.00622035 s, 169 MB/s 00:05:33.737 17:27:04 -- spdk/autotest.sh@97 -- # for dev in /dev/nvme*n!(*p*) 00:05:33.737 17:27:04 -- spdk/autotest.sh@99 -- # [[ -z '' ]] 00:05:33.737 17:27:04 -- spdk/autotest.sh@100 -- # block_in_use /dev/nvme1n3 00:05:33.737 17:27:04 -- scripts/common.sh@381 -- # local block=/dev/nvme1n3 pt 00:05:33.737 17:27:04 -- scripts/common.sh@390 -- # /home/vagrant/spdk_repo/spdk/scripts/spdk-gpt.py /dev/nvme1n3 00:05:33.996 No valid GPT data, bailing 00:05:33.996 17:27:04 -- scripts/common.sh@394 -- # blkid -s PTTYPE -o value /dev/nvme1n3 00:05:33.996 17:27:04 -- scripts/common.sh@394 -- # pt= 00:05:33.996 17:27:04 -- scripts/common.sh@395 -- # return 1 00:05:33.996 17:27:04 -- spdk/autotest.sh@101 -- # dd if=/dev/zero of=/dev/nvme1n3 bs=1M count=1 00:05:33.996 1+0 records in 00:05:33.996 1+0 records out 00:05:33.996 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.00627409 s, 167 MB/s 00:05:33.996 17:27:04 -- spdk/autotest.sh@105 -- # sync 00:05:34.255 17:27:05 -- spdk/autotest.sh@107 -- # xtrace_disable_per_cmd reap_spdk_processes 00:05:34.255 17:27:05 -- common/autotest_common.sh@22 -- # eval 'reap_spdk_processes 12> /dev/null' 00:05:34.255 17:27:05 -- common/autotest_common.sh@22 -- # reap_spdk_processes 00:05:37.550 17:27:08 -- spdk/autotest.sh@111 -- # uname -s 00:05:37.550 17:27:08 -- spdk/autotest.sh@111 -- # [[ Linux == Linux ]] 00:05:37.550 17:27:08 -- spdk/autotest.sh@111 -- # [[ 0 -eq 1 ]] 00:05:37.550 17:27:08 -- spdk/autotest.sh@115 -- # /home/vagrant/spdk_repo/spdk/scripts/setup.sh status 00:05:38.121 0000:00:03.0 (1af4 1001): Active devices: mount@vda:vda2,mount@vda:vda3,mount@vda:vda5, so not binding PCI dev 00:05:38.121 Hugepages 00:05:38.121 node hugesize free / total 00:05:38.121 node0 1048576kB 0 / 0 00:05:38.121 node0 2048kB 0 / 0 00:05:38.121 00:05:38.121 Type BDF Vendor Device NUMA Driver Device Block devices 00:05:38.121 virtio 0000:00:03.0 1af4 1001 unknown virtio-pci - vda 00:05:38.121 NVMe 0000:00:10.0 1b36 0010 unknown nvme nvme0 nvme0n1 00:05:38.381 NVMe 0000:00:11.0 1b36 0010 unknown nvme nvme1 nvme1n1 nvme1n2 nvme1n3 00:05:38.381 17:27:09 -- spdk/autotest.sh@117 -- # uname -s 00:05:38.381 17:27:09 -- spdk/autotest.sh@117 -- # [[ Linux == Linux ]] 00:05:38.381 17:27:09 -- spdk/autotest.sh@119 -- # nvme_namespace_revert 00:05:38.381 17:27:09 -- common/autotest_common.sh@1514 -- # /home/vagrant/spdk_repo/spdk/scripts/setup.sh 00:05:39.322 0000:00:03.0 (1af4 1001): Active devices: mount@vda:vda2,mount@vda:vda3,mount@vda:vda5, so not binding PCI dev 00:05:39.322 0000:00:10.0 (1b36 0010): nvme -> uio_pci_generic 00:05:39.322 0000:00:11.0 (1b36 0010): nvme -> uio_pci_generic 00:05:39.322 17:27:10 -- common/autotest_common.sh@1515 -- # sleep 1 00:05:40.261 17:27:11 -- common/autotest_common.sh@1516 -- # bdfs=() 00:05:40.262 17:27:11 -- common/autotest_common.sh@1516 -- # local bdfs 00:05:40.262 17:27:11 -- common/autotest_common.sh@1518 -- # bdfs=($(get_nvme_bdfs)) 00:05:40.262 17:27:11 -- common/autotest_common.sh@1518 -- # get_nvme_bdfs 00:05:40.262 17:27:11 -- common/autotest_common.sh@1496 -- # bdfs=() 00:05:40.262 17:27:11 -- common/autotest_common.sh@1496 -- # local bdfs 00:05:40.262 17:27:11 -- common/autotest_common.sh@1497 -- # bdfs=($("$rootdir/scripts/gen_nvme.sh" | jq -r '.config[].params.traddr')) 00:05:40.262 17:27:11 -- common/autotest_common.sh@1497 -- # /home/vagrant/spdk_repo/spdk/scripts/gen_nvme.sh 00:05:40.262 17:27:11 -- common/autotest_common.sh@1497 -- # jq -r '.config[].params.traddr' 00:05:40.521 17:27:11 -- common/autotest_common.sh@1498 -- # (( 2 == 0 )) 00:05:40.521 17:27:11 -- common/autotest_common.sh@1502 -- # printf '%s\n' 0000:00:10.0 0000:00:11.0 00:05:40.521 17:27:11 -- common/autotest_common.sh@1520 -- # /home/vagrant/spdk_repo/spdk/scripts/setup.sh reset 00:05:40.781 0000:00:03.0 (1af4 1001): Active devices: mount@vda:vda2,mount@vda:vda3,mount@vda:vda5, so not binding PCI dev 00:05:40.781 Waiting for block devices as requested 00:05:41.041 0000:00:11.0 (1b36 0010): uio_pci_generic -> nvme 00:05:41.041 0000:00:10.0 (1b36 0010): uio_pci_generic -> nvme 00:05:41.041 17:27:12 -- common/autotest_common.sh@1522 -- # for bdf in "${bdfs[@]}" 00:05:41.041 17:27:12 -- common/autotest_common.sh@1523 -- # get_nvme_ctrlr_from_bdf 0000:00:10.0 00:05:41.041 17:27:12 -- common/autotest_common.sh@1485 -- # readlink -f /sys/class/nvme/nvme0 /sys/class/nvme/nvme1 00:05:41.041 17:27:12 -- common/autotest_common.sh@1485 -- # grep 0000:00:10.0/nvme/nvme 00:05:41.041 17:27:12 -- common/autotest_common.sh@1485 -- # bdf_sysfs_path=/sys/devices/pci0000:00/0000:00:10.0/nvme/nvme1 00:05:41.041 17:27:12 -- common/autotest_common.sh@1486 -- # [[ -z /sys/devices/pci0000:00/0000:00:10.0/nvme/nvme1 ]] 00:05:41.041 17:27:12 -- common/autotest_common.sh@1490 -- # basename /sys/devices/pci0000:00/0000:00:10.0/nvme/nvme1 00:05:41.302 17:27:12 -- common/autotest_common.sh@1490 -- # printf '%s\n' nvme1 00:05:41.302 17:27:12 -- common/autotest_common.sh@1523 -- # nvme_ctrlr=/dev/nvme1 00:05:41.302 17:27:12 -- common/autotest_common.sh@1524 -- # [[ -z /dev/nvme1 ]] 00:05:41.302 17:27:12 -- common/autotest_common.sh@1529 -- # cut -d: -f2 00:05:41.302 17:27:12 -- common/autotest_common.sh@1529 -- # nvme id-ctrl /dev/nvme1 00:05:41.302 17:27:12 -- common/autotest_common.sh@1529 -- # grep oacs 00:05:41.302 17:27:12 -- common/autotest_common.sh@1529 -- # oacs=' 0x12a' 00:05:41.302 17:27:12 -- common/autotest_common.sh@1530 -- # oacs_ns_manage=8 00:05:41.302 17:27:12 -- common/autotest_common.sh@1532 -- # [[ 8 -ne 0 ]] 00:05:41.302 17:27:12 -- common/autotest_common.sh@1538 -- # nvme id-ctrl /dev/nvme1 00:05:41.302 17:27:12 -- common/autotest_common.sh@1538 -- # grep unvmcap 00:05:41.302 17:27:12 -- common/autotest_common.sh@1538 -- # cut -d: -f2 00:05:41.302 17:27:12 -- common/autotest_common.sh@1538 -- # unvmcap=' 0' 00:05:41.302 17:27:12 -- common/autotest_common.sh@1539 -- # [[ 0 -eq 0 ]] 00:05:41.302 17:27:12 -- common/autotest_common.sh@1541 -- # continue 00:05:41.302 17:27:12 -- common/autotest_common.sh@1522 -- # for bdf in "${bdfs[@]}" 00:05:41.302 17:27:12 -- common/autotest_common.sh@1523 -- # get_nvme_ctrlr_from_bdf 0000:00:11.0 00:05:41.302 17:27:12 -- common/autotest_common.sh@1485 -- # readlink -f /sys/class/nvme/nvme0 /sys/class/nvme/nvme1 00:05:41.302 17:27:12 -- common/autotest_common.sh@1485 -- # grep 0000:00:11.0/nvme/nvme 00:05:41.302 17:27:12 -- common/autotest_common.sh@1485 -- # bdf_sysfs_path=/sys/devices/pci0000:00/0000:00:11.0/nvme/nvme0 00:05:41.302 17:27:12 -- common/autotest_common.sh@1486 -- # [[ -z /sys/devices/pci0000:00/0000:00:11.0/nvme/nvme0 ]] 00:05:41.302 17:27:12 -- common/autotest_common.sh@1490 -- # basename /sys/devices/pci0000:00/0000:00:11.0/nvme/nvme0 00:05:41.302 17:27:12 -- common/autotest_common.sh@1490 -- # printf '%s\n' nvme0 00:05:41.302 17:27:12 -- common/autotest_common.sh@1523 -- # nvme_ctrlr=/dev/nvme0 00:05:41.302 17:27:12 -- common/autotest_common.sh@1524 -- # [[ -z /dev/nvme0 ]] 00:05:41.302 17:27:12 -- common/autotest_common.sh@1529 -- # cut -d: -f2 00:05:41.302 17:27:12 -- common/autotest_common.sh@1529 -- # nvme id-ctrl /dev/nvme0 00:05:41.302 17:27:12 -- common/autotest_common.sh@1529 -- # grep oacs 00:05:41.302 17:27:12 -- common/autotest_common.sh@1529 -- # oacs=' 0x12a' 00:05:41.302 17:27:12 -- common/autotest_common.sh@1530 -- # oacs_ns_manage=8 00:05:41.302 17:27:12 -- common/autotest_common.sh@1532 -- # [[ 8 -ne 0 ]] 00:05:41.302 17:27:12 -- common/autotest_common.sh@1538 -- # nvme id-ctrl /dev/nvme0 00:05:41.302 17:27:12 -- common/autotest_common.sh@1538 -- # grep unvmcap 00:05:41.302 17:27:12 -- common/autotest_common.sh@1538 -- # cut -d: -f2 00:05:41.302 17:27:12 -- common/autotest_common.sh@1538 -- # unvmcap=' 0' 00:05:41.302 17:27:12 -- common/autotest_common.sh@1539 -- # [[ 0 -eq 0 ]] 00:05:41.302 17:27:12 -- common/autotest_common.sh@1541 -- # continue 00:05:41.302 17:27:12 -- spdk/autotest.sh@122 -- # timing_exit pre_cleanup 00:05:41.302 17:27:12 -- common/autotest_common.sh@730 -- # xtrace_disable 00:05:41.302 17:27:12 -- common/autotest_common.sh@10 -- # set +x 00:05:41.302 17:27:12 -- spdk/autotest.sh@125 -- # timing_enter afterboot 00:05:41.302 17:27:12 -- common/autotest_common.sh@724 -- # xtrace_disable 00:05:41.302 17:27:12 -- common/autotest_common.sh@10 -- # set +x 00:05:41.302 17:27:12 -- spdk/autotest.sh@126 -- # /home/vagrant/spdk_repo/spdk/scripts/setup.sh 00:05:42.260 0000:00:03.0 (1af4 1001): Active devices: mount@vda:vda2,mount@vda:vda3,mount@vda:vda5, so not binding PCI dev 00:05:42.260 0000:00:10.0 (1b36 0010): nvme -> uio_pci_generic 00:05:42.260 0000:00:11.0 (1b36 0010): nvme -> uio_pci_generic 00:05:42.260 17:27:13 -- spdk/autotest.sh@127 -- # timing_exit afterboot 00:05:42.260 17:27:13 -- common/autotest_common.sh@730 -- # xtrace_disable 00:05:42.260 17:27:13 -- common/autotest_common.sh@10 -- # set +x 00:05:42.260 17:27:13 -- spdk/autotest.sh@131 -- # opal_revert_cleanup 00:05:42.260 17:27:13 -- common/autotest_common.sh@1576 -- # mapfile -t bdfs 00:05:42.260 17:27:13 -- common/autotest_common.sh@1576 -- # get_nvme_bdfs_by_id 0x0a54 00:05:42.260 17:27:13 -- common/autotest_common.sh@1561 -- # bdfs=() 00:05:42.260 17:27:13 -- common/autotest_common.sh@1561 -- # _bdfs=() 00:05:42.260 17:27:13 -- common/autotest_common.sh@1561 -- # local bdfs _bdfs 00:05:42.260 17:27:13 -- common/autotest_common.sh@1562 -- # _bdfs=($(get_nvme_bdfs)) 00:05:42.260 17:27:13 -- common/autotest_common.sh@1562 -- # get_nvme_bdfs 00:05:42.260 17:27:13 -- common/autotest_common.sh@1496 -- # bdfs=() 00:05:42.260 17:27:13 -- common/autotest_common.sh@1496 -- # local bdfs 00:05:42.260 17:27:13 -- common/autotest_common.sh@1497 -- # bdfs=($("$rootdir/scripts/gen_nvme.sh" | jq -r '.config[].params.traddr')) 00:05:42.260 17:27:13 -- common/autotest_common.sh@1497 -- # /home/vagrant/spdk_repo/spdk/scripts/gen_nvme.sh 00:05:42.260 17:27:13 -- common/autotest_common.sh@1497 -- # jq -r '.config[].params.traddr' 00:05:42.520 17:27:13 -- common/autotest_common.sh@1498 -- # (( 2 == 0 )) 00:05:42.520 17:27:13 -- common/autotest_common.sh@1502 -- # printf '%s\n' 0000:00:10.0 0000:00:11.0 00:05:42.520 17:27:13 -- common/autotest_common.sh@1563 -- # for bdf in "${_bdfs[@]}" 00:05:42.520 17:27:13 -- common/autotest_common.sh@1564 -- # cat /sys/bus/pci/devices/0000:00:10.0/device 00:05:42.520 17:27:13 -- common/autotest_common.sh@1564 -- # device=0x0010 00:05:42.520 17:27:13 -- common/autotest_common.sh@1565 -- # [[ 0x0010 == \0\x\0\a\5\4 ]] 00:05:42.520 17:27:13 -- common/autotest_common.sh@1563 -- # for bdf in "${_bdfs[@]}" 00:05:42.520 17:27:13 -- common/autotest_common.sh@1564 -- # cat /sys/bus/pci/devices/0000:00:11.0/device 00:05:42.520 17:27:13 -- common/autotest_common.sh@1564 -- # device=0x0010 00:05:42.520 17:27:13 -- common/autotest_common.sh@1565 -- # [[ 0x0010 == \0\x\0\a\5\4 ]] 00:05:42.520 17:27:13 -- common/autotest_common.sh@1570 -- # (( 0 > 0 )) 00:05:42.520 17:27:13 -- common/autotest_common.sh@1570 -- # return 0 00:05:42.520 17:27:13 -- common/autotest_common.sh@1577 -- # [[ -z '' ]] 00:05:42.520 17:27:13 -- common/autotest_common.sh@1578 -- # return 0 00:05:42.520 17:27:13 -- spdk/autotest.sh@137 -- # '[' 0 -eq 1 ']' 00:05:42.520 17:27:13 -- spdk/autotest.sh@141 -- # '[' 1 -eq 1 ']' 00:05:42.520 17:27:13 -- spdk/autotest.sh@142 -- # [[ 0 -eq 1 ]] 00:05:42.520 17:27:13 -- spdk/autotest.sh@142 -- # [[ 0 -eq 1 ]] 00:05:42.520 17:27:13 -- spdk/autotest.sh@149 -- # timing_enter lib 00:05:42.520 17:27:13 -- common/autotest_common.sh@724 -- # xtrace_disable 00:05:42.520 17:27:13 -- common/autotest_common.sh@10 -- # set +x 00:05:42.520 17:27:13 -- spdk/autotest.sh@151 -- # [[ 0 -eq 1 ]] 00:05:42.520 17:27:13 -- spdk/autotest.sh@155 -- # run_test env /home/vagrant/spdk_repo/spdk/test/env/env.sh 00:05:42.520 17:27:13 -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:05:42.520 17:27:13 -- common/autotest_common.sh@1107 -- # xtrace_disable 00:05:42.520 17:27:13 -- common/autotest_common.sh@10 -- # set +x 00:05:42.520 ************************************ 00:05:42.520 START TEST env 00:05:42.520 ************************************ 00:05:42.520 17:27:13 env -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/env/env.sh 00:05:42.520 * Looking for test storage... 00:05:42.520 * Found test storage at /home/vagrant/spdk_repo/spdk/test/env 00:05:42.520 17:27:13 env -- common/autotest_common.sh@1680 -- # [[ y == y ]] 00:05:42.520 17:27:13 env -- common/autotest_common.sh@1681 -- # lcov --version 00:05:42.520 17:27:13 env -- common/autotest_common.sh@1681 -- # awk '{print $NF}' 00:05:42.781 17:27:13 env -- common/autotest_common.sh@1681 -- # lt 1.15 2 00:05:42.781 17:27:13 env -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:05:42.781 17:27:13 env -- scripts/common.sh@333 -- # local ver1 ver1_l 00:05:42.781 17:27:13 env -- scripts/common.sh@334 -- # local ver2 ver2_l 00:05:42.781 17:27:13 env -- scripts/common.sh@336 -- # IFS=.-: 00:05:42.781 17:27:13 env -- scripts/common.sh@336 -- # read -ra ver1 00:05:42.781 17:27:13 env -- scripts/common.sh@337 -- # IFS=.-: 00:05:42.781 17:27:13 env -- scripts/common.sh@337 -- # read -ra ver2 00:05:42.781 17:27:13 env -- scripts/common.sh@338 -- # local 'op=<' 00:05:42.781 17:27:13 env -- scripts/common.sh@340 -- # ver1_l=2 00:05:42.781 17:27:13 env -- scripts/common.sh@341 -- # ver2_l=1 00:05:42.781 17:27:13 env -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:05:42.781 17:27:13 env -- scripts/common.sh@344 -- # case "$op" in 00:05:42.781 17:27:13 env -- scripts/common.sh@345 -- # : 1 00:05:42.781 17:27:13 env -- scripts/common.sh@364 -- # (( v = 0 )) 00:05:42.781 17:27:13 env -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:05:42.781 17:27:13 env -- scripts/common.sh@365 -- # decimal 1 00:05:42.781 17:27:13 env -- scripts/common.sh@353 -- # local d=1 00:05:42.781 17:27:13 env -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:05:42.781 17:27:13 env -- scripts/common.sh@355 -- # echo 1 00:05:42.781 17:27:13 env -- scripts/common.sh@365 -- # ver1[v]=1 00:05:42.781 17:27:13 env -- scripts/common.sh@366 -- # decimal 2 00:05:42.781 17:27:13 env -- scripts/common.sh@353 -- # local d=2 00:05:42.781 17:27:13 env -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:05:42.781 17:27:13 env -- scripts/common.sh@355 -- # echo 2 00:05:42.781 17:27:13 env -- scripts/common.sh@366 -- # ver2[v]=2 00:05:42.781 17:27:13 env -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:05:42.781 17:27:13 env -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:05:42.781 17:27:13 env -- scripts/common.sh@368 -- # return 0 00:05:42.781 17:27:13 env -- common/autotest_common.sh@1682 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:05:42.781 17:27:13 env -- common/autotest_common.sh@1694 -- # export 'LCOV_OPTS= 00:05:42.781 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:42.781 --rc genhtml_branch_coverage=1 00:05:42.781 --rc genhtml_function_coverage=1 00:05:42.781 --rc genhtml_legend=1 00:05:42.781 --rc geninfo_all_blocks=1 00:05:42.781 --rc geninfo_unexecuted_blocks=1 00:05:42.781 00:05:42.781 ' 00:05:42.781 17:27:13 env -- common/autotest_common.sh@1694 -- # LCOV_OPTS=' 00:05:42.781 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:42.781 --rc genhtml_branch_coverage=1 00:05:42.781 --rc genhtml_function_coverage=1 00:05:42.781 --rc genhtml_legend=1 00:05:42.781 --rc geninfo_all_blocks=1 00:05:42.781 --rc geninfo_unexecuted_blocks=1 00:05:42.781 00:05:42.781 ' 00:05:42.781 17:27:13 env -- common/autotest_common.sh@1695 -- # export 'LCOV=lcov 00:05:42.781 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:42.781 --rc genhtml_branch_coverage=1 00:05:42.781 --rc genhtml_function_coverage=1 00:05:42.781 --rc genhtml_legend=1 00:05:42.781 --rc geninfo_all_blocks=1 00:05:42.781 --rc geninfo_unexecuted_blocks=1 00:05:42.781 00:05:42.781 ' 00:05:42.781 17:27:13 env -- common/autotest_common.sh@1695 -- # LCOV='lcov 00:05:42.781 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:42.781 --rc genhtml_branch_coverage=1 00:05:42.781 --rc genhtml_function_coverage=1 00:05:42.781 --rc genhtml_legend=1 00:05:42.781 --rc geninfo_all_blocks=1 00:05:42.781 --rc geninfo_unexecuted_blocks=1 00:05:42.781 00:05:42.781 ' 00:05:42.781 17:27:13 env -- env/env.sh@10 -- # run_test env_memory /home/vagrant/spdk_repo/spdk/test/env/memory/memory_ut 00:05:42.781 17:27:13 env -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:05:42.781 17:27:13 env -- common/autotest_common.sh@1107 -- # xtrace_disable 00:05:42.781 17:27:13 env -- common/autotest_common.sh@10 -- # set +x 00:05:42.781 ************************************ 00:05:42.781 START TEST env_memory 00:05:42.781 ************************************ 00:05:42.781 17:27:13 env.env_memory -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/env/memory/memory_ut 00:05:42.781 00:05:42.781 00:05:42.781 CUnit - A unit testing framework for C - Version 2.1-3 00:05:42.781 http://cunit.sourceforge.net/ 00:05:42.781 00:05:42.781 00:05:42.781 Suite: memory 00:05:42.781 Test: alloc and free memory map ...[2024-11-27 17:27:13.845263] /home/vagrant/spdk_repo/spdk/lib/env_dpdk/memory.c: 283:spdk_mem_map_alloc: *ERROR*: Initial mem_map notify failed 00:05:42.781 passed 00:05:42.781 Test: mem map translation ...[2024-11-27 17:27:13.885131] /home/vagrant/spdk_repo/spdk/lib/env_dpdk/memory.c: 595:spdk_mem_map_set_translation: *ERROR*: invalid spdk_mem_map_set_translation parameters, vaddr=2097152 len=1234 00:05:42.781 [2024-11-27 17:27:13.885174] /home/vagrant/spdk_repo/spdk/lib/env_dpdk/memory.c: 595:spdk_mem_map_set_translation: *ERROR*: invalid spdk_mem_map_set_translation parameters, vaddr=1234 len=2097152 00:05:42.781 [2024-11-27 17:27:13.885224] /home/vagrant/spdk_repo/spdk/lib/env_dpdk/memory.c: 589:spdk_mem_map_set_translation: *ERROR*: invalid usermode virtual address 281474976710656 00:05:42.781 [2024-11-27 17:27:13.885257] /home/vagrant/spdk_repo/spdk/lib/env_dpdk/memory.c: 605:spdk_mem_map_set_translation: *ERROR*: could not get 0xffffffe00000 map 00:05:42.781 passed 00:05:42.781 Test: mem map registration ...[2024-11-27 17:27:13.945936] /home/vagrant/spdk_repo/spdk/lib/env_dpdk/memory.c: 347:spdk_mem_register: *ERROR*: invalid spdk_mem_register parameters, vaddr=200000 len=1234 00:05:42.781 [2024-11-27 17:27:13.945970] /home/vagrant/spdk_repo/spdk/lib/env_dpdk/memory.c: 347:spdk_mem_register: *ERROR*: invalid spdk_mem_register parameters, vaddr=4d2 len=2097152 00:05:42.781 passed 00:05:43.042 Test: mem map adjacent registrations ...passed 00:05:43.042 00:05:43.042 Run Summary: Type Total Ran Passed Failed Inactive 00:05:43.042 suites 1 1 n/a 0 0 00:05:43.042 tests 4 4 4 0 0 00:05:43.042 asserts 152 152 152 0 n/a 00:05:43.042 00:05:43.042 Elapsed time = 0.219 seconds 00:05:43.042 00:05:43.042 real 0m0.271s 00:05:43.042 user 0m0.228s 00:05:43.042 sys 0m0.033s 00:05:43.042 17:27:14 env.env_memory -- common/autotest_common.sh@1126 -- # xtrace_disable 00:05:43.042 17:27:14 env.env_memory -- common/autotest_common.sh@10 -- # set +x 00:05:43.042 ************************************ 00:05:43.042 END TEST env_memory 00:05:43.042 ************************************ 00:05:43.042 17:27:14 env -- env/env.sh@11 -- # run_test env_vtophys /home/vagrant/spdk_repo/spdk/test/env/vtophys/vtophys 00:05:43.042 17:27:14 env -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:05:43.042 17:27:14 env -- common/autotest_common.sh@1107 -- # xtrace_disable 00:05:43.042 17:27:14 env -- common/autotest_common.sh@10 -- # set +x 00:05:43.042 ************************************ 00:05:43.042 START TEST env_vtophys 00:05:43.042 ************************************ 00:05:43.042 17:27:14 env.env_vtophys -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/env/vtophys/vtophys 00:05:43.042 EAL: lib.eal log level changed from notice to debug 00:05:43.042 EAL: Detected lcore 0 as core 0 on socket 0 00:05:43.042 EAL: Detected lcore 1 as core 0 on socket 0 00:05:43.042 EAL: Detected lcore 2 as core 0 on socket 0 00:05:43.042 EAL: Detected lcore 3 as core 0 on socket 0 00:05:43.042 EAL: Detected lcore 4 as core 0 on socket 0 00:05:43.042 EAL: Detected lcore 5 as core 0 on socket 0 00:05:43.042 EAL: Detected lcore 6 as core 0 on socket 0 00:05:43.042 EAL: Detected lcore 7 as core 0 on socket 0 00:05:43.042 EAL: Detected lcore 8 as core 0 on socket 0 00:05:43.042 EAL: Detected lcore 9 as core 0 on socket 0 00:05:43.042 EAL: Maximum logical cores by configuration: 128 00:05:43.042 EAL: Detected CPU lcores: 10 00:05:43.042 EAL: Detected NUMA nodes: 1 00:05:43.042 EAL: Checking presence of .so 'librte_eal.so.23.0' 00:05:43.042 EAL: Detected shared linkage of DPDK 00:05:43.042 EAL: open shared lib /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0/librte_bus_pci.so.23.0 00:05:43.042 EAL: open shared lib /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0/librte_bus_vdev.so.23.0 00:05:43.042 EAL: Registered [vdev] bus. 00:05:43.042 EAL: bus.vdev log level changed from disabled to notice 00:05:43.042 EAL: open shared lib /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0/librte_mempool_ring.so.23.0 00:05:43.042 EAL: open shared lib /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0/librte_net_i40e.so.23.0 00:05:43.042 EAL: pmd.net.i40e.init log level changed from disabled to notice 00:05:43.042 EAL: pmd.net.i40e.driver log level changed from disabled to notice 00:05:43.042 EAL: open shared lib /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0/librte_bus_pci.so 00:05:43.043 EAL: open shared lib /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0/librte_bus_vdev.so 00:05:43.043 EAL: open shared lib /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0/librte_mempool_ring.so 00:05:43.043 EAL: open shared lib /home/vagrant/spdk_repo/dpdk/build/lib/dpdk/pmds-23.0/librte_net_i40e.so 00:05:43.043 EAL: No shared files mode enabled, IPC will be disabled 00:05:43.043 EAL: No shared files mode enabled, IPC is disabled 00:05:43.043 EAL: Selected IOVA mode 'PA' 00:05:43.043 EAL: Probing VFIO support... 00:05:43.043 EAL: Module /sys/module/vfio not found! error 2 (No such file or directory) 00:05:43.043 EAL: VFIO modules not loaded, skipping VFIO support... 00:05:43.043 EAL: Ask a virtual area of 0x2e000 bytes 00:05:43.043 EAL: Virtual area found at 0x200000000000 (size = 0x2e000) 00:05:43.043 EAL: Setting up physically contiguous memory... 00:05:43.043 EAL: Setting maximum number of open files to 524288 00:05:43.043 EAL: Detected memory type: socket_id:0 hugepage_sz:2097152 00:05:43.043 EAL: Creating 4 segment lists: n_segs:8192 socket_id:0 hugepage_sz:2097152 00:05:43.043 EAL: Ask a virtual area of 0x61000 bytes 00:05:43.043 EAL: Virtual area found at 0x20000002e000 (size = 0x61000) 00:05:43.043 EAL: Memseg list allocated at socket 0, page size 0x800kB 00:05:43.043 EAL: Ask a virtual area of 0x400000000 bytes 00:05:43.043 EAL: Virtual area found at 0x200000200000 (size = 0x400000000) 00:05:43.043 EAL: VA reserved for memseg list at 0x200000200000, size 400000000 00:05:43.043 EAL: Ask a virtual area of 0x61000 bytes 00:05:43.043 EAL: Virtual area found at 0x200400200000 (size = 0x61000) 00:05:43.043 EAL: Memseg list allocated at socket 0, page size 0x800kB 00:05:43.043 EAL: Ask a virtual area of 0x400000000 bytes 00:05:43.043 EAL: Virtual area found at 0x200400400000 (size = 0x400000000) 00:05:43.043 EAL: VA reserved for memseg list at 0x200400400000, size 400000000 00:05:43.043 EAL: Ask a virtual area of 0x61000 bytes 00:05:43.043 EAL: Virtual area found at 0x200800400000 (size = 0x61000) 00:05:43.043 EAL: Memseg list allocated at socket 0, page size 0x800kB 00:05:43.043 EAL: Ask a virtual area of 0x400000000 bytes 00:05:43.043 EAL: Virtual area found at 0x200800600000 (size = 0x400000000) 00:05:43.043 EAL: VA reserved for memseg list at 0x200800600000, size 400000000 00:05:43.043 EAL: Ask a virtual area of 0x61000 bytes 00:05:43.043 EAL: Virtual area found at 0x200c00600000 (size = 0x61000) 00:05:43.043 EAL: Memseg list allocated at socket 0, page size 0x800kB 00:05:43.043 EAL: Ask a virtual area of 0x400000000 bytes 00:05:43.043 EAL: Virtual area found at 0x200c00800000 (size = 0x400000000) 00:05:43.043 EAL: VA reserved for memseg list at 0x200c00800000, size 400000000 00:05:43.043 EAL: Hugepages will be freed exactly as allocated. 00:05:43.043 EAL: No shared files mode enabled, IPC is disabled 00:05:43.043 EAL: No shared files mode enabled, IPC is disabled 00:05:43.304 EAL: TSC frequency is ~2290000 KHz 00:05:43.304 EAL: Main lcore 0 is ready (tid=7f38d6a82a40;cpuset=[0]) 00:05:43.304 EAL: Trying to obtain current memory policy. 00:05:43.304 EAL: Setting policy MPOL_PREFERRED for socket 0 00:05:43.304 EAL: Restoring previous memory policy: 0 00:05:43.304 EAL: request: mp_malloc_sync 00:05:43.304 EAL: No shared files mode enabled, IPC is disabled 00:05:43.304 EAL: Heap on socket 0 was expanded by 2MB 00:05:43.304 EAL: Module /sys/module/vfio not found! error 2 (No such file or directory) 00:05:43.304 EAL: No shared files mode enabled, IPC is disabled 00:05:43.304 EAL: No PCI address specified using 'addr=' in: bus=pci 00:05:43.304 EAL: Mem event callback 'spdk:(nil)' registered 00:05:43.304 EAL: Module /sys/module/vfio_pci not found! error 2 (No such file or directory) 00:05:43.304 00:05:43.304 00:05:43.304 CUnit - A unit testing framework for C - Version 2.1-3 00:05:43.304 http://cunit.sourceforge.net/ 00:05:43.304 00:05:43.304 00:05:43.304 Suite: components_suite 00:05:43.564 Test: vtophys_malloc_test ...passed 00:05:43.564 Test: vtophys_spdk_malloc_test ...EAL: Trying to obtain current memory policy. 00:05:43.564 EAL: Setting policy MPOL_PREFERRED for socket 0 00:05:43.564 EAL: Restoring previous memory policy: 4 00:05:43.564 EAL: Calling mem event callback 'spdk:(nil)' 00:05:43.564 EAL: request: mp_malloc_sync 00:05:43.564 EAL: No shared files mode enabled, IPC is disabled 00:05:43.564 EAL: Heap on socket 0 was expanded by 4MB 00:05:43.564 EAL: Calling mem event callback 'spdk:(nil)' 00:05:43.564 EAL: request: mp_malloc_sync 00:05:43.564 EAL: No shared files mode enabled, IPC is disabled 00:05:43.564 EAL: Heap on socket 0 was shrunk by 4MB 00:05:43.564 EAL: Trying to obtain current memory policy. 00:05:43.564 EAL: Setting policy MPOL_PREFERRED for socket 0 00:05:43.564 EAL: Restoring previous memory policy: 4 00:05:43.564 EAL: Calling mem event callback 'spdk:(nil)' 00:05:43.564 EAL: request: mp_malloc_sync 00:05:43.564 EAL: No shared files mode enabled, IPC is disabled 00:05:43.564 EAL: Heap on socket 0 was expanded by 6MB 00:05:43.564 EAL: Calling mem event callback 'spdk:(nil)' 00:05:43.564 EAL: request: mp_malloc_sync 00:05:43.564 EAL: No shared files mode enabled, IPC is disabled 00:05:43.564 EAL: Heap on socket 0 was shrunk by 6MB 00:05:43.564 EAL: Trying to obtain current memory policy. 00:05:43.564 EAL: Setting policy MPOL_PREFERRED for socket 0 00:05:43.564 EAL: Restoring previous memory policy: 4 00:05:43.564 EAL: Calling mem event callback 'spdk:(nil)' 00:05:43.564 EAL: request: mp_malloc_sync 00:05:43.564 EAL: No shared files mode enabled, IPC is disabled 00:05:43.564 EAL: Heap on socket 0 was expanded by 10MB 00:05:43.564 EAL: Calling mem event callback 'spdk:(nil)' 00:05:43.564 EAL: request: mp_malloc_sync 00:05:43.565 EAL: No shared files mode enabled, IPC is disabled 00:05:43.565 EAL: Heap on socket 0 was shrunk by 10MB 00:05:43.565 EAL: Trying to obtain current memory policy. 00:05:43.565 EAL: Setting policy MPOL_PREFERRED for socket 0 00:05:43.565 EAL: Restoring previous memory policy: 4 00:05:43.565 EAL: Calling mem event callback 'spdk:(nil)' 00:05:43.565 EAL: request: mp_malloc_sync 00:05:43.565 EAL: No shared files mode enabled, IPC is disabled 00:05:43.565 EAL: Heap on socket 0 was expanded by 18MB 00:05:43.825 EAL: Calling mem event callback 'spdk:(nil)' 00:05:43.825 EAL: request: mp_malloc_sync 00:05:43.825 EAL: No shared files mode enabled, IPC is disabled 00:05:43.825 EAL: Heap on socket 0 was shrunk by 18MB 00:05:43.825 EAL: Trying to obtain current memory policy. 00:05:43.825 EAL: Setting policy MPOL_PREFERRED for socket 0 00:05:43.825 EAL: Restoring previous memory policy: 4 00:05:43.825 EAL: Calling mem event callback 'spdk:(nil)' 00:05:43.825 EAL: request: mp_malloc_sync 00:05:43.825 EAL: No shared files mode enabled, IPC is disabled 00:05:43.825 EAL: Heap on socket 0 was expanded by 34MB 00:05:43.825 EAL: Calling mem event callback 'spdk:(nil)' 00:05:43.825 EAL: request: mp_malloc_sync 00:05:43.825 EAL: No shared files mode enabled, IPC is disabled 00:05:43.825 EAL: Heap on socket 0 was shrunk by 34MB 00:05:43.825 EAL: Trying to obtain current memory policy. 00:05:43.825 EAL: Setting policy MPOL_PREFERRED for socket 0 00:05:43.825 EAL: Restoring previous memory policy: 4 00:05:43.825 EAL: Calling mem event callback 'spdk:(nil)' 00:05:43.825 EAL: request: mp_malloc_sync 00:05:43.825 EAL: No shared files mode enabled, IPC is disabled 00:05:43.825 EAL: Heap on socket 0 was expanded by 66MB 00:05:43.825 EAL: Calling mem event callback 'spdk:(nil)' 00:05:43.825 EAL: request: mp_malloc_sync 00:05:43.825 EAL: No shared files mode enabled, IPC is disabled 00:05:43.825 EAL: Heap on socket 0 was shrunk by 66MB 00:05:43.825 EAL: Trying to obtain current memory policy. 00:05:43.825 EAL: Setting policy MPOL_PREFERRED for socket 0 00:05:43.825 EAL: Restoring previous memory policy: 4 00:05:43.825 EAL: Calling mem event callback 'spdk:(nil)' 00:05:43.825 EAL: request: mp_malloc_sync 00:05:43.825 EAL: No shared files mode enabled, IPC is disabled 00:05:43.825 EAL: Heap on socket 0 was expanded by 130MB 00:05:43.825 EAL: Calling mem event callback 'spdk:(nil)' 00:05:43.825 EAL: request: mp_malloc_sync 00:05:43.825 EAL: No shared files mode enabled, IPC is disabled 00:05:43.825 EAL: Heap on socket 0 was shrunk by 130MB 00:05:43.825 EAL: Trying to obtain current memory policy. 00:05:43.825 EAL: Setting policy MPOL_PREFERRED for socket 0 00:05:44.084 EAL: Restoring previous memory policy: 4 00:05:44.084 EAL: Calling mem event callback 'spdk:(nil)' 00:05:44.084 EAL: request: mp_malloc_sync 00:05:44.084 EAL: No shared files mode enabled, IPC is disabled 00:05:44.084 EAL: Heap on socket 0 was expanded by 258MB 00:05:44.084 EAL: Calling mem event callback 'spdk:(nil)' 00:05:44.084 EAL: request: mp_malloc_sync 00:05:44.084 EAL: No shared files mode enabled, IPC is disabled 00:05:44.084 EAL: Heap on socket 0 was shrunk by 258MB 00:05:44.084 EAL: Trying to obtain current memory policy. 00:05:44.084 EAL: Setting policy MPOL_PREFERRED for socket 0 00:05:44.345 EAL: Restoring previous memory policy: 4 00:05:44.345 EAL: Calling mem event callback 'spdk:(nil)' 00:05:44.345 EAL: request: mp_malloc_sync 00:05:44.345 EAL: No shared files mode enabled, IPC is disabled 00:05:44.345 EAL: Heap on socket 0 was expanded by 514MB 00:05:44.345 EAL: Calling mem event callback 'spdk:(nil)' 00:05:44.605 EAL: request: mp_malloc_sync 00:05:44.605 EAL: No shared files mode enabled, IPC is disabled 00:05:44.605 EAL: Heap on socket 0 was shrunk by 514MB 00:05:44.605 EAL: Trying to obtain current memory policy. 00:05:44.605 EAL: Setting policy MPOL_PREFERRED for socket 0 00:05:44.865 EAL: Restoring previous memory policy: 4 00:05:44.865 EAL: Calling mem event callback 'spdk:(nil)' 00:05:44.865 EAL: request: mp_malloc_sync 00:05:44.865 EAL: No shared files mode enabled, IPC is disabled 00:05:44.865 EAL: Heap on socket 0 was expanded by 1026MB 00:05:45.434 EAL: Calling mem event callback 'spdk:(nil)' 00:05:45.694 passed 00:05:45.694 00:05:45.694 Run Summary: Type Total Ran Passed Failed Inactive 00:05:45.694 suites 1 1 n/a 0 0 00:05:45.694 tests 2 2 2 0 0 00:05:45.694 asserts 5582 5582 5582 0 n/a 00:05:45.694 00:05:45.694 Elapsed time = 2.347 seconds 00:05:45.694 EAL: request: mp_malloc_sync 00:05:45.694 EAL: No shared files mode enabled, IPC is disabled 00:05:45.694 EAL: Heap on socket 0 was shrunk by 1026MB 00:05:45.694 EAL: Calling mem event callback 'spdk:(nil)' 00:05:45.694 EAL: request: mp_malloc_sync 00:05:45.694 EAL: No shared files mode enabled, IPC is disabled 00:05:45.694 EAL: Heap on socket 0 was shrunk by 2MB 00:05:45.694 EAL: No shared files mode enabled, IPC is disabled 00:05:45.694 EAL: No shared files mode enabled, IPC is disabled 00:05:45.694 EAL: No shared files mode enabled, IPC is disabled 00:05:45.694 00:05:45.694 real 0m2.596s 00:05:45.694 user 0m1.367s 00:05:45.694 sys 0m1.091s 00:05:45.694 ************************************ 00:05:45.694 END TEST env_vtophys 00:05:45.694 ************************************ 00:05:45.694 17:27:16 env.env_vtophys -- common/autotest_common.sh@1126 -- # xtrace_disable 00:05:45.694 17:27:16 env.env_vtophys -- common/autotest_common.sh@10 -- # set +x 00:05:45.694 17:27:16 env -- env/env.sh@12 -- # run_test env_pci /home/vagrant/spdk_repo/spdk/test/env/pci/pci_ut 00:05:45.694 17:27:16 env -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:05:45.694 17:27:16 env -- common/autotest_common.sh@1107 -- # xtrace_disable 00:05:45.694 17:27:16 env -- common/autotest_common.sh@10 -- # set +x 00:05:45.694 ************************************ 00:05:45.694 START TEST env_pci 00:05:45.694 ************************************ 00:05:45.694 17:27:16 env.env_pci -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/env/pci/pci_ut 00:05:45.694 00:05:45.694 00:05:45.694 CUnit - A unit testing framework for C - Version 2.1-3 00:05:45.694 http://cunit.sourceforge.net/ 00:05:45.694 00:05:45.694 00:05:45.694 Suite: pci 00:05:45.694 Test: pci_hook ...[2024-11-27 17:27:16.815231] /home/vagrant/spdk_repo/spdk/lib/env_dpdk/pci.c:1049:spdk_pci_device_claim: *ERROR*: Cannot create lock on device /var/tmp/spdk_pci_lock_10000:00:01.0, probably process 68685 has claimed it 00:05:45.694 passed 00:05:45.694 00:05:45.694 Run Summary: Type Total Ran Passed Failed Inactive 00:05:45.694 suites 1 1 n/a 0 0 00:05:45.694 tests 1 1 1 0 0 00:05:45.694 asserts 25 25 25 0 n/a 00:05:45.694 00:05:45.694 Elapsed time = 0.010 seconds 00:05:45.694 EAL: Cannot find device (10000:00:01.0) 00:05:45.694 EAL: Failed to attach device on primary process 00:05:45.694 00:05:45.694 real 0m0.098s 00:05:45.694 user 0m0.040s 00:05:45.694 sys 0m0.057s 00:05:45.694 ************************************ 00:05:45.694 END TEST env_pci 00:05:45.694 ************************************ 00:05:45.694 17:27:16 env.env_pci -- common/autotest_common.sh@1126 -- # xtrace_disable 00:05:45.694 17:27:16 env.env_pci -- common/autotest_common.sh@10 -- # set +x 00:05:45.953 17:27:16 env -- env/env.sh@14 -- # argv='-c 0x1 ' 00:05:45.953 17:27:16 env -- env/env.sh@15 -- # uname 00:05:45.953 17:27:16 env -- env/env.sh@15 -- # '[' Linux = Linux ']' 00:05:45.953 17:27:16 env -- env/env.sh@22 -- # argv+=--base-virtaddr=0x200000000000 00:05:45.953 17:27:16 env -- env/env.sh@24 -- # run_test env_dpdk_post_init /home/vagrant/spdk_repo/spdk/test/env/env_dpdk_post_init/env_dpdk_post_init -c 0x1 --base-virtaddr=0x200000000000 00:05:45.953 17:27:16 env -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:05:45.953 17:27:16 env -- common/autotest_common.sh@1107 -- # xtrace_disable 00:05:45.953 17:27:16 env -- common/autotest_common.sh@10 -- # set +x 00:05:45.953 ************************************ 00:05:45.953 START TEST env_dpdk_post_init 00:05:45.953 ************************************ 00:05:45.953 17:27:16 env.env_dpdk_post_init -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/env/env_dpdk_post_init/env_dpdk_post_init -c 0x1 --base-virtaddr=0x200000000000 00:05:45.953 EAL: Detected CPU lcores: 10 00:05:45.953 EAL: Detected NUMA nodes: 1 00:05:45.953 EAL: Detected shared linkage of DPDK 00:05:45.953 EAL: Multi-process socket /var/run/dpdk/rte/mp_socket 00:05:45.953 EAL: Selected IOVA mode 'PA' 00:05:45.953 TELEMETRY: No legacy callbacks, legacy socket not created 00:05:46.214 EAL: Probe PCI driver: spdk_nvme (1b36:0010) device: 0000:00:10.0 (socket -1) 00:05:46.214 EAL: Probe PCI driver: spdk_nvme (1b36:0010) device: 0000:00:11.0 (socket -1) 00:05:46.214 Starting DPDK initialization... 00:05:46.214 Starting SPDK post initialization... 00:05:46.214 SPDK NVMe probe 00:05:46.214 Attaching to 0000:00:10.0 00:05:46.214 Attaching to 0000:00:11.0 00:05:46.214 Attached to 0000:00:10.0 00:05:46.214 Attached to 0000:00:11.0 00:05:46.214 Cleaning up... 00:05:46.214 00:05:46.214 real 0m0.239s 00:05:46.214 user 0m0.065s 00:05:46.214 sys 0m0.075s 00:05:46.214 17:27:17 env.env_dpdk_post_init -- common/autotest_common.sh@1126 -- # xtrace_disable 00:05:46.214 ************************************ 00:05:46.214 END TEST env_dpdk_post_init 00:05:46.214 ************************************ 00:05:46.214 17:27:17 env.env_dpdk_post_init -- common/autotest_common.sh@10 -- # set +x 00:05:46.214 17:27:17 env -- env/env.sh@26 -- # uname 00:05:46.214 17:27:17 env -- env/env.sh@26 -- # '[' Linux = Linux ']' 00:05:46.214 17:27:17 env -- env/env.sh@29 -- # run_test env_mem_callbacks /home/vagrant/spdk_repo/spdk/test/env/mem_callbacks/mem_callbacks 00:05:46.214 17:27:17 env -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:05:46.214 17:27:17 env -- common/autotest_common.sh@1107 -- # xtrace_disable 00:05:46.214 17:27:17 env -- common/autotest_common.sh@10 -- # set +x 00:05:46.214 ************************************ 00:05:46.214 START TEST env_mem_callbacks 00:05:46.214 ************************************ 00:05:46.214 17:27:17 env.env_mem_callbacks -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/env/mem_callbacks/mem_callbacks 00:05:46.214 EAL: Detected CPU lcores: 10 00:05:46.214 EAL: Detected NUMA nodes: 1 00:05:46.214 EAL: Detected shared linkage of DPDK 00:05:46.214 EAL: Multi-process socket /var/run/dpdk/rte/mp_socket 00:05:46.214 EAL: Selected IOVA mode 'PA' 00:05:46.474 TELEMETRY: No legacy callbacks, legacy socket not created 00:05:46.474 00:05:46.474 00:05:46.474 CUnit - A unit testing framework for C - Version 2.1-3 00:05:46.474 http://cunit.sourceforge.net/ 00:05:46.474 00:05:46.474 00:05:46.474 Suite: memory 00:05:46.474 Test: test ... 00:05:46.474 register 0x200000200000 2097152 00:05:46.474 malloc 3145728 00:05:46.474 register 0x200000400000 4194304 00:05:46.474 buf 0x200000500000 len 3145728 PASSED 00:05:46.474 malloc 64 00:05:46.474 buf 0x2000004fff40 len 64 PASSED 00:05:46.474 malloc 4194304 00:05:46.474 register 0x200000800000 6291456 00:05:46.474 buf 0x200000a00000 len 4194304 PASSED 00:05:46.474 free 0x200000500000 3145728 00:05:46.474 free 0x2000004fff40 64 00:05:46.474 unregister 0x200000400000 4194304 PASSED 00:05:46.474 free 0x200000a00000 4194304 00:05:46.474 unregister 0x200000800000 6291456 PASSED 00:05:46.474 malloc 8388608 00:05:46.474 register 0x200000400000 10485760 00:05:46.474 buf 0x200000600000 len 8388608 PASSED 00:05:46.474 free 0x200000600000 8388608 00:05:46.474 unregister 0x200000400000 10485760 PASSED 00:05:46.474 passed 00:05:46.474 00:05:46.474 Run Summary: Type Total Ran Passed Failed Inactive 00:05:46.474 suites 1 1 n/a 0 0 00:05:46.474 tests 1 1 1 0 0 00:05:46.474 asserts 15 15 15 0 n/a 00:05:46.474 00:05:46.474 Elapsed time = 0.013 seconds 00:05:46.474 00:05:46.474 real 0m0.187s 00:05:46.474 user 0m0.034s 00:05:46.474 sys 0m0.050s 00:05:46.474 17:27:17 env.env_mem_callbacks -- common/autotest_common.sh@1126 -- # xtrace_disable 00:05:46.474 17:27:17 env.env_mem_callbacks -- common/autotest_common.sh@10 -- # set +x 00:05:46.474 ************************************ 00:05:46.474 END TEST env_mem_callbacks 00:05:46.474 ************************************ 00:05:46.474 00:05:46.474 real 0m3.973s 00:05:46.474 user 0m1.978s 00:05:46.474 sys 0m1.660s 00:05:46.474 ************************************ 00:05:46.474 END TEST env 00:05:46.474 ************************************ 00:05:46.474 17:27:17 env -- common/autotest_common.sh@1126 -- # xtrace_disable 00:05:46.474 17:27:17 env -- common/autotest_common.sh@10 -- # set +x 00:05:46.474 17:27:17 -- spdk/autotest.sh@156 -- # run_test rpc /home/vagrant/spdk_repo/spdk/test/rpc/rpc.sh 00:05:46.474 17:27:17 -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:05:46.474 17:27:17 -- common/autotest_common.sh@1107 -- # xtrace_disable 00:05:46.474 17:27:17 -- common/autotest_common.sh@10 -- # set +x 00:05:46.474 ************************************ 00:05:46.474 START TEST rpc 00:05:46.474 ************************************ 00:05:46.474 17:27:17 rpc -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/rpc/rpc.sh 00:05:46.734 * Looking for test storage... 00:05:46.734 * Found test storage at /home/vagrant/spdk_repo/spdk/test/rpc 00:05:46.734 17:27:17 rpc -- common/autotest_common.sh@1680 -- # [[ y == y ]] 00:05:46.734 17:27:17 rpc -- common/autotest_common.sh@1681 -- # lcov --version 00:05:46.734 17:27:17 rpc -- common/autotest_common.sh@1681 -- # awk '{print $NF}' 00:05:46.734 17:27:17 rpc -- common/autotest_common.sh@1681 -- # lt 1.15 2 00:05:46.734 17:27:17 rpc -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:05:46.734 17:27:17 rpc -- scripts/common.sh@333 -- # local ver1 ver1_l 00:05:46.734 17:27:17 rpc -- scripts/common.sh@334 -- # local ver2 ver2_l 00:05:46.734 17:27:17 rpc -- scripts/common.sh@336 -- # IFS=.-: 00:05:46.734 17:27:17 rpc -- scripts/common.sh@336 -- # read -ra ver1 00:05:46.734 17:27:17 rpc -- scripts/common.sh@337 -- # IFS=.-: 00:05:46.734 17:27:17 rpc -- scripts/common.sh@337 -- # read -ra ver2 00:05:46.734 17:27:17 rpc -- scripts/common.sh@338 -- # local 'op=<' 00:05:46.734 17:27:17 rpc -- scripts/common.sh@340 -- # ver1_l=2 00:05:46.734 17:27:17 rpc -- scripts/common.sh@341 -- # ver2_l=1 00:05:46.734 17:27:17 rpc -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:05:46.734 17:27:17 rpc -- scripts/common.sh@344 -- # case "$op" in 00:05:46.734 17:27:17 rpc -- scripts/common.sh@345 -- # : 1 00:05:46.734 17:27:17 rpc -- scripts/common.sh@364 -- # (( v = 0 )) 00:05:46.734 17:27:17 rpc -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:05:46.734 17:27:17 rpc -- scripts/common.sh@365 -- # decimal 1 00:05:46.734 17:27:17 rpc -- scripts/common.sh@353 -- # local d=1 00:05:46.734 17:27:17 rpc -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:05:46.734 17:27:17 rpc -- scripts/common.sh@355 -- # echo 1 00:05:46.734 17:27:17 rpc -- scripts/common.sh@365 -- # ver1[v]=1 00:05:46.734 17:27:17 rpc -- scripts/common.sh@366 -- # decimal 2 00:05:46.734 17:27:17 rpc -- scripts/common.sh@353 -- # local d=2 00:05:46.734 17:27:17 rpc -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:05:46.734 17:27:17 rpc -- scripts/common.sh@355 -- # echo 2 00:05:46.734 17:27:17 rpc -- scripts/common.sh@366 -- # ver2[v]=2 00:05:46.734 17:27:17 rpc -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:05:46.734 17:27:17 rpc -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:05:46.734 17:27:17 rpc -- scripts/common.sh@368 -- # return 0 00:05:46.734 17:27:17 rpc -- common/autotest_common.sh@1682 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:05:46.734 17:27:17 rpc -- common/autotest_common.sh@1694 -- # export 'LCOV_OPTS= 00:05:46.734 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:46.734 --rc genhtml_branch_coverage=1 00:05:46.734 --rc genhtml_function_coverage=1 00:05:46.734 --rc genhtml_legend=1 00:05:46.734 --rc geninfo_all_blocks=1 00:05:46.734 --rc geninfo_unexecuted_blocks=1 00:05:46.734 00:05:46.734 ' 00:05:46.734 17:27:17 rpc -- common/autotest_common.sh@1694 -- # LCOV_OPTS=' 00:05:46.734 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:46.734 --rc genhtml_branch_coverage=1 00:05:46.734 --rc genhtml_function_coverage=1 00:05:46.734 --rc genhtml_legend=1 00:05:46.734 --rc geninfo_all_blocks=1 00:05:46.734 --rc geninfo_unexecuted_blocks=1 00:05:46.734 00:05:46.734 ' 00:05:46.734 17:27:17 rpc -- common/autotest_common.sh@1695 -- # export 'LCOV=lcov 00:05:46.734 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:46.734 --rc genhtml_branch_coverage=1 00:05:46.734 --rc genhtml_function_coverage=1 00:05:46.734 --rc genhtml_legend=1 00:05:46.734 --rc geninfo_all_blocks=1 00:05:46.734 --rc geninfo_unexecuted_blocks=1 00:05:46.734 00:05:46.734 ' 00:05:46.734 17:27:17 rpc -- common/autotest_common.sh@1695 -- # LCOV='lcov 00:05:46.734 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:46.734 --rc genhtml_branch_coverage=1 00:05:46.734 --rc genhtml_function_coverage=1 00:05:46.734 --rc genhtml_legend=1 00:05:46.734 --rc geninfo_all_blocks=1 00:05:46.734 --rc geninfo_unexecuted_blocks=1 00:05:46.734 00:05:46.734 ' 00:05:46.734 17:27:17 rpc -- rpc/rpc.sh@65 -- # spdk_pid=68812 00:05:46.734 17:27:17 rpc -- rpc/rpc.sh@64 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -e bdev 00:05:46.734 17:27:17 rpc -- rpc/rpc.sh@66 -- # trap 'killprocess $spdk_pid; exit 1' SIGINT SIGTERM EXIT 00:05:46.734 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:05:46.734 17:27:17 rpc -- rpc/rpc.sh@67 -- # waitforlisten 68812 00:05:46.734 17:27:17 rpc -- common/autotest_common.sh@831 -- # '[' -z 68812 ']' 00:05:46.734 17:27:17 rpc -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:05:46.734 17:27:17 rpc -- common/autotest_common.sh@836 -- # local max_retries=100 00:05:46.734 17:27:17 rpc -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:05:46.734 17:27:17 rpc -- common/autotest_common.sh@840 -- # xtrace_disable 00:05:46.734 17:27:17 rpc -- common/autotest_common.sh@10 -- # set +x 00:05:46.734 [2024-11-27 17:27:17.910640] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:05:46.734 [2024-11-27 17:27:17.910879] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid68812 ] 00:05:46.994 [2024-11-27 17:27:18.055673] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:05:46.994 [2024-11-27 17:27:18.124838] app.c: 610:app_setup_trace: *NOTICE*: Tracepoint Group Mask bdev specified. 00:05:46.994 [2024-11-27 17:27:18.124999] app.c: 611:app_setup_trace: *NOTICE*: Use 'spdk_trace -s spdk_tgt -p 68812' to capture a snapshot of events at runtime. 00:05:46.994 [2024-11-27 17:27:18.125046] app.c: 616:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:05:46.994 [2024-11-27 17:27:18.125075] app.c: 617:app_setup_trace: *NOTICE*: SPDK application currently running. 00:05:46.994 [2024-11-27 17:27:18.125117] app.c: 618:app_setup_trace: *NOTICE*: Or copy /dev/shm/spdk_tgt_trace.pid68812 for offline analysis/debug. 00:05:46.994 [2024-11-27 17:27:18.125211] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:05:47.562 17:27:18 rpc -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:05:47.562 17:27:18 rpc -- common/autotest_common.sh@864 -- # return 0 00:05:47.562 17:27:18 rpc -- rpc/rpc.sh@69 -- # export PYTHONPATH=:/home/vagrant/spdk_repo/spdk/python:/home/vagrant/spdk_repo/spdk/test/rpc_plugins:/home/vagrant/spdk_repo/spdk/python:/home/vagrant/spdk_repo/spdk/test/rpc_plugins:/home/vagrant/spdk_repo/spdk/test/rpc 00:05:47.562 17:27:18 rpc -- rpc/rpc.sh@69 -- # PYTHONPATH=:/home/vagrant/spdk_repo/spdk/python:/home/vagrant/spdk_repo/spdk/test/rpc_plugins:/home/vagrant/spdk_repo/spdk/python:/home/vagrant/spdk_repo/spdk/test/rpc_plugins:/home/vagrant/spdk_repo/spdk/test/rpc 00:05:47.562 17:27:18 rpc -- rpc/rpc.sh@72 -- # rpc=rpc_cmd 00:05:47.562 17:27:18 rpc -- rpc/rpc.sh@73 -- # run_test rpc_integrity rpc_integrity 00:05:47.562 17:27:18 rpc -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:05:47.562 17:27:18 rpc -- common/autotest_common.sh@1107 -- # xtrace_disable 00:05:47.562 17:27:18 rpc -- common/autotest_common.sh@10 -- # set +x 00:05:47.562 ************************************ 00:05:47.562 START TEST rpc_integrity 00:05:47.562 ************************************ 00:05:47.562 17:27:18 rpc.rpc_integrity -- common/autotest_common.sh@1125 -- # rpc_integrity 00:05:47.562 17:27:18 rpc.rpc_integrity -- rpc/rpc.sh@12 -- # rpc_cmd bdev_get_bdevs 00:05:47.562 17:27:18 rpc.rpc_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:47.562 17:27:18 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:47.822 17:27:18 rpc.rpc_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:05:47.822 17:27:18 rpc.rpc_integrity -- rpc/rpc.sh@12 -- # bdevs='[]' 00:05:47.822 17:27:18 rpc.rpc_integrity -- rpc/rpc.sh@13 -- # jq length 00:05:47.822 17:27:18 rpc.rpc_integrity -- rpc/rpc.sh@13 -- # '[' 0 == 0 ']' 00:05:47.822 17:27:18 rpc.rpc_integrity -- rpc/rpc.sh@15 -- # rpc_cmd bdev_malloc_create 8 512 00:05:47.822 17:27:18 rpc.rpc_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:47.822 17:27:18 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:47.822 17:27:18 rpc.rpc_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:05:47.822 17:27:18 rpc.rpc_integrity -- rpc/rpc.sh@15 -- # malloc=Malloc0 00:05:47.822 17:27:18 rpc.rpc_integrity -- rpc/rpc.sh@16 -- # rpc_cmd bdev_get_bdevs 00:05:47.822 17:27:18 rpc.rpc_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:47.822 17:27:18 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:47.822 17:27:18 rpc.rpc_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:05:47.822 17:27:18 rpc.rpc_integrity -- rpc/rpc.sh@16 -- # bdevs='[ 00:05:47.822 { 00:05:47.822 "name": "Malloc0", 00:05:47.822 "aliases": [ 00:05:47.822 "a81fffd7-4da7-48bd-8201-b992c12ef4fa" 00:05:47.822 ], 00:05:47.822 "product_name": "Malloc disk", 00:05:47.822 "block_size": 512, 00:05:47.822 "num_blocks": 16384, 00:05:47.822 "uuid": "a81fffd7-4da7-48bd-8201-b992c12ef4fa", 00:05:47.822 "assigned_rate_limits": { 00:05:47.822 "rw_ios_per_sec": 0, 00:05:47.822 "rw_mbytes_per_sec": 0, 00:05:47.822 "r_mbytes_per_sec": 0, 00:05:47.822 "w_mbytes_per_sec": 0 00:05:47.822 }, 00:05:47.822 "claimed": false, 00:05:47.822 "zoned": false, 00:05:47.822 "supported_io_types": { 00:05:47.822 "read": true, 00:05:47.822 "write": true, 00:05:47.822 "unmap": true, 00:05:47.822 "flush": true, 00:05:47.822 "reset": true, 00:05:47.822 "nvme_admin": false, 00:05:47.822 "nvme_io": false, 00:05:47.822 "nvme_io_md": false, 00:05:47.822 "write_zeroes": true, 00:05:47.822 "zcopy": true, 00:05:47.822 "get_zone_info": false, 00:05:47.822 "zone_management": false, 00:05:47.822 "zone_append": false, 00:05:47.822 "compare": false, 00:05:47.822 "compare_and_write": false, 00:05:47.822 "abort": true, 00:05:47.822 "seek_hole": false, 00:05:47.822 "seek_data": false, 00:05:47.822 "copy": true, 00:05:47.822 "nvme_iov_md": false 00:05:47.822 }, 00:05:47.822 "memory_domains": [ 00:05:47.822 { 00:05:47.822 "dma_device_id": "system", 00:05:47.822 "dma_device_type": 1 00:05:47.822 }, 00:05:47.822 { 00:05:47.822 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:05:47.822 "dma_device_type": 2 00:05:47.822 } 00:05:47.822 ], 00:05:47.822 "driver_specific": {} 00:05:47.822 } 00:05:47.822 ]' 00:05:47.822 17:27:18 rpc.rpc_integrity -- rpc/rpc.sh@17 -- # jq length 00:05:47.822 17:27:18 rpc.rpc_integrity -- rpc/rpc.sh@17 -- # '[' 1 == 1 ']' 00:05:47.822 17:27:18 rpc.rpc_integrity -- rpc/rpc.sh@19 -- # rpc_cmd bdev_passthru_create -b Malloc0 -p Passthru0 00:05:47.822 17:27:18 rpc.rpc_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:47.822 17:27:18 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:47.822 [2024-11-27 17:27:18.897483] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on Malloc0 00:05:47.822 [2024-11-27 17:27:18.897605] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:05:47.822 [2024-11-27 17:27:18.897664] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006c80 00:05:47.822 [2024-11-27 17:27:18.897714] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:05:47.822 [2024-11-27 17:27:18.900329] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:05:47.822 [2024-11-27 17:27:18.900396] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: Passthru0 00:05:47.822 Passthru0 00:05:47.822 17:27:18 rpc.rpc_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:05:47.822 17:27:18 rpc.rpc_integrity -- rpc/rpc.sh@20 -- # rpc_cmd bdev_get_bdevs 00:05:47.822 17:27:18 rpc.rpc_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:47.822 17:27:18 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:47.822 17:27:18 rpc.rpc_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:05:47.822 17:27:18 rpc.rpc_integrity -- rpc/rpc.sh@20 -- # bdevs='[ 00:05:47.822 { 00:05:47.822 "name": "Malloc0", 00:05:47.822 "aliases": [ 00:05:47.822 "a81fffd7-4da7-48bd-8201-b992c12ef4fa" 00:05:47.822 ], 00:05:47.822 "product_name": "Malloc disk", 00:05:47.822 "block_size": 512, 00:05:47.822 "num_blocks": 16384, 00:05:47.822 "uuid": "a81fffd7-4da7-48bd-8201-b992c12ef4fa", 00:05:47.822 "assigned_rate_limits": { 00:05:47.822 "rw_ios_per_sec": 0, 00:05:47.822 "rw_mbytes_per_sec": 0, 00:05:47.822 "r_mbytes_per_sec": 0, 00:05:47.822 "w_mbytes_per_sec": 0 00:05:47.822 }, 00:05:47.822 "claimed": true, 00:05:47.822 "claim_type": "exclusive_write", 00:05:47.822 "zoned": false, 00:05:47.822 "supported_io_types": { 00:05:47.822 "read": true, 00:05:47.822 "write": true, 00:05:47.822 "unmap": true, 00:05:47.822 "flush": true, 00:05:47.822 "reset": true, 00:05:47.822 "nvme_admin": false, 00:05:47.822 "nvme_io": false, 00:05:47.822 "nvme_io_md": false, 00:05:47.822 "write_zeroes": true, 00:05:47.822 "zcopy": true, 00:05:47.822 "get_zone_info": false, 00:05:47.822 "zone_management": false, 00:05:47.822 "zone_append": false, 00:05:47.822 "compare": false, 00:05:47.822 "compare_and_write": false, 00:05:47.822 "abort": true, 00:05:47.822 "seek_hole": false, 00:05:47.822 "seek_data": false, 00:05:47.822 "copy": true, 00:05:47.822 "nvme_iov_md": false 00:05:47.822 }, 00:05:47.822 "memory_domains": [ 00:05:47.822 { 00:05:47.822 "dma_device_id": "system", 00:05:47.823 "dma_device_type": 1 00:05:47.823 }, 00:05:47.823 { 00:05:47.823 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:05:47.823 "dma_device_type": 2 00:05:47.823 } 00:05:47.823 ], 00:05:47.823 "driver_specific": {} 00:05:47.823 }, 00:05:47.823 { 00:05:47.823 "name": "Passthru0", 00:05:47.823 "aliases": [ 00:05:47.823 "9fe615d5-1abe-5185-87f1-00a55a85f25f" 00:05:47.823 ], 00:05:47.823 "product_name": "passthru", 00:05:47.823 "block_size": 512, 00:05:47.823 "num_blocks": 16384, 00:05:47.823 "uuid": "9fe615d5-1abe-5185-87f1-00a55a85f25f", 00:05:47.823 "assigned_rate_limits": { 00:05:47.823 "rw_ios_per_sec": 0, 00:05:47.823 "rw_mbytes_per_sec": 0, 00:05:47.823 "r_mbytes_per_sec": 0, 00:05:47.823 "w_mbytes_per_sec": 0 00:05:47.823 }, 00:05:47.823 "claimed": false, 00:05:47.823 "zoned": false, 00:05:47.823 "supported_io_types": { 00:05:47.823 "read": true, 00:05:47.823 "write": true, 00:05:47.823 "unmap": true, 00:05:47.823 "flush": true, 00:05:47.823 "reset": true, 00:05:47.823 "nvme_admin": false, 00:05:47.823 "nvme_io": false, 00:05:47.823 "nvme_io_md": false, 00:05:47.823 "write_zeroes": true, 00:05:47.823 "zcopy": true, 00:05:47.823 "get_zone_info": false, 00:05:47.823 "zone_management": false, 00:05:47.823 "zone_append": false, 00:05:47.823 "compare": false, 00:05:47.823 "compare_and_write": false, 00:05:47.823 "abort": true, 00:05:47.823 "seek_hole": false, 00:05:47.823 "seek_data": false, 00:05:47.823 "copy": true, 00:05:47.823 "nvme_iov_md": false 00:05:47.823 }, 00:05:47.823 "memory_domains": [ 00:05:47.823 { 00:05:47.823 "dma_device_id": "system", 00:05:47.823 "dma_device_type": 1 00:05:47.823 }, 00:05:47.823 { 00:05:47.823 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:05:47.823 "dma_device_type": 2 00:05:47.823 } 00:05:47.823 ], 00:05:47.823 "driver_specific": { 00:05:47.823 "passthru": { 00:05:47.823 "name": "Passthru0", 00:05:47.823 "base_bdev_name": "Malloc0" 00:05:47.823 } 00:05:47.823 } 00:05:47.823 } 00:05:47.823 ]' 00:05:47.823 17:27:18 rpc.rpc_integrity -- rpc/rpc.sh@21 -- # jq length 00:05:47.823 17:27:18 rpc.rpc_integrity -- rpc/rpc.sh@21 -- # '[' 2 == 2 ']' 00:05:47.823 17:27:18 rpc.rpc_integrity -- rpc/rpc.sh@23 -- # rpc_cmd bdev_passthru_delete Passthru0 00:05:47.823 17:27:18 rpc.rpc_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:47.823 17:27:18 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:47.823 17:27:18 rpc.rpc_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:05:47.823 17:27:18 rpc.rpc_integrity -- rpc/rpc.sh@24 -- # rpc_cmd bdev_malloc_delete Malloc0 00:05:47.823 17:27:18 rpc.rpc_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:47.823 17:27:18 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:47.823 17:27:19 rpc.rpc_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:05:47.823 17:27:19 rpc.rpc_integrity -- rpc/rpc.sh@25 -- # rpc_cmd bdev_get_bdevs 00:05:47.823 17:27:19 rpc.rpc_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:47.823 17:27:19 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:48.083 17:27:19 rpc.rpc_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:05:48.083 17:27:19 rpc.rpc_integrity -- rpc/rpc.sh@25 -- # bdevs='[]' 00:05:48.083 17:27:19 rpc.rpc_integrity -- rpc/rpc.sh@26 -- # jq length 00:05:48.083 17:27:19 rpc.rpc_integrity -- rpc/rpc.sh@26 -- # '[' 0 == 0 ']' 00:05:48.083 00:05:48.083 real 0m0.333s 00:05:48.083 user 0m0.201s 00:05:48.083 ************************************ 00:05:48.083 END TEST rpc_integrity 00:05:48.083 ************************************ 00:05:48.083 sys 0m0.053s 00:05:48.083 17:27:19 rpc.rpc_integrity -- common/autotest_common.sh@1126 -- # xtrace_disable 00:05:48.083 17:27:19 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:48.083 17:27:19 rpc -- rpc/rpc.sh@74 -- # run_test rpc_plugins rpc_plugins 00:05:48.083 17:27:19 rpc -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:05:48.083 17:27:19 rpc -- common/autotest_common.sh@1107 -- # xtrace_disable 00:05:48.083 17:27:19 rpc -- common/autotest_common.sh@10 -- # set +x 00:05:48.083 ************************************ 00:05:48.083 START TEST rpc_plugins 00:05:48.083 ************************************ 00:05:48.083 17:27:19 rpc.rpc_plugins -- common/autotest_common.sh@1125 -- # rpc_plugins 00:05:48.083 17:27:19 rpc.rpc_plugins -- rpc/rpc.sh@30 -- # rpc_cmd --plugin rpc_plugin create_malloc 00:05:48.083 17:27:19 rpc.rpc_plugins -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:48.083 17:27:19 rpc.rpc_plugins -- common/autotest_common.sh@10 -- # set +x 00:05:48.083 17:27:19 rpc.rpc_plugins -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:05:48.083 17:27:19 rpc.rpc_plugins -- rpc/rpc.sh@30 -- # malloc=Malloc1 00:05:48.083 17:27:19 rpc.rpc_plugins -- rpc/rpc.sh@31 -- # rpc_cmd bdev_get_bdevs 00:05:48.083 17:27:19 rpc.rpc_plugins -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:48.083 17:27:19 rpc.rpc_plugins -- common/autotest_common.sh@10 -- # set +x 00:05:48.083 17:27:19 rpc.rpc_plugins -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:05:48.083 17:27:19 rpc.rpc_plugins -- rpc/rpc.sh@31 -- # bdevs='[ 00:05:48.083 { 00:05:48.083 "name": "Malloc1", 00:05:48.083 "aliases": [ 00:05:48.083 "580b03ef-76a4-448d-b791-9e8467bc6f74" 00:05:48.083 ], 00:05:48.083 "product_name": "Malloc disk", 00:05:48.083 "block_size": 4096, 00:05:48.083 "num_blocks": 256, 00:05:48.083 "uuid": "580b03ef-76a4-448d-b791-9e8467bc6f74", 00:05:48.083 "assigned_rate_limits": { 00:05:48.083 "rw_ios_per_sec": 0, 00:05:48.083 "rw_mbytes_per_sec": 0, 00:05:48.083 "r_mbytes_per_sec": 0, 00:05:48.083 "w_mbytes_per_sec": 0 00:05:48.083 }, 00:05:48.083 "claimed": false, 00:05:48.083 "zoned": false, 00:05:48.083 "supported_io_types": { 00:05:48.083 "read": true, 00:05:48.083 "write": true, 00:05:48.083 "unmap": true, 00:05:48.083 "flush": true, 00:05:48.083 "reset": true, 00:05:48.083 "nvme_admin": false, 00:05:48.083 "nvme_io": false, 00:05:48.083 "nvme_io_md": false, 00:05:48.083 "write_zeroes": true, 00:05:48.083 "zcopy": true, 00:05:48.083 "get_zone_info": false, 00:05:48.083 "zone_management": false, 00:05:48.083 "zone_append": false, 00:05:48.083 "compare": false, 00:05:48.083 "compare_and_write": false, 00:05:48.083 "abort": true, 00:05:48.083 "seek_hole": false, 00:05:48.083 "seek_data": false, 00:05:48.083 "copy": true, 00:05:48.083 "nvme_iov_md": false 00:05:48.083 }, 00:05:48.083 "memory_domains": [ 00:05:48.083 { 00:05:48.083 "dma_device_id": "system", 00:05:48.083 "dma_device_type": 1 00:05:48.083 }, 00:05:48.083 { 00:05:48.083 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:05:48.083 "dma_device_type": 2 00:05:48.083 } 00:05:48.083 ], 00:05:48.083 "driver_specific": {} 00:05:48.083 } 00:05:48.083 ]' 00:05:48.083 17:27:19 rpc.rpc_plugins -- rpc/rpc.sh@32 -- # jq length 00:05:48.083 17:27:19 rpc.rpc_plugins -- rpc/rpc.sh@32 -- # '[' 1 == 1 ']' 00:05:48.083 17:27:19 rpc.rpc_plugins -- rpc/rpc.sh@34 -- # rpc_cmd --plugin rpc_plugin delete_malloc Malloc1 00:05:48.083 17:27:19 rpc.rpc_plugins -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:48.083 17:27:19 rpc.rpc_plugins -- common/autotest_common.sh@10 -- # set +x 00:05:48.083 17:27:19 rpc.rpc_plugins -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:05:48.083 17:27:19 rpc.rpc_plugins -- rpc/rpc.sh@35 -- # rpc_cmd bdev_get_bdevs 00:05:48.083 17:27:19 rpc.rpc_plugins -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:48.083 17:27:19 rpc.rpc_plugins -- common/autotest_common.sh@10 -- # set +x 00:05:48.083 17:27:19 rpc.rpc_plugins -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:05:48.083 17:27:19 rpc.rpc_plugins -- rpc/rpc.sh@35 -- # bdevs='[]' 00:05:48.083 17:27:19 rpc.rpc_plugins -- rpc/rpc.sh@36 -- # jq length 00:05:48.344 ************************************ 00:05:48.344 END TEST rpc_plugins 00:05:48.344 ************************************ 00:05:48.344 17:27:19 rpc.rpc_plugins -- rpc/rpc.sh@36 -- # '[' 0 == 0 ']' 00:05:48.344 00:05:48.344 real 0m0.166s 00:05:48.344 user 0m0.096s 00:05:48.344 sys 0m0.028s 00:05:48.344 17:27:19 rpc.rpc_plugins -- common/autotest_common.sh@1126 -- # xtrace_disable 00:05:48.344 17:27:19 rpc.rpc_plugins -- common/autotest_common.sh@10 -- # set +x 00:05:48.344 17:27:19 rpc -- rpc/rpc.sh@75 -- # run_test rpc_trace_cmd_test rpc_trace_cmd_test 00:05:48.344 17:27:19 rpc -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:05:48.344 17:27:19 rpc -- common/autotest_common.sh@1107 -- # xtrace_disable 00:05:48.345 17:27:19 rpc -- common/autotest_common.sh@10 -- # set +x 00:05:48.345 ************************************ 00:05:48.345 START TEST rpc_trace_cmd_test 00:05:48.345 ************************************ 00:05:48.345 17:27:19 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@1125 -- # rpc_trace_cmd_test 00:05:48.345 17:27:19 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@40 -- # local info 00:05:48.345 17:27:19 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@42 -- # rpc_cmd trace_get_info 00:05:48.345 17:27:19 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:48.345 17:27:19 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@10 -- # set +x 00:05:48.345 17:27:19 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:05:48.345 17:27:19 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@42 -- # info='{ 00:05:48.345 "tpoint_shm_path": "/dev/shm/spdk_tgt_trace.pid68812", 00:05:48.345 "tpoint_group_mask": "0x8", 00:05:48.345 "iscsi_conn": { 00:05:48.345 "mask": "0x2", 00:05:48.345 "tpoint_mask": "0x0" 00:05:48.345 }, 00:05:48.345 "scsi": { 00:05:48.345 "mask": "0x4", 00:05:48.345 "tpoint_mask": "0x0" 00:05:48.345 }, 00:05:48.345 "bdev": { 00:05:48.345 "mask": "0x8", 00:05:48.345 "tpoint_mask": "0xffffffffffffffff" 00:05:48.345 }, 00:05:48.345 "nvmf_rdma": { 00:05:48.345 "mask": "0x10", 00:05:48.345 "tpoint_mask": "0x0" 00:05:48.345 }, 00:05:48.345 "nvmf_tcp": { 00:05:48.345 "mask": "0x20", 00:05:48.345 "tpoint_mask": "0x0" 00:05:48.345 }, 00:05:48.345 "ftl": { 00:05:48.345 "mask": "0x40", 00:05:48.345 "tpoint_mask": "0x0" 00:05:48.345 }, 00:05:48.345 "blobfs": { 00:05:48.345 "mask": "0x80", 00:05:48.345 "tpoint_mask": "0x0" 00:05:48.345 }, 00:05:48.345 "dsa": { 00:05:48.345 "mask": "0x200", 00:05:48.345 "tpoint_mask": "0x0" 00:05:48.345 }, 00:05:48.345 "thread": { 00:05:48.345 "mask": "0x400", 00:05:48.345 "tpoint_mask": "0x0" 00:05:48.345 }, 00:05:48.345 "nvme_pcie": { 00:05:48.345 "mask": "0x800", 00:05:48.345 "tpoint_mask": "0x0" 00:05:48.345 }, 00:05:48.345 "iaa": { 00:05:48.345 "mask": "0x1000", 00:05:48.345 "tpoint_mask": "0x0" 00:05:48.345 }, 00:05:48.345 "nvme_tcp": { 00:05:48.345 "mask": "0x2000", 00:05:48.345 "tpoint_mask": "0x0" 00:05:48.345 }, 00:05:48.345 "bdev_nvme": { 00:05:48.345 "mask": "0x4000", 00:05:48.345 "tpoint_mask": "0x0" 00:05:48.345 }, 00:05:48.345 "sock": { 00:05:48.345 "mask": "0x8000", 00:05:48.345 "tpoint_mask": "0x0" 00:05:48.345 }, 00:05:48.345 "blob": { 00:05:48.345 "mask": "0x10000", 00:05:48.345 "tpoint_mask": "0x0" 00:05:48.345 }, 00:05:48.345 "bdev_raid": { 00:05:48.345 "mask": "0x20000", 00:05:48.345 "tpoint_mask": "0x0" 00:05:48.345 } 00:05:48.345 }' 00:05:48.345 17:27:19 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@43 -- # jq length 00:05:48.345 17:27:19 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@43 -- # '[' 18 -gt 2 ']' 00:05:48.345 17:27:19 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@44 -- # jq 'has("tpoint_group_mask")' 00:05:48.345 17:27:19 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@44 -- # '[' true = true ']' 00:05:48.345 17:27:19 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@45 -- # jq 'has("tpoint_shm_path")' 00:05:48.345 17:27:19 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@45 -- # '[' true = true ']' 00:05:48.345 17:27:19 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@46 -- # jq 'has("bdev")' 00:05:48.605 17:27:19 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@46 -- # '[' true = true ']' 00:05:48.605 17:27:19 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@47 -- # jq -r .bdev.tpoint_mask 00:05:48.605 17:27:19 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@47 -- # '[' 0xffffffffffffffff '!=' 0x0 ']' 00:05:48.605 00:05:48.605 real 0m0.247s 00:05:48.605 user 0m0.189s 00:05:48.605 sys 0m0.048s 00:05:48.605 17:27:19 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:05:48.605 17:27:19 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@10 -- # set +x 00:05:48.605 ************************************ 00:05:48.605 END TEST rpc_trace_cmd_test 00:05:48.605 ************************************ 00:05:48.605 17:27:19 rpc -- rpc/rpc.sh@76 -- # [[ 0 -eq 1 ]] 00:05:48.605 17:27:19 rpc -- rpc/rpc.sh@80 -- # rpc=rpc_cmd 00:05:48.605 17:27:19 rpc -- rpc/rpc.sh@81 -- # run_test rpc_daemon_integrity rpc_integrity 00:05:48.605 17:27:19 rpc -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:05:48.605 17:27:19 rpc -- common/autotest_common.sh@1107 -- # xtrace_disable 00:05:48.605 17:27:19 rpc -- common/autotest_common.sh@10 -- # set +x 00:05:48.605 ************************************ 00:05:48.605 START TEST rpc_daemon_integrity 00:05:48.605 ************************************ 00:05:48.605 17:27:19 rpc.rpc_daemon_integrity -- common/autotest_common.sh@1125 -- # rpc_integrity 00:05:48.605 17:27:19 rpc.rpc_daemon_integrity -- rpc/rpc.sh@12 -- # rpc_cmd bdev_get_bdevs 00:05:48.605 17:27:19 rpc.rpc_daemon_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:48.605 17:27:19 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:48.605 17:27:19 rpc.rpc_daemon_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:05:48.605 17:27:19 rpc.rpc_daemon_integrity -- rpc/rpc.sh@12 -- # bdevs='[]' 00:05:48.605 17:27:19 rpc.rpc_daemon_integrity -- rpc/rpc.sh@13 -- # jq length 00:05:48.605 17:27:19 rpc.rpc_daemon_integrity -- rpc/rpc.sh@13 -- # '[' 0 == 0 ']' 00:05:48.605 17:27:19 rpc.rpc_daemon_integrity -- rpc/rpc.sh@15 -- # rpc_cmd bdev_malloc_create 8 512 00:05:48.605 17:27:19 rpc.rpc_daemon_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:48.605 17:27:19 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:48.605 17:27:19 rpc.rpc_daemon_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:05:48.605 17:27:19 rpc.rpc_daemon_integrity -- rpc/rpc.sh@15 -- # malloc=Malloc2 00:05:48.605 17:27:19 rpc.rpc_daemon_integrity -- rpc/rpc.sh@16 -- # rpc_cmd bdev_get_bdevs 00:05:48.605 17:27:19 rpc.rpc_daemon_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:48.605 17:27:19 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:48.605 17:27:19 rpc.rpc_daemon_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:05:48.605 17:27:19 rpc.rpc_daemon_integrity -- rpc/rpc.sh@16 -- # bdevs='[ 00:05:48.605 { 00:05:48.605 "name": "Malloc2", 00:05:48.605 "aliases": [ 00:05:48.606 "8e3f8e35-8f50-44a1-8dfd-53147cca3d96" 00:05:48.606 ], 00:05:48.606 "product_name": "Malloc disk", 00:05:48.606 "block_size": 512, 00:05:48.606 "num_blocks": 16384, 00:05:48.606 "uuid": "8e3f8e35-8f50-44a1-8dfd-53147cca3d96", 00:05:48.606 "assigned_rate_limits": { 00:05:48.606 "rw_ios_per_sec": 0, 00:05:48.606 "rw_mbytes_per_sec": 0, 00:05:48.606 "r_mbytes_per_sec": 0, 00:05:48.606 "w_mbytes_per_sec": 0 00:05:48.606 }, 00:05:48.606 "claimed": false, 00:05:48.606 "zoned": false, 00:05:48.606 "supported_io_types": { 00:05:48.606 "read": true, 00:05:48.606 "write": true, 00:05:48.606 "unmap": true, 00:05:48.606 "flush": true, 00:05:48.606 "reset": true, 00:05:48.606 "nvme_admin": false, 00:05:48.606 "nvme_io": false, 00:05:48.606 "nvme_io_md": false, 00:05:48.606 "write_zeroes": true, 00:05:48.606 "zcopy": true, 00:05:48.606 "get_zone_info": false, 00:05:48.606 "zone_management": false, 00:05:48.606 "zone_append": false, 00:05:48.606 "compare": false, 00:05:48.606 "compare_and_write": false, 00:05:48.606 "abort": true, 00:05:48.606 "seek_hole": false, 00:05:48.606 "seek_data": false, 00:05:48.606 "copy": true, 00:05:48.606 "nvme_iov_md": false 00:05:48.606 }, 00:05:48.606 "memory_domains": [ 00:05:48.606 { 00:05:48.606 "dma_device_id": "system", 00:05:48.606 "dma_device_type": 1 00:05:48.606 }, 00:05:48.606 { 00:05:48.606 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:05:48.606 "dma_device_type": 2 00:05:48.606 } 00:05:48.606 ], 00:05:48.606 "driver_specific": {} 00:05:48.606 } 00:05:48.606 ]' 00:05:48.606 17:27:19 rpc.rpc_daemon_integrity -- rpc/rpc.sh@17 -- # jq length 00:05:48.866 17:27:19 rpc.rpc_daemon_integrity -- rpc/rpc.sh@17 -- # '[' 1 == 1 ']' 00:05:48.866 17:27:19 rpc.rpc_daemon_integrity -- rpc/rpc.sh@19 -- # rpc_cmd bdev_passthru_create -b Malloc2 -p Passthru0 00:05:48.866 17:27:19 rpc.rpc_daemon_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:48.866 17:27:19 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:48.866 [2024-11-27 17:27:19.831651] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on Malloc2 00:05:48.866 [2024-11-27 17:27:19.831720] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:05:48.866 [2024-11-27 17:27:19.831746] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:05:48.866 [2024-11-27 17:27:19.831757] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:05:48.866 [2024-11-27 17:27:19.834511] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:05:48.866 [2024-11-27 17:27:19.834595] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: Passthru0 00:05:48.866 Passthru0 00:05:48.866 17:27:19 rpc.rpc_daemon_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:05:48.866 17:27:19 rpc.rpc_daemon_integrity -- rpc/rpc.sh@20 -- # rpc_cmd bdev_get_bdevs 00:05:48.866 17:27:19 rpc.rpc_daemon_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:48.866 17:27:19 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:48.866 17:27:19 rpc.rpc_daemon_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:05:48.866 17:27:19 rpc.rpc_daemon_integrity -- rpc/rpc.sh@20 -- # bdevs='[ 00:05:48.866 { 00:05:48.866 "name": "Malloc2", 00:05:48.866 "aliases": [ 00:05:48.866 "8e3f8e35-8f50-44a1-8dfd-53147cca3d96" 00:05:48.866 ], 00:05:48.866 "product_name": "Malloc disk", 00:05:48.866 "block_size": 512, 00:05:48.866 "num_blocks": 16384, 00:05:48.866 "uuid": "8e3f8e35-8f50-44a1-8dfd-53147cca3d96", 00:05:48.866 "assigned_rate_limits": { 00:05:48.866 "rw_ios_per_sec": 0, 00:05:48.866 "rw_mbytes_per_sec": 0, 00:05:48.866 "r_mbytes_per_sec": 0, 00:05:48.866 "w_mbytes_per_sec": 0 00:05:48.866 }, 00:05:48.866 "claimed": true, 00:05:48.866 "claim_type": "exclusive_write", 00:05:48.866 "zoned": false, 00:05:48.866 "supported_io_types": { 00:05:48.866 "read": true, 00:05:48.866 "write": true, 00:05:48.866 "unmap": true, 00:05:48.866 "flush": true, 00:05:48.866 "reset": true, 00:05:48.866 "nvme_admin": false, 00:05:48.866 "nvme_io": false, 00:05:48.866 "nvme_io_md": false, 00:05:48.866 "write_zeroes": true, 00:05:48.866 "zcopy": true, 00:05:48.866 "get_zone_info": false, 00:05:48.866 "zone_management": false, 00:05:48.866 "zone_append": false, 00:05:48.866 "compare": false, 00:05:48.866 "compare_and_write": false, 00:05:48.866 "abort": true, 00:05:48.866 "seek_hole": false, 00:05:48.866 "seek_data": false, 00:05:48.866 "copy": true, 00:05:48.866 "nvme_iov_md": false 00:05:48.866 }, 00:05:48.866 "memory_domains": [ 00:05:48.866 { 00:05:48.866 "dma_device_id": "system", 00:05:48.866 "dma_device_type": 1 00:05:48.866 }, 00:05:48.866 { 00:05:48.866 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:05:48.866 "dma_device_type": 2 00:05:48.866 } 00:05:48.866 ], 00:05:48.866 "driver_specific": {} 00:05:48.866 }, 00:05:48.866 { 00:05:48.866 "name": "Passthru0", 00:05:48.866 "aliases": [ 00:05:48.866 "4392581d-3b7b-5786-8eca-70eb2403bc2d" 00:05:48.866 ], 00:05:48.866 "product_name": "passthru", 00:05:48.866 "block_size": 512, 00:05:48.866 "num_blocks": 16384, 00:05:48.866 "uuid": "4392581d-3b7b-5786-8eca-70eb2403bc2d", 00:05:48.866 "assigned_rate_limits": { 00:05:48.866 "rw_ios_per_sec": 0, 00:05:48.866 "rw_mbytes_per_sec": 0, 00:05:48.866 "r_mbytes_per_sec": 0, 00:05:48.866 "w_mbytes_per_sec": 0 00:05:48.866 }, 00:05:48.866 "claimed": false, 00:05:48.866 "zoned": false, 00:05:48.867 "supported_io_types": { 00:05:48.867 "read": true, 00:05:48.867 "write": true, 00:05:48.867 "unmap": true, 00:05:48.867 "flush": true, 00:05:48.867 "reset": true, 00:05:48.867 "nvme_admin": false, 00:05:48.867 "nvme_io": false, 00:05:48.867 "nvme_io_md": false, 00:05:48.867 "write_zeroes": true, 00:05:48.867 "zcopy": true, 00:05:48.867 "get_zone_info": false, 00:05:48.867 "zone_management": false, 00:05:48.867 "zone_append": false, 00:05:48.867 "compare": false, 00:05:48.867 "compare_and_write": false, 00:05:48.867 "abort": true, 00:05:48.867 "seek_hole": false, 00:05:48.867 "seek_data": false, 00:05:48.867 "copy": true, 00:05:48.867 "nvme_iov_md": false 00:05:48.867 }, 00:05:48.867 "memory_domains": [ 00:05:48.867 { 00:05:48.867 "dma_device_id": "system", 00:05:48.867 "dma_device_type": 1 00:05:48.867 }, 00:05:48.867 { 00:05:48.867 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:05:48.867 "dma_device_type": 2 00:05:48.867 } 00:05:48.867 ], 00:05:48.867 "driver_specific": { 00:05:48.867 "passthru": { 00:05:48.867 "name": "Passthru0", 00:05:48.867 "base_bdev_name": "Malloc2" 00:05:48.867 } 00:05:48.867 } 00:05:48.867 } 00:05:48.867 ]' 00:05:48.867 17:27:19 rpc.rpc_daemon_integrity -- rpc/rpc.sh@21 -- # jq length 00:05:48.867 17:27:19 rpc.rpc_daemon_integrity -- rpc/rpc.sh@21 -- # '[' 2 == 2 ']' 00:05:48.867 17:27:19 rpc.rpc_daemon_integrity -- rpc/rpc.sh@23 -- # rpc_cmd bdev_passthru_delete Passthru0 00:05:48.867 17:27:19 rpc.rpc_daemon_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:48.867 17:27:19 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:48.867 17:27:19 rpc.rpc_daemon_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:05:48.867 17:27:19 rpc.rpc_daemon_integrity -- rpc/rpc.sh@24 -- # rpc_cmd bdev_malloc_delete Malloc2 00:05:48.867 17:27:19 rpc.rpc_daemon_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:48.867 17:27:19 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:48.867 17:27:19 rpc.rpc_daemon_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:05:48.867 17:27:19 rpc.rpc_daemon_integrity -- rpc/rpc.sh@25 -- # rpc_cmd bdev_get_bdevs 00:05:48.867 17:27:19 rpc.rpc_daemon_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:48.867 17:27:19 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:48.867 17:27:19 rpc.rpc_daemon_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:05:48.867 17:27:19 rpc.rpc_daemon_integrity -- rpc/rpc.sh@25 -- # bdevs='[]' 00:05:48.867 17:27:19 rpc.rpc_daemon_integrity -- rpc/rpc.sh@26 -- # jq length 00:05:48.867 ************************************ 00:05:48.867 END TEST rpc_daemon_integrity 00:05:48.867 ************************************ 00:05:48.867 17:27:19 rpc.rpc_daemon_integrity -- rpc/rpc.sh@26 -- # '[' 0 == 0 ']' 00:05:48.867 00:05:48.867 real 0m0.314s 00:05:48.867 user 0m0.185s 00:05:48.867 sys 0m0.055s 00:05:48.867 17:27:19 rpc.rpc_daemon_integrity -- common/autotest_common.sh@1126 -- # xtrace_disable 00:05:48.867 17:27:19 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:05:48.867 17:27:20 rpc -- rpc/rpc.sh@83 -- # trap - SIGINT SIGTERM EXIT 00:05:48.867 17:27:20 rpc -- rpc/rpc.sh@84 -- # killprocess 68812 00:05:48.867 17:27:20 rpc -- common/autotest_common.sh@950 -- # '[' -z 68812 ']' 00:05:48.867 17:27:20 rpc -- common/autotest_common.sh@954 -- # kill -0 68812 00:05:48.867 17:27:20 rpc -- common/autotest_common.sh@955 -- # uname 00:05:48.867 17:27:20 rpc -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:05:48.867 17:27:20 rpc -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 68812 00:05:49.127 17:27:20 rpc -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:05:49.127 17:27:20 rpc -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:05:49.127 killing process with pid 68812 00:05:49.127 17:27:20 rpc -- common/autotest_common.sh@968 -- # echo 'killing process with pid 68812' 00:05:49.127 17:27:20 rpc -- common/autotest_common.sh@969 -- # kill 68812 00:05:49.127 17:27:20 rpc -- common/autotest_common.sh@974 -- # wait 68812 00:05:49.698 ************************************ 00:05:49.698 END TEST rpc 00:05:49.698 ************************************ 00:05:49.698 00:05:49.698 real 0m3.162s 00:05:49.698 user 0m3.551s 00:05:49.698 sys 0m1.050s 00:05:49.698 17:27:20 rpc -- common/autotest_common.sh@1126 -- # xtrace_disable 00:05:49.698 17:27:20 rpc -- common/autotest_common.sh@10 -- # set +x 00:05:49.698 17:27:20 -- spdk/autotest.sh@157 -- # run_test skip_rpc /home/vagrant/spdk_repo/spdk/test/rpc/skip_rpc.sh 00:05:49.698 17:27:20 -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:05:49.698 17:27:20 -- common/autotest_common.sh@1107 -- # xtrace_disable 00:05:49.698 17:27:20 -- common/autotest_common.sh@10 -- # set +x 00:05:49.698 ************************************ 00:05:49.698 START TEST skip_rpc 00:05:49.698 ************************************ 00:05:49.698 17:27:20 skip_rpc -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/rpc/skip_rpc.sh 00:05:49.958 * Looking for test storage... 00:05:49.958 * Found test storage at /home/vagrant/spdk_repo/spdk/test/rpc 00:05:49.958 17:27:20 skip_rpc -- common/autotest_common.sh@1680 -- # [[ y == y ]] 00:05:49.958 17:27:20 skip_rpc -- common/autotest_common.sh@1681 -- # lcov --version 00:05:49.958 17:27:20 skip_rpc -- common/autotest_common.sh@1681 -- # awk '{print $NF}' 00:05:49.958 17:27:21 skip_rpc -- common/autotest_common.sh@1681 -- # lt 1.15 2 00:05:49.958 17:27:21 skip_rpc -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:05:49.958 17:27:21 skip_rpc -- scripts/common.sh@333 -- # local ver1 ver1_l 00:05:49.958 17:27:21 skip_rpc -- scripts/common.sh@334 -- # local ver2 ver2_l 00:05:49.958 17:27:21 skip_rpc -- scripts/common.sh@336 -- # IFS=.-: 00:05:49.958 17:27:21 skip_rpc -- scripts/common.sh@336 -- # read -ra ver1 00:05:49.958 17:27:21 skip_rpc -- scripts/common.sh@337 -- # IFS=.-: 00:05:49.958 17:27:21 skip_rpc -- scripts/common.sh@337 -- # read -ra ver2 00:05:49.958 17:27:21 skip_rpc -- scripts/common.sh@338 -- # local 'op=<' 00:05:49.958 17:27:21 skip_rpc -- scripts/common.sh@340 -- # ver1_l=2 00:05:49.958 17:27:21 skip_rpc -- scripts/common.sh@341 -- # ver2_l=1 00:05:49.958 17:27:21 skip_rpc -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:05:49.958 17:27:21 skip_rpc -- scripts/common.sh@344 -- # case "$op" in 00:05:49.958 17:27:21 skip_rpc -- scripts/common.sh@345 -- # : 1 00:05:49.959 17:27:21 skip_rpc -- scripts/common.sh@364 -- # (( v = 0 )) 00:05:49.959 17:27:21 skip_rpc -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:05:49.959 17:27:21 skip_rpc -- scripts/common.sh@365 -- # decimal 1 00:05:49.959 17:27:21 skip_rpc -- scripts/common.sh@353 -- # local d=1 00:05:49.959 17:27:21 skip_rpc -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:05:49.959 17:27:21 skip_rpc -- scripts/common.sh@355 -- # echo 1 00:05:49.959 17:27:21 skip_rpc -- scripts/common.sh@365 -- # ver1[v]=1 00:05:49.959 17:27:21 skip_rpc -- scripts/common.sh@366 -- # decimal 2 00:05:49.959 17:27:21 skip_rpc -- scripts/common.sh@353 -- # local d=2 00:05:49.959 17:27:21 skip_rpc -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:05:49.959 17:27:21 skip_rpc -- scripts/common.sh@355 -- # echo 2 00:05:49.959 17:27:21 skip_rpc -- scripts/common.sh@366 -- # ver2[v]=2 00:05:49.959 17:27:21 skip_rpc -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:05:49.959 17:27:21 skip_rpc -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:05:49.959 17:27:21 skip_rpc -- scripts/common.sh@368 -- # return 0 00:05:49.959 17:27:21 skip_rpc -- common/autotest_common.sh@1682 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:05:49.959 17:27:21 skip_rpc -- common/autotest_common.sh@1694 -- # export 'LCOV_OPTS= 00:05:49.959 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:49.959 --rc genhtml_branch_coverage=1 00:05:49.959 --rc genhtml_function_coverage=1 00:05:49.959 --rc genhtml_legend=1 00:05:49.959 --rc geninfo_all_blocks=1 00:05:49.959 --rc geninfo_unexecuted_blocks=1 00:05:49.959 00:05:49.959 ' 00:05:49.959 17:27:21 skip_rpc -- common/autotest_common.sh@1694 -- # LCOV_OPTS=' 00:05:49.959 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:49.959 --rc genhtml_branch_coverage=1 00:05:49.959 --rc genhtml_function_coverage=1 00:05:49.959 --rc genhtml_legend=1 00:05:49.959 --rc geninfo_all_blocks=1 00:05:49.959 --rc geninfo_unexecuted_blocks=1 00:05:49.959 00:05:49.959 ' 00:05:49.959 17:27:21 skip_rpc -- common/autotest_common.sh@1695 -- # export 'LCOV=lcov 00:05:49.959 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:49.959 --rc genhtml_branch_coverage=1 00:05:49.959 --rc genhtml_function_coverage=1 00:05:49.959 --rc genhtml_legend=1 00:05:49.959 --rc geninfo_all_blocks=1 00:05:49.959 --rc geninfo_unexecuted_blocks=1 00:05:49.959 00:05:49.959 ' 00:05:49.959 17:27:21 skip_rpc -- common/autotest_common.sh@1695 -- # LCOV='lcov 00:05:49.959 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:05:49.959 --rc genhtml_branch_coverage=1 00:05:49.959 --rc genhtml_function_coverage=1 00:05:49.959 --rc genhtml_legend=1 00:05:49.959 --rc geninfo_all_blocks=1 00:05:49.959 --rc geninfo_unexecuted_blocks=1 00:05:49.959 00:05:49.959 ' 00:05:49.959 17:27:21 skip_rpc -- rpc/skip_rpc.sh@11 -- # CONFIG_PATH=/home/vagrant/spdk_repo/spdk/test/rpc/config.json 00:05:49.959 17:27:21 skip_rpc -- rpc/skip_rpc.sh@12 -- # LOG_PATH=/home/vagrant/spdk_repo/spdk/test/rpc/log.txt 00:05:49.959 17:27:21 skip_rpc -- rpc/skip_rpc.sh@73 -- # run_test skip_rpc test_skip_rpc 00:05:49.959 17:27:21 skip_rpc -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:05:49.959 17:27:21 skip_rpc -- common/autotest_common.sh@1107 -- # xtrace_disable 00:05:49.959 17:27:21 skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:05:49.959 ************************************ 00:05:49.959 START TEST skip_rpc 00:05:49.959 ************************************ 00:05:49.959 17:27:21 skip_rpc.skip_rpc -- common/autotest_common.sh@1125 -- # test_skip_rpc 00:05:49.959 17:27:21 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@16 -- # local spdk_pid=69019 00:05:49.959 17:27:21 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@15 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt --no-rpc-server -m 0x1 00:05:49.959 17:27:21 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@18 -- # trap 'killprocess $spdk_pid; exit 1' SIGINT SIGTERM EXIT 00:05:49.959 17:27:21 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@19 -- # sleep 5 00:05:50.220 [2024-11-27 17:27:21.153668] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:05:50.220 [2024-11-27 17:27:21.153871] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid69019 ] 00:05:50.220 [2024-11-27 17:27:21.297791] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:05:50.220 [2024-11-27 17:27:21.371360] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:05:55.498 17:27:26 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@21 -- # NOT rpc_cmd spdk_get_version 00:05:55.498 17:27:26 skip_rpc.skip_rpc -- common/autotest_common.sh@650 -- # local es=0 00:05:55.498 17:27:26 skip_rpc.skip_rpc -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd spdk_get_version 00:05:55.498 17:27:26 skip_rpc.skip_rpc -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:05:55.498 17:27:26 skip_rpc.skip_rpc -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:05:55.498 17:27:26 skip_rpc.skip_rpc -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:05:55.498 17:27:26 skip_rpc.skip_rpc -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:05:55.498 17:27:26 skip_rpc.skip_rpc -- common/autotest_common.sh@653 -- # rpc_cmd spdk_get_version 00:05:55.498 17:27:26 skip_rpc.skip_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:55.498 17:27:26 skip_rpc.skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:05:55.498 17:27:26 skip_rpc.skip_rpc -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:05:55.498 17:27:26 skip_rpc.skip_rpc -- common/autotest_common.sh@653 -- # es=1 00:05:55.498 17:27:26 skip_rpc.skip_rpc -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:05:55.498 17:27:26 skip_rpc.skip_rpc -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:05:55.498 17:27:26 skip_rpc.skip_rpc -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:05:55.498 17:27:26 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@22 -- # trap - SIGINT SIGTERM EXIT 00:05:55.498 17:27:26 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@23 -- # killprocess 69019 00:05:55.498 17:27:26 skip_rpc.skip_rpc -- common/autotest_common.sh@950 -- # '[' -z 69019 ']' 00:05:55.498 17:27:26 skip_rpc.skip_rpc -- common/autotest_common.sh@954 -- # kill -0 69019 00:05:55.498 17:27:26 skip_rpc.skip_rpc -- common/autotest_common.sh@955 -- # uname 00:05:55.498 17:27:26 skip_rpc.skip_rpc -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:05:55.498 17:27:26 skip_rpc.skip_rpc -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 69019 00:05:55.498 killing process with pid 69019 00:05:55.498 17:27:26 skip_rpc.skip_rpc -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:05:55.499 17:27:26 skip_rpc.skip_rpc -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:05:55.499 17:27:26 skip_rpc.skip_rpc -- common/autotest_common.sh@968 -- # echo 'killing process with pid 69019' 00:05:55.499 17:27:26 skip_rpc.skip_rpc -- common/autotest_common.sh@969 -- # kill 69019 00:05:55.499 17:27:26 skip_rpc.skip_rpc -- common/autotest_common.sh@974 -- # wait 69019 00:05:55.757 00:05:55.757 real 0m5.712s 00:05:55.757 user 0m5.156s 00:05:55.757 sys 0m0.484s 00:05:55.757 ************************************ 00:05:55.757 END TEST skip_rpc 00:05:55.757 ************************************ 00:05:55.757 17:27:26 skip_rpc.skip_rpc -- common/autotest_common.sh@1126 -- # xtrace_disable 00:05:55.757 17:27:26 skip_rpc.skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:05:55.757 17:27:26 skip_rpc -- rpc/skip_rpc.sh@74 -- # run_test skip_rpc_with_json test_skip_rpc_with_json 00:05:55.757 17:27:26 skip_rpc -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:05:55.757 17:27:26 skip_rpc -- common/autotest_common.sh@1107 -- # xtrace_disable 00:05:55.757 17:27:26 skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:05:55.757 ************************************ 00:05:55.757 START TEST skip_rpc_with_json 00:05:55.757 ************************************ 00:05:55.757 17:27:26 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@1125 -- # test_skip_rpc_with_json 00:05:55.757 17:27:26 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@44 -- # gen_json_config 00:05:55.757 17:27:26 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@28 -- # local spdk_pid=69106 00:05:55.757 17:27:26 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@27 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 00:05:55.757 17:27:26 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@30 -- # trap 'killprocess $spdk_pid; exit 1' SIGINT SIGTERM EXIT 00:05:55.757 17:27:26 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@31 -- # waitforlisten 69106 00:05:55.757 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:05:55.757 17:27:26 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@831 -- # '[' -z 69106 ']' 00:05:55.757 17:27:26 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:05:55.757 17:27:26 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@836 -- # local max_retries=100 00:05:55.757 17:27:26 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:05:55.757 17:27:26 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@840 -- # xtrace_disable 00:05:55.757 17:27:26 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@10 -- # set +x 00:05:55.757 [2024-11-27 17:27:26.933807] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:05:55.757 [2024-11-27 17:27:26.934021] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid69106 ] 00:05:56.015 [2024-11-27 17:27:27.080299] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:05:56.015 [2024-11-27 17:27:27.148922] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:05:56.587 17:27:27 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:05:56.587 17:27:27 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@864 -- # return 0 00:05:56.587 17:27:27 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@34 -- # rpc_cmd nvmf_get_transports --trtype tcp 00:05:56.587 17:27:27 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:56.587 17:27:27 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@10 -- # set +x 00:05:56.587 [2024-11-27 17:27:27.733247] nvmf_rpc.c:2703:rpc_nvmf_get_transports: *ERROR*: transport 'tcp' does not exist 00:05:56.587 request: 00:05:56.587 { 00:05:56.587 "trtype": "tcp", 00:05:56.587 "method": "nvmf_get_transports", 00:05:56.587 "req_id": 1 00:05:56.587 } 00:05:56.587 Got JSON-RPC error response 00:05:56.587 response: 00:05:56.587 { 00:05:56.587 "code": -19, 00:05:56.587 "message": "No such device" 00:05:56.587 } 00:05:56.587 17:27:27 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:05:56.587 17:27:27 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@34 -- # rpc_cmd nvmf_create_transport -t tcp 00:05:56.587 17:27:27 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:56.587 17:27:27 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@10 -- # set +x 00:05:56.587 [2024-11-27 17:27:27.745369] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:05:56.587 17:27:27 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:05:56.587 17:27:27 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@36 -- # rpc_cmd save_config 00:05:56.587 17:27:27 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@561 -- # xtrace_disable 00:05:56.587 17:27:27 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@10 -- # set +x 00:05:56.852 17:27:27 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:05:56.852 17:27:27 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@37 -- # cat /home/vagrant/spdk_repo/spdk/test/rpc/config.json 00:05:56.852 { 00:05:56.852 "subsystems": [ 00:05:56.852 { 00:05:56.852 "subsystem": "fsdev", 00:05:56.852 "config": [ 00:05:56.852 { 00:05:56.852 "method": "fsdev_set_opts", 00:05:56.852 "params": { 00:05:56.852 "fsdev_io_pool_size": 65535, 00:05:56.852 "fsdev_io_cache_size": 256 00:05:56.852 } 00:05:56.852 } 00:05:56.852 ] 00:05:56.852 }, 00:05:56.852 { 00:05:56.852 "subsystem": "keyring", 00:05:56.852 "config": [] 00:05:56.852 }, 00:05:56.852 { 00:05:56.852 "subsystem": "iobuf", 00:05:56.852 "config": [ 00:05:56.852 { 00:05:56.852 "method": "iobuf_set_options", 00:05:56.852 "params": { 00:05:56.852 "small_pool_count": 8192, 00:05:56.852 "large_pool_count": 1024, 00:05:56.852 "small_bufsize": 8192, 00:05:56.852 "large_bufsize": 135168 00:05:56.852 } 00:05:56.852 } 00:05:56.852 ] 00:05:56.852 }, 00:05:56.852 { 00:05:56.852 "subsystem": "sock", 00:05:56.852 "config": [ 00:05:56.852 { 00:05:56.852 "method": "sock_set_default_impl", 00:05:56.852 "params": { 00:05:56.852 "impl_name": "posix" 00:05:56.852 } 00:05:56.852 }, 00:05:56.852 { 00:05:56.852 "method": "sock_impl_set_options", 00:05:56.852 "params": { 00:05:56.852 "impl_name": "ssl", 00:05:56.852 "recv_buf_size": 4096, 00:05:56.852 "send_buf_size": 4096, 00:05:56.852 "enable_recv_pipe": true, 00:05:56.852 "enable_quickack": false, 00:05:56.852 "enable_placement_id": 0, 00:05:56.852 "enable_zerocopy_send_server": true, 00:05:56.852 "enable_zerocopy_send_client": false, 00:05:56.852 "zerocopy_threshold": 0, 00:05:56.852 "tls_version": 0, 00:05:56.852 "enable_ktls": false 00:05:56.852 } 00:05:56.852 }, 00:05:56.852 { 00:05:56.852 "method": "sock_impl_set_options", 00:05:56.852 "params": { 00:05:56.852 "impl_name": "posix", 00:05:56.852 "recv_buf_size": 2097152, 00:05:56.852 "send_buf_size": 2097152, 00:05:56.852 "enable_recv_pipe": true, 00:05:56.852 "enable_quickack": false, 00:05:56.852 "enable_placement_id": 0, 00:05:56.852 "enable_zerocopy_send_server": true, 00:05:56.852 "enable_zerocopy_send_client": false, 00:05:56.852 "zerocopy_threshold": 0, 00:05:56.852 "tls_version": 0, 00:05:56.852 "enable_ktls": false 00:05:56.852 } 00:05:56.852 } 00:05:56.852 ] 00:05:56.852 }, 00:05:56.852 { 00:05:56.852 "subsystem": "vmd", 00:05:56.852 "config": [] 00:05:56.852 }, 00:05:56.852 { 00:05:56.852 "subsystem": "accel", 00:05:56.852 "config": [ 00:05:56.852 { 00:05:56.852 "method": "accel_set_options", 00:05:56.852 "params": { 00:05:56.852 "small_cache_size": 128, 00:05:56.852 "large_cache_size": 16, 00:05:56.852 "task_count": 2048, 00:05:56.852 "sequence_count": 2048, 00:05:56.852 "buf_count": 2048 00:05:56.852 } 00:05:56.852 } 00:05:56.852 ] 00:05:56.852 }, 00:05:56.852 { 00:05:56.852 "subsystem": "bdev", 00:05:56.852 "config": [ 00:05:56.852 { 00:05:56.852 "method": "bdev_set_options", 00:05:56.852 "params": { 00:05:56.852 "bdev_io_pool_size": 65535, 00:05:56.852 "bdev_io_cache_size": 256, 00:05:56.852 "bdev_auto_examine": true, 00:05:56.852 "iobuf_small_cache_size": 128, 00:05:56.852 "iobuf_large_cache_size": 16 00:05:56.852 } 00:05:56.852 }, 00:05:56.852 { 00:05:56.852 "method": "bdev_raid_set_options", 00:05:56.852 "params": { 00:05:56.852 "process_window_size_kb": 1024, 00:05:56.852 "process_max_bandwidth_mb_sec": 0 00:05:56.852 } 00:05:56.852 }, 00:05:56.852 { 00:05:56.852 "method": "bdev_iscsi_set_options", 00:05:56.852 "params": { 00:05:56.852 "timeout_sec": 30 00:05:56.852 } 00:05:56.852 }, 00:05:56.852 { 00:05:56.852 "method": "bdev_nvme_set_options", 00:05:56.852 "params": { 00:05:56.852 "action_on_timeout": "none", 00:05:56.852 "timeout_us": 0, 00:05:56.852 "timeout_admin_us": 0, 00:05:56.852 "keep_alive_timeout_ms": 10000, 00:05:56.852 "arbitration_burst": 0, 00:05:56.852 "low_priority_weight": 0, 00:05:56.852 "medium_priority_weight": 0, 00:05:56.852 "high_priority_weight": 0, 00:05:56.852 "nvme_adminq_poll_period_us": 10000, 00:05:56.852 "nvme_ioq_poll_period_us": 0, 00:05:56.852 "io_queue_requests": 0, 00:05:56.852 "delay_cmd_submit": true, 00:05:56.852 "transport_retry_count": 4, 00:05:56.852 "bdev_retry_count": 3, 00:05:56.852 "transport_ack_timeout": 0, 00:05:56.852 "ctrlr_loss_timeout_sec": 0, 00:05:56.852 "reconnect_delay_sec": 0, 00:05:56.852 "fast_io_fail_timeout_sec": 0, 00:05:56.852 "disable_auto_failback": false, 00:05:56.852 "generate_uuids": false, 00:05:56.852 "transport_tos": 0, 00:05:56.852 "nvme_error_stat": false, 00:05:56.852 "rdma_srq_size": 0, 00:05:56.852 "io_path_stat": false, 00:05:56.852 "allow_accel_sequence": false, 00:05:56.852 "rdma_max_cq_size": 0, 00:05:56.852 "rdma_cm_event_timeout_ms": 0, 00:05:56.852 "dhchap_digests": [ 00:05:56.852 "sha256", 00:05:56.852 "sha384", 00:05:56.852 "sha512" 00:05:56.852 ], 00:05:56.852 "dhchap_dhgroups": [ 00:05:56.852 "null", 00:05:56.852 "ffdhe2048", 00:05:56.852 "ffdhe3072", 00:05:56.852 "ffdhe4096", 00:05:56.852 "ffdhe6144", 00:05:56.852 "ffdhe8192" 00:05:56.852 ] 00:05:56.852 } 00:05:56.852 }, 00:05:56.852 { 00:05:56.852 "method": "bdev_nvme_set_hotplug", 00:05:56.852 "params": { 00:05:56.852 "period_us": 100000, 00:05:56.852 "enable": false 00:05:56.852 } 00:05:56.852 }, 00:05:56.852 { 00:05:56.852 "method": "bdev_wait_for_examine" 00:05:56.852 } 00:05:56.852 ] 00:05:56.852 }, 00:05:56.852 { 00:05:56.852 "subsystem": "scsi", 00:05:56.852 "config": null 00:05:56.852 }, 00:05:56.852 { 00:05:56.852 "subsystem": "scheduler", 00:05:56.852 "config": [ 00:05:56.852 { 00:05:56.852 "method": "framework_set_scheduler", 00:05:56.852 "params": { 00:05:56.852 "name": "static" 00:05:56.852 } 00:05:56.852 } 00:05:56.852 ] 00:05:56.852 }, 00:05:56.852 { 00:05:56.852 "subsystem": "vhost_scsi", 00:05:56.852 "config": [] 00:05:56.852 }, 00:05:56.852 { 00:05:56.852 "subsystem": "vhost_blk", 00:05:56.852 "config": [] 00:05:56.852 }, 00:05:56.852 { 00:05:56.852 "subsystem": "ublk", 00:05:56.852 "config": [] 00:05:56.852 }, 00:05:56.852 { 00:05:56.852 "subsystem": "nbd", 00:05:56.852 "config": [] 00:05:56.852 }, 00:05:56.852 { 00:05:56.852 "subsystem": "nvmf", 00:05:56.852 "config": [ 00:05:56.852 { 00:05:56.852 "method": "nvmf_set_config", 00:05:56.852 "params": { 00:05:56.852 "discovery_filter": "match_any", 00:05:56.852 "admin_cmd_passthru": { 00:05:56.852 "identify_ctrlr": false 00:05:56.852 }, 00:05:56.852 "dhchap_digests": [ 00:05:56.852 "sha256", 00:05:56.852 "sha384", 00:05:56.852 "sha512" 00:05:56.852 ], 00:05:56.852 "dhchap_dhgroups": [ 00:05:56.852 "null", 00:05:56.852 "ffdhe2048", 00:05:56.852 "ffdhe3072", 00:05:56.852 "ffdhe4096", 00:05:56.852 "ffdhe6144", 00:05:56.852 "ffdhe8192" 00:05:56.852 ] 00:05:56.852 } 00:05:56.852 }, 00:05:56.852 { 00:05:56.852 "method": "nvmf_set_max_subsystems", 00:05:56.852 "params": { 00:05:56.852 "max_subsystems": 1024 00:05:56.852 } 00:05:56.852 }, 00:05:56.852 { 00:05:56.852 "method": "nvmf_set_crdt", 00:05:56.852 "params": { 00:05:56.852 "crdt1": 0, 00:05:56.852 "crdt2": 0, 00:05:56.852 "crdt3": 0 00:05:56.852 } 00:05:56.852 }, 00:05:56.852 { 00:05:56.852 "method": "nvmf_create_transport", 00:05:56.852 "params": { 00:05:56.852 "trtype": "TCP", 00:05:56.852 "max_queue_depth": 128, 00:05:56.852 "max_io_qpairs_per_ctrlr": 127, 00:05:56.852 "in_capsule_data_size": 4096, 00:05:56.852 "max_io_size": 131072, 00:05:56.852 "io_unit_size": 131072, 00:05:56.852 "max_aq_depth": 128, 00:05:56.852 "num_shared_buffers": 511, 00:05:56.852 "buf_cache_size": 4294967295, 00:05:56.852 "dif_insert_or_strip": false, 00:05:56.852 "zcopy": false, 00:05:56.852 "c2h_success": true, 00:05:56.852 "sock_priority": 0, 00:05:56.852 "abort_timeout_sec": 1, 00:05:56.853 "ack_timeout": 0, 00:05:56.853 "data_wr_pool_size": 0 00:05:56.853 } 00:05:56.853 } 00:05:56.853 ] 00:05:56.853 }, 00:05:56.853 { 00:05:56.853 "subsystem": "iscsi", 00:05:56.853 "config": [ 00:05:56.853 { 00:05:56.853 "method": "iscsi_set_options", 00:05:56.853 "params": { 00:05:56.853 "node_base": "iqn.2016-06.io.spdk", 00:05:56.853 "max_sessions": 128, 00:05:56.853 "max_connections_per_session": 2, 00:05:56.853 "max_queue_depth": 64, 00:05:56.853 "default_time2wait": 2, 00:05:56.853 "default_time2retain": 20, 00:05:56.853 "first_burst_length": 8192, 00:05:56.853 "immediate_data": true, 00:05:56.853 "allow_duplicated_isid": false, 00:05:56.853 "error_recovery_level": 0, 00:05:56.853 "nop_timeout": 60, 00:05:56.853 "nop_in_interval": 30, 00:05:56.853 "disable_chap": false, 00:05:56.853 "require_chap": false, 00:05:56.853 "mutual_chap": false, 00:05:56.853 "chap_group": 0, 00:05:56.853 "max_large_datain_per_connection": 64, 00:05:56.853 "max_r2t_per_connection": 4, 00:05:56.853 "pdu_pool_size": 36864, 00:05:56.853 "immediate_data_pool_size": 16384, 00:05:56.853 "data_out_pool_size": 2048 00:05:56.853 } 00:05:56.853 } 00:05:56.853 ] 00:05:56.853 } 00:05:56.853 ] 00:05:56.853 } 00:05:56.853 17:27:27 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@39 -- # trap - SIGINT SIGTERM EXIT 00:05:56.853 17:27:27 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@40 -- # killprocess 69106 00:05:56.853 17:27:27 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@950 -- # '[' -z 69106 ']' 00:05:56.853 17:27:27 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@954 -- # kill -0 69106 00:05:56.853 17:27:27 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@955 -- # uname 00:05:56.853 17:27:27 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:05:56.853 17:27:27 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 69106 00:05:56.853 17:27:27 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:05:56.853 17:27:27 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:05:56.853 17:27:27 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@968 -- # echo 'killing process with pid 69106' 00:05:56.853 killing process with pid 69106 00:05:56.853 17:27:27 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@969 -- # kill 69106 00:05:56.853 17:27:27 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@974 -- # wait 69106 00:05:57.794 17:27:28 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@47 -- # local spdk_pid=69135 00:05:57.794 17:27:28 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@46 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt --no-rpc-server -m 0x1 --json /home/vagrant/spdk_repo/spdk/test/rpc/config.json 00:05:57.794 17:27:28 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@48 -- # sleep 5 00:06:03.076 17:27:33 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@50 -- # killprocess 69135 00:06:03.076 17:27:33 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@950 -- # '[' -z 69135 ']' 00:06:03.076 17:27:33 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@954 -- # kill -0 69135 00:06:03.076 17:27:33 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@955 -- # uname 00:06:03.076 17:27:33 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:06:03.076 17:27:33 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 69135 00:06:03.076 killing process with pid 69135 00:06:03.076 17:27:33 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:06:03.076 17:27:33 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:06:03.076 17:27:33 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@968 -- # echo 'killing process with pid 69135' 00:06:03.076 17:27:33 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@969 -- # kill 69135 00:06:03.076 17:27:33 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@974 -- # wait 69135 00:06:03.336 17:27:34 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@51 -- # grep -q 'TCP Transport Init' /home/vagrant/spdk_repo/spdk/test/rpc/log.txt 00:06:03.336 17:27:34 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@52 -- # rm /home/vagrant/spdk_repo/spdk/test/rpc/log.txt 00:06:03.336 00:06:03.336 real 0m7.482s 00:06:03.336 user 0m6.722s 00:06:03.336 sys 0m1.025s 00:06:03.336 17:27:34 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@1126 -- # xtrace_disable 00:06:03.336 ************************************ 00:06:03.336 END TEST skip_rpc_with_json 00:06:03.336 ************************************ 00:06:03.336 17:27:34 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@10 -- # set +x 00:06:03.336 17:27:34 skip_rpc -- rpc/skip_rpc.sh@75 -- # run_test skip_rpc_with_delay test_skip_rpc_with_delay 00:06:03.336 17:27:34 skip_rpc -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:06:03.336 17:27:34 skip_rpc -- common/autotest_common.sh@1107 -- # xtrace_disable 00:06:03.336 17:27:34 skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:03.336 ************************************ 00:06:03.336 START TEST skip_rpc_with_delay 00:06:03.336 ************************************ 00:06:03.336 17:27:34 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@1125 -- # test_skip_rpc_with_delay 00:06:03.336 17:27:34 skip_rpc.skip_rpc_with_delay -- rpc/skip_rpc.sh@57 -- # NOT /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt --no-rpc-server -m 0x1 --wait-for-rpc 00:06:03.336 17:27:34 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@650 -- # local es=0 00:06:03.336 17:27:34 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@652 -- # valid_exec_arg /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt --no-rpc-server -m 0x1 --wait-for-rpc 00:06:03.336 17:27:34 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@638 -- # local arg=/home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:06:03.336 17:27:34 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:06:03.336 17:27:34 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@642 -- # type -t /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:06:03.336 17:27:34 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:06:03.336 17:27:34 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@644 -- # type -P /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:06:03.336 17:27:34 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:06:03.336 17:27:34 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@644 -- # arg=/home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:06:03.336 17:27:34 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@644 -- # [[ -x /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt ]] 00:06:03.336 17:27:34 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@653 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt --no-rpc-server -m 0x1 --wait-for-rpc 00:06:03.336 [2024-11-27 17:27:34.485295] app.c: 840:spdk_app_start: *ERROR*: Cannot use '--wait-for-rpc' if no RPC server is going to be started. 00:06:03.336 [2024-11-27 17:27:34.485426] app.c: 719:unclaim_cpu_cores: *ERROR*: Failed to unlink lock fd for core 0, errno: 2 00:06:03.596 17:27:34 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@653 -- # es=1 00:06:03.596 17:27:34 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:06:03.596 17:27:34 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:06:03.596 17:27:34 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:06:03.596 00:06:03.596 real 0m0.161s 00:06:03.596 user 0m0.087s 00:06:03.596 sys 0m0.073s 00:06:03.596 17:27:34 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@1126 -- # xtrace_disable 00:06:03.596 17:27:34 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@10 -- # set +x 00:06:03.596 ************************************ 00:06:03.596 END TEST skip_rpc_with_delay 00:06:03.596 ************************************ 00:06:03.596 17:27:34 skip_rpc -- rpc/skip_rpc.sh@77 -- # uname 00:06:03.596 17:27:34 skip_rpc -- rpc/skip_rpc.sh@77 -- # '[' Linux '!=' FreeBSD ']' 00:06:03.596 17:27:34 skip_rpc -- rpc/skip_rpc.sh@78 -- # run_test exit_on_failed_rpc_init test_exit_on_failed_rpc_init 00:06:03.596 17:27:34 skip_rpc -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:06:03.596 17:27:34 skip_rpc -- common/autotest_common.sh@1107 -- # xtrace_disable 00:06:03.596 17:27:34 skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:03.596 ************************************ 00:06:03.596 START TEST exit_on_failed_rpc_init 00:06:03.596 ************************************ 00:06:03.596 17:27:34 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@1125 -- # test_exit_on_failed_rpc_init 00:06:03.596 17:27:34 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@62 -- # local spdk_pid=69252 00:06:03.596 17:27:34 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@61 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 00:06:03.596 17:27:34 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@63 -- # waitforlisten 69252 00:06:03.596 17:27:34 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@831 -- # '[' -z 69252 ']' 00:06:03.596 17:27:34 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:03.596 17:27:34 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@836 -- # local max_retries=100 00:06:03.596 17:27:34 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:03.596 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:03.596 17:27:34 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@840 -- # xtrace_disable 00:06:03.596 17:27:34 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@10 -- # set +x 00:06:03.596 [2024-11-27 17:27:34.721393] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:06:03.596 [2024-11-27 17:27:34.721518] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid69252 ] 00:06:03.868 [2024-11-27 17:27:34.865693] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:03.868 [2024-11-27 17:27:34.933633] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:06:04.455 17:27:35 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:06:04.455 17:27:35 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@864 -- # return 0 00:06:04.455 17:27:35 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@65 -- # trap 'killprocess $spdk_pid; exit 1' SIGINT SIGTERM EXIT 00:06:04.455 17:27:35 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@67 -- # NOT /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x2 00:06:04.455 17:27:35 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@650 -- # local es=0 00:06:04.455 17:27:35 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@652 -- # valid_exec_arg /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x2 00:06:04.455 17:27:35 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@638 -- # local arg=/home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:06:04.455 17:27:35 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:06:04.455 17:27:35 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@642 -- # type -t /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:06:04.455 17:27:35 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:06:04.455 17:27:35 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@644 -- # type -P /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:06:04.455 17:27:35 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:06:04.455 17:27:35 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@644 -- # arg=/home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:06:04.455 17:27:35 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@644 -- # [[ -x /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt ]] 00:06:04.455 17:27:35 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@653 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x2 00:06:04.455 [2024-11-27 17:27:35.628892] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:06:04.455 [2024-11-27 17:27:35.629092] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid69270 ] 00:06:04.714 [2024-11-27 17:27:35.772428] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:04.714 [2024-11-27 17:27:35.818050] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 1 00:06:04.714 [2024-11-27 17:27:35.818253] rpc.c: 180:_spdk_rpc_listen: *ERROR*: RPC Unix domain socket path /var/tmp/spdk.sock in use. Specify another. 00:06:04.714 [2024-11-27 17:27:35.818309] rpc.c: 166:spdk_rpc_initialize: *ERROR*: Unable to start RPC service at /var/tmp/spdk.sock 00:06:04.714 [2024-11-27 17:27:35.818392] app.c:1061:spdk_app_stop: *WARNING*: spdk_app_stop'd on non-zero 00:06:04.975 17:27:35 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@653 -- # es=234 00:06:04.975 17:27:35 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:06:04.975 17:27:35 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@662 -- # es=106 00:06:04.975 17:27:35 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@663 -- # case "$es" in 00:06:04.975 17:27:35 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@670 -- # es=1 00:06:04.975 17:27:35 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:06:04.975 17:27:35 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@69 -- # trap - SIGINT SIGTERM EXIT 00:06:04.975 17:27:35 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@70 -- # killprocess 69252 00:06:04.975 17:27:35 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@950 -- # '[' -z 69252 ']' 00:06:04.975 17:27:35 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@954 -- # kill -0 69252 00:06:04.975 17:27:35 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@955 -- # uname 00:06:04.975 17:27:35 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:06:04.975 17:27:35 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 69252 00:06:04.975 17:27:35 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:06:04.975 17:27:35 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:06:04.975 17:27:35 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@968 -- # echo 'killing process with pid 69252' 00:06:04.975 killing process with pid 69252 00:06:04.975 17:27:35 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@969 -- # kill 69252 00:06:04.975 17:27:35 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@974 -- # wait 69252 00:06:05.546 00:06:05.546 real 0m1.987s 00:06:05.546 user 0m1.986s 00:06:05.546 sys 0m0.612s 00:06:05.546 17:27:36 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@1126 -- # xtrace_disable 00:06:05.546 17:27:36 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@10 -- # set +x 00:06:05.546 ************************************ 00:06:05.546 END TEST exit_on_failed_rpc_init 00:06:05.546 ************************************ 00:06:05.546 17:27:36 skip_rpc -- rpc/skip_rpc.sh@81 -- # rm /home/vagrant/spdk_repo/spdk/test/rpc/config.json 00:06:05.546 00:06:05.546 real 0m15.867s 00:06:05.546 user 0m14.158s 00:06:05.546 sys 0m2.524s 00:06:05.546 17:27:36 skip_rpc -- common/autotest_common.sh@1126 -- # xtrace_disable 00:06:05.546 ************************************ 00:06:05.546 END TEST skip_rpc 00:06:05.546 ************************************ 00:06:05.546 17:27:36 skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:05.546 17:27:36 -- spdk/autotest.sh@158 -- # run_test rpc_client /home/vagrant/spdk_repo/spdk/test/rpc_client/rpc_client.sh 00:06:05.546 17:27:36 -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:06:05.546 17:27:36 -- common/autotest_common.sh@1107 -- # xtrace_disable 00:06:05.546 17:27:36 -- common/autotest_common.sh@10 -- # set +x 00:06:05.546 ************************************ 00:06:05.546 START TEST rpc_client 00:06:05.546 ************************************ 00:06:05.546 17:27:36 rpc_client -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/rpc_client/rpc_client.sh 00:06:05.807 * Looking for test storage... 00:06:05.807 * Found test storage at /home/vagrant/spdk_repo/spdk/test/rpc_client 00:06:05.807 17:27:36 rpc_client -- common/autotest_common.sh@1680 -- # [[ y == y ]] 00:06:05.807 17:27:36 rpc_client -- common/autotest_common.sh@1681 -- # awk '{print $NF}' 00:06:05.807 17:27:36 rpc_client -- common/autotest_common.sh@1681 -- # lcov --version 00:06:05.807 17:27:36 rpc_client -- common/autotest_common.sh@1681 -- # lt 1.15 2 00:06:05.807 17:27:36 rpc_client -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:06:05.807 17:27:36 rpc_client -- scripts/common.sh@333 -- # local ver1 ver1_l 00:06:05.807 17:27:36 rpc_client -- scripts/common.sh@334 -- # local ver2 ver2_l 00:06:05.807 17:27:36 rpc_client -- scripts/common.sh@336 -- # IFS=.-: 00:06:05.807 17:27:36 rpc_client -- scripts/common.sh@336 -- # read -ra ver1 00:06:05.807 17:27:36 rpc_client -- scripts/common.sh@337 -- # IFS=.-: 00:06:05.807 17:27:36 rpc_client -- scripts/common.sh@337 -- # read -ra ver2 00:06:05.807 17:27:36 rpc_client -- scripts/common.sh@338 -- # local 'op=<' 00:06:05.807 17:27:36 rpc_client -- scripts/common.sh@340 -- # ver1_l=2 00:06:05.807 17:27:36 rpc_client -- scripts/common.sh@341 -- # ver2_l=1 00:06:05.807 17:27:36 rpc_client -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:06:05.807 17:27:36 rpc_client -- scripts/common.sh@344 -- # case "$op" in 00:06:05.807 17:27:36 rpc_client -- scripts/common.sh@345 -- # : 1 00:06:05.807 17:27:36 rpc_client -- scripts/common.sh@364 -- # (( v = 0 )) 00:06:05.807 17:27:36 rpc_client -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:06:05.807 17:27:36 rpc_client -- scripts/common.sh@365 -- # decimal 1 00:06:05.807 17:27:36 rpc_client -- scripts/common.sh@353 -- # local d=1 00:06:05.807 17:27:36 rpc_client -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:06:05.807 17:27:36 rpc_client -- scripts/common.sh@355 -- # echo 1 00:06:05.807 17:27:36 rpc_client -- scripts/common.sh@365 -- # ver1[v]=1 00:06:05.807 17:27:36 rpc_client -- scripts/common.sh@366 -- # decimal 2 00:06:05.807 17:27:36 rpc_client -- scripts/common.sh@353 -- # local d=2 00:06:05.807 17:27:36 rpc_client -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:06:05.807 17:27:36 rpc_client -- scripts/common.sh@355 -- # echo 2 00:06:05.807 17:27:36 rpc_client -- scripts/common.sh@366 -- # ver2[v]=2 00:06:05.807 17:27:36 rpc_client -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:06:05.807 17:27:36 rpc_client -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:06:05.807 17:27:36 rpc_client -- scripts/common.sh@368 -- # return 0 00:06:05.807 17:27:36 rpc_client -- common/autotest_common.sh@1682 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:06:05.807 17:27:36 rpc_client -- common/autotest_common.sh@1694 -- # export 'LCOV_OPTS= 00:06:05.807 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:05.807 --rc genhtml_branch_coverage=1 00:06:05.807 --rc genhtml_function_coverage=1 00:06:05.807 --rc genhtml_legend=1 00:06:05.807 --rc geninfo_all_blocks=1 00:06:05.807 --rc geninfo_unexecuted_blocks=1 00:06:05.807 00:06:05.807 ' 00:06:05.807 17:27:36 rpc_client -- common/autotest_common.sh@1694 -- # LCOV_OPTS=' 00:06:05.807 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:05.807 --rc genhtml_branch_coverage=1 00:06:05.807 --rc genhtml_function_coverage=1 00:06:05.807 --rc genhtml_legend=1 00:06:05.807 --rc geninfo_all_blocks=1 00:06:05.807 --rc geninfo_unexecuted_blocks=1 00:06:05.807 00:06:05.807 ' 00:06:05.807 17:27:36 rpc_client -- common/autotest_common.sh@1695 -- # export 'LCOV=lcov 00:06:05.807 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:05.807 --rc genhtml_branch_coverage=1 00:06:05.807 --rc genhtml_function_coverage=1 00:06:05.807 --rc genhtml_legend=1 00:06:05.807 --rc geninfo_all_blocks=1 00:06:05.807 --rc geninfo_unexecuted_blocks=1 00:06:05.807 00:06:05.807 ' 00:06:05.807 17:27:36 rpc_client -- common/autotest_common.sh@1695 -- # LCOV='lcov 00:06:05.807 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:05.807 --rc genhtml_branch_coverage=1 00:06:05.807 --rc genhtml_function_coverage=1 00:06:05.807 --rc genhtml_legend=1 00:06:05.807 --rc geninfo_all_blocks=1 00:06:05.807 --rc geninfo_unexecuted_blocks=1 00:06:05.807 00:06:05.807 ' 00:06:05.807 17:27:36 rpc_client -- rpc_client/rpc_client.sh@10 -- # /home/vagrant/spdk_repo/spdk/test/rpc_client/rpc_client_test 00:06:05.807 OK 00:06:06.068 17:27:37 rpc_client -- rpc_client/rpc_client.sh@12 -- # trap - SIGINT SIGTERM EXIT 00:06:06.068 ************************************ 00:06:06.068 END TEST rpc_client 00:06:06.068 ************************************ 00:06:06.068 00:06:06.068 real 0m0.289s 00:06:06.068 user 0m0.158s 00:06:06.068 sys 0m0.146s 00:06:06.068 17:27:37 rpc_client -- common/autotest_common.sh@1126 -- # xtrace_disable 00:06:06.068 17:27:37 rpc_client -- common/autotest_common.sh@10 -- # set +x 00:06:06.068 17:27:37 -- spdk/autotest.sh@159 -- # run_test json_config /home/vagrant/spdk_repo/spdk/test/json_config/json_config.sh 00:06:06.068 17:27:37 -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:06:06.068 17:27:37 -- common/autotest_common.sh@1107 -- # xtrace_disable 00:06:06.068 17:27:37 -- common/autotest_common.sh@10 -- # set +x 00:06:06.068 ************************************ 00:06:06.068 START TEST json_config 00:06:06.068 ************************************ 00:06:06.068 17:27:37 json_config -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/json_config/json_config.sh 00:06:06.068 17:27:37 json_config -- common/autotest_common.sh@1680 -- # [[ y == y ]] 00:06:06.068 17:27:37 json_config -- common/autotest_common.sh@1681 -- # lcov --version 00:06:06.068 17:27:37 json_config -- common/autotest_common.sh@1681 -- # awk '{print $NF}' 00:06:06.068 17:27:37 json_config -- common/autotest_common.sh@1681 -- # lt 1.15 2 00:06:06.068 17:27:37 json_config -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:06:06.068 17:27:37 json_config -- scripts/common.sh@333 -- # local ver1 ver1_l 00:06:06.068 17:27:37 json_config -- scripts/common.sh@334 -- # local ver2 ver2_l 00:06:06.068 17:27:37 json_config -- scripts/common.sh@336 -- # IFS=.-: 00:06:06.068 17:27:37 json_config -- scripts/common.sh@336 -- # read -ra ver1 00:06:06.068 17:27:37 json_config -- scripts/common.sh@337 -- # IFS=.-: 00:06:06.068 17:27:37 json_config -- scripts/common.sh@337 -- # read -ra ver2 00:06:06.069 17:27:37 json_config -- scripts/common.sh@338 -- # local 'op=<' 00:06:06.069 17:27:37 json_config -- scripts/common.sh@340 -- # ver1_l=2 00:06:06.069 17:27:37 json_config -- scripts/common.sh@341 -- # ver2_l=1 00:06:06.069 17:27:37 json_config -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:06:06.069 17:27:37 json_config -- scripts/common.sh@344 -- # case "$op" in 00:06:06.069 17:27:37 json_config -- scripts/common.sh@345 -- # : 1 00:06:06.069 17:27:37 json_config -- scripts/common.sh@364 -- # (( v = 0 )) 00:06:06.069 17:27:37 json_config -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:06:06.330 17:27:37 json_config -- scripts/common.sh@365 -- # decimal 1 00:06:06.330 17:27:37 json_config -- scripts/common.sh@353 -- # local d=1 00:06:06.330 17:27:37 json_config -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:06:06.330 17:27:37 json_config -- scripts/common.sh@355 -- # echo 1 00:06:06.330 17:27:37 json_config -- scripts/common.sh@365 -- # ver1[v]=1 00:06:06.330 17:27:37 json_config -- scripts/common.sh@366 -- # decimal 2 00:06:06.330 17:27:37 json_config -- scripts/common.sh@353 -- # local d=2 00:06:06.330 17:27:37 json_config -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:06:06.330 17:27:37 json_config -- scripts/common.sh@355 -- # echo 2 00:06:06.330 17:27:37 json_config -- scripts/common.sh@366 -- # ver2[v]=2 00:06:06.330 17:27:37 json_config -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:06:06.330 17:27:37 json_config -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:06:06.330 17:27:37 json_config -- scripts/common.sh@368 -- # return 0 00:06:06.330 17:27:37 json_config -- common/autotest_common.sh@1682 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:06:06.330 17:27:37 json_config -- common/autotest_common.sh@1694 -- # export 'LCOV_OPTS= 00:06:06.330 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:06.330 --rc genhtml_branch_coverage=1 00:06:06.330 --rc genhtml_function_coverage=1 00:06:06.330 --rc genhtml_legend=1 00:06:06.330 --rc geninfo_all_blocks=1 00:06:06.330 --rc geninfo_unexecuted_blocks=1 00:06:06.330 00:06:06.330 ' 00:06:06.330 17:27:37 json_config -- common/autotest_common.sh@1694 -- # LCOV_OPTS=' 00:06:06.330 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:06.330 --rc genhtml_branch_coverage=1 00:06:06.330 --rc genhtml_function_coverage=1 00:06:06.330 --rc genhtml_legend=1 00:06:06.330 --rc geninfo_all_blocks=1 00:06:06.330 --rc geninfo_unexecuted_blocks=1 00:06:06.330 00:06:06.330 ' 00:06:06.330 17:27:37 json_config -- common/autotest_common.sh@1695 -- # export 'LCOV=lcov 00:06:06.330 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:06.330 --rc genhtml_branch_coverage=1 00:06:06.330 --rc genhtml_function_coverage=1 00:06:06.330 --rc genhtml_legend=1 00:06:06.330 --rc geninfo_all_blocks=1 00:06:06.330 --rc geninfo_unexecuted_blocks=1 00:06:06.330 00:06:06.330 ' 00:06:06.330 17:27:37 json_config -- common/autotest_common.sh@1695 -- # LCOV='lcov 00:06:06.330 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:06.330 --rc genhtml_branch_coverage=1 00:06:06.330 --rc genhtml_function_coverage=1 00:06:06.330 --rc genhtml_legend=1 00:06:06.330 --rc geninfo_all_blocks=1 00:06:06.330 --rc geninfo_unexecuted_blocks=1 00:06:06.330 00:06:06.330 ' 00:06:06.330 17:27:37 json_config -- json_config/json_config.sh@8 -- # source /home/vagrant/spdk_repo/spdk/test/nvmf/common.sh 00:06:06.330 17:27:37 json_config -- nvmf/common.sh@7 -- # uname -s 00:06:06.330 17:27:37 json_config -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:06:06.330 17:27:37 json_config -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:06:06.330 17:27:37 json_config -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:06:06.331 17:27:37 json_config -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:06:06.331 17:27:37 json_config -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:06:06.331 17:27:37 json_config -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:06:06.331 17:27:37 json_config -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:06:06.331 17:27:37 json_config -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:06:06.331 17:27:37 json_config -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:06:06.331 17:27:37 json_config -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:06:06.331 17:27:37 json_config -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:8044f8d2-4aeb-4cab-84e2-f73e0a4751e5 00:06:06.331 17:27:37 json_config -- nvmf/common.sh@18 -- # NVME_HOSTID=8044f8d2-4aeb-4cab-84e2-f73e0a4751e5 00:06:06.331 17:27:37 json_config -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:06:06.331 17:27:37 json_config -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:06:06.331 17:27:37 json_config -- nvmf/common.sh@21 -- # NET_TYPE=phy-fallback 00:06:06.331 17:27:37 json_config -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:06:06.331 17:27:37 json_config -- nvmf/common.sh@49 -- # source /home/vagrant/spdk_repo/spdk/scripts/common.sh 00:06:06.331 17:27:37 json_config -- scripts/common.sh@15 -- # shopt -s extglob 00:06:06.331 17:27:37 json_config -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:06:06.331 17:27:37 json_config -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:06:06.331 17:27:37 json_config -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:06:06.331 17:27:37 json_config -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:06:06.331 17:27:37 json_config -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:06:06.331 17:27:37 json_config -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:06:06.331 17:27:37 json_config -- paths/export.sh@5 -- # export PATH 00:06:06.331 17:27:37 json_config -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:06:06.331 17:27:37 json_config -- nvmf/common.sh@51 -- # : 0 00:06:06.331 17:27:37 json_config -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:06:06.331 17:27:37 json_config -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:06:06.331 17:27:37 json_config -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:06:06.331 17:27:37 json_config -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:06:06.331 17:27:37 json_config -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:06:06.331 17:27:37 json_config -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:06:06.331 /home/vagrant/spdk_repo/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:06:06.331 17:27:37 json_config -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:06:06.331 17:27:37 json_config -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:06:06.331 17:27:37 json_config -- nvmf/common.sh@55 -- # have_pci_nics=0 00:06:06.331 17:27:37 json_config -- json_config/json_config.sh@9 -- # source /home/vagrant/spdk_repo/spdk/test/json_config/common.sh 00:06:06.331 WARNING: No tests are enabled so not running JSON configuration tests 00:06:06.331 17:27:37 json_config -- json_config/json_config.sh@11 -- # [[ 0 -eq 1 ]] 00:06:06.331 17:27:37 json_config -- json_config/json_config.sh@15 -- # [[ 0 -ne 1 ]] 00:06:06.331 17:27:37 json_config -- json_config/json_config.sh@15 -- # [[ 0 -eq 1 ]] 00:06:06.331 17:27:37 json_config -- json_config/json_config.sh@26 -- # (( SPDK_TEST_BLOCKDEV + SPDK_TEST_ISCSI + SPDK_TEST_NVMF + SPDK_TEST_VHOST + SPDK_TEST_VHOST_INIT + SPDK_TEST_RBD == 0 )) 00:06:06.331 17:27:37 json_config -- json_config/json_config.sh@27 -- # echo 'WARNING: No tests are enabled so not running JSON configuration tests' 00:06:06.331 17:27:37 json_config -- json_config/json_config.sh@28 -- # exit 0 00:06:06.331 ************************************ 00:06:06.331 END TEST json_config 00:06:06.331 ************************************ 00:06:06.331 00:06:06.331 real 0m0.241s 00:06:06.331 user 0m0.150s 00:06:06.331 sys 0m0.095s 00:06:06.331 17:27:37 json_config -- common/autotest_common.sh@1126 -- # xtrace_disable 00:06:06.331 17:27:37 json_config -- common/autotest_common.sh@10 -- # set +x 00:06:06.331 17:27:37 -- spdk/autotest.sh@160 -- # run_test json_config_extra_key /home/vagrant/spdk_repo/spdk/test/json_config/json_config_extra_key.sh 00:06:06.331 17:27:37 -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:06:06.331 17:27:37 -- common/autotest_common.sh@1107 -- # xtrace_disable 00:06:06.331 17:27:37 -- common/autotest_common.sh@10 -- # set +x 00:06:06.331 ************************************ 00:06:06.331 START TEST json_config_extra_key 00:06:06.331 ************************************ 00:06:06.331 17:27:37 json_config_extra_key -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/json_config/json_config_extra_key.sh 00:06:06.331 17:27:37 json_config_extra_key -- common/autotest_common.sh@1680 -- # [[ y == y ]] 00:06:06.331 17:27:37 json_config_extra_key -- common/autotest_common.sh@1681 -- # lcov --version 00:06:06.331 17:27:37 json_config_extra_key -- common/autotest_common.sh@1681 -- # awk '{print $NF}' 00:06:06.593 17:27:37 json_config_extra_key -- common/autotest_common.sh@1681 -- # lt 1.15 2 00:06:06.593 17:27:37 json_config_extra_key -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:06:06.593 17:27:37 json_config_extra_key -- scripts/common.sh@333 -- # local ver1 ver1_l 00:06:06.593 17:27:37 json_config_extra_key -- scripts/common.sh@334 -- # local ver2 ver2_l 00:06:06.593 17:27:37 json_config_extra_key -- scripts/common.sh@336 -- # IFS=.-: 00:06:06.593 17:27:37 json_config_extra_key -- scripts/common.sh@336 -- # read -ra ver1 00:06:06.593 17:27:37 json_config_extra_key -- scripts/common.sh@337 -- # IFS=.-: 00:06:06.593 17:27:37 json_config_extra_key -- scripts/common.sh@337 -- # read -ra ver2 00:06:06.593 17:27:37 json_config_extra_key -- scripts/common.sh@338 -- # local 'op=<' 00:06:06.593 17:27:37 json_config_extra_key -- scripts/common.sh@340 -- # ver1_l=2 00:06:06.593 17:27:37 json_config_extra_key -- scripts/common.sh@341 -- # ver2_l=1 00:06:06.593 17:27:37 json_config_extra_key -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:06:06.593 17:27:37 json_config_extra_key -- scripts/common.sh@344 -- # case "$op" in 00:06:06.593 17:27:37 json_config_extra_key -- scripts/common.sh@345 -- # : 1 00:06:06.593 17:27:37 json_config_extra_key -- scripts/common.sh@364 -- # (( v = 0 )) 00:06:06.593 17:27:37 json_config_extra_key -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:06:06.593 17:27:37 json_config_extra_key -- scripts/common.sh@365 -- # decimal 1 00:06:06.593 17:27:37 json_config_extra_key -- scripts/common.sh@353 -- # local d=1 00:06:06.593 17:27:37 json_config_extra_key -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:06:06.593 17:27:37 json_config_extra_key -- scripts/common.sh@355 -- # echo 1 00:06:06.593 17:27:37 json_config_extra_key -- scripts/common.sh@365 -- # ver1[v]=1 00:06:06.593 17:27:37 json_config_extra_key -- scripts/common.sh@366 -- # decimal 2 00:06:06.593 17:27:37 json_config_extra_key -- scripts/common.sh@353 -- # local d=2 00:06:06.593 17:27:37 json_config_extra_key -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:06:06.593 17:27:37 json_config_extra_key -- scripts/common.sh@355 -- # echo 2 00:06:06.593 17:27:37 json_config_extra_key -- scripts/common.sh@366 -- # ver2[v]=2 00:06:06.593 17:27:37 json_config_extra_key -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:06:06.593 17:27:37 json_config_extra_key -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:06:06.593 17:27:37 json_config_extra_key -- scripts/common.sh@368 -- # return 0 00:06:06.593 17:27:37 json_config_extra_key -- common/autotest_common.sh@1682 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:06:06.593 17:27:37 json_config_extra_key -- common/autotest_common.sh@1694 -- # export 'LCOV_OPTS= 00:06:06.593 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:06.593 --rc genhtml_branch_coverage=1 00:06:06.593 --rc genhtml_function_coverage=1 00:06:06.593 --rc genhtml_legend=1 00:06:06.593 --rc geninfo_all_blocks=1 00:06:06.593 --rc geninfo_unexecuted_blocks=1 00:06:06.593 00:06:06.593 ' 00:06:06.593 17:27:37 json_config_extra_key -- common/autotest_common.sh@1694 -- # LCOV_OPTS=' 00:06:06.593 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:06.593 --rc genhtml_branch_coverage=1 00:06:06.593 --rc genhtml_function_coverage=1 00:06:06.593 --rc genhtml_legend=1 00:06:06.593 --rc geninfo_all_blocks=1 00:06:06.593 --rc geninfo_unexecuted_blocks=1 00:06:06.593 00:06:06.593 ' 00:06:06.593 17:27:37 json_config_extra_key -- common/autotest_common.sh@1695 -- # export 'LCOV=lcov 00:06:06.593 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:06.593 --rc genhtml_branch_coverage=1 00:06:06.593 --rc genhtml_function_coverage=1 00:06:06.593 --rc genhtml_legend=1 00:06:06.593 --rc geninfo_all_blocks=1 00:06:06.593 --rc geninfo_unexecuted_blocks=1 00:06:06.593 00:06:06.593 ' 00:06:06.593 17:27:37 json_config_extra_key -- common/autotest_common.sh@1695 -- # LCOV='lcov 00:06:06.593 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:06.593 --rc genhtml_branch_coverage=1 00:06:06.593 --rc genhtml_function_coverage=1 00:06:06.593 --rc genhtml_legend=1 00:06:06.593 --rc geninfo_all_blocks=1 00:06:06.593 --rc geninfo_unexecuted_blocks=1 00:06:06.593 00:06:06.593 ' 00:06:06.593 17:27:37 json_config_extra_key -- json_config/json_config_extra_key.sh@9 -- # source /home/vagrant/spdk_repo/spdk/test/nvmf/common.sh 00:06:06.593 17:27:37 json_config_extra_key -- nvmf/common.sh@7 -- # uname -s 00:06:06.593 17:27:37 json_config_extra_key -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:06:06.593 17:27:37 json_config_extra_key -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:06:06.593 17:27:37 json_config_extra_key -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:06:06.593 17:27:37 json_config_extra_key -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:06:06.593 17:27:37 json_config_extra_key -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:06:06.594 17:27:37 json_config_extra_key -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:06:06.594 17:27:37 json_config_extra_key -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:06:06.594 17:27:37 json_config_extra_key -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:06:06.594 17:27:37 json_config_extra_key -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:06:06.594 17:27:37 json_config_extra_key -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:06:06.594 17:27:37 json_config_extra_key -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:8044f8d2-4aeb-4cab-84e2-f73e0a4751e5 00:06:06.594 17:27:37 json_config_extra_key -- nvmf/common.sh@18 -- # NVME_HOSTID=8044f8d2-4aeb-4cab-84e2-f73e0a4751e5 00:06:06.594 17:27:37 json_config_extra_key -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:06:06.594 17:27:37 json_config_extra_key -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:06:06.594 17:27:37 json_config_extra_key -- nvmf/common.sh@21 -- # NET_TYPE=phy-fallback 00:06:06.594 17:27:37 json_config_extra_key -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:06:06.594 17:27:37 json_config_extra_key -- nvmf/common.sh@49 -- # source /home/vagrant/spdk_repo/spdk/scripts/common.sh 00:06:06.594 17:27:37 json_config_extra_key -- scripts/common.sh@15 -- # shopt -s extglob 00:06:06.594 17:27:37 json_config_extra_key -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:06:06.594 17:27:37 json_config_extra_key -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:06:06.594 17:27:37 json_config_extra_key -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:06:06.594 17:27:37 json_config_extra_key -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:06:06.594 17:27:37 json_config_extra_key -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:06:06.594 17:27:37 json_config_extra_key -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:06:06.594 17:27:37 json_config_extra_key -- paths/export.sh@5 -- # export PATH 00:06:06.594 17:27:37 json_config_extra_key -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:06:06.594 17:27:37 json_config_extra_key -- nvmf/common.sh@51 -- # : 0 00:06:06.594 17:27:37 json_config_extra_key -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:06:06.594 17:27:37 json_config_extra_key -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:06:06.594 17:27:37 json_config_extra_key -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:06:06.594 17:27:37 json_config_extra_key -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:06:06.594 17:27:37 json_config_extra_key -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:06:06.594 17:27:37 json_config_extra_key -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:06:06.594 /home/vagrant/spdk_repo/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:06:06.594 17:27:37 json_config_extra_key -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:06:06.594 17:27:37 json_config_extra_key -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:06:06.594 17:27:37 json_config_extra_key -- nvmf/common.sh@55 -- # have_pci_nics=0 00:06:06.594 17:27:37 json_config_extra_key -- json_config/json_config_extra_key.sh@10 -- # source /home/vagrant/spdk_repo/spdk/test/json_config/common.sh 00:06:06.594 17:27:37 json_config_extra_key -- json_config/json_config_extra_key.sh@17 -- # app_pid=(['target']='') 00:06:06.594 17:27:37 json_config_extra_key -- json_config/json_config_extra_key.sh@17 -- # declare -A app_pid 00:06:06.594 17:27:37 json_config_extra_key -- json_config/json_config_extra_key.sh@18 -- # app_socket=(['target']='/var/tmp/spdk_tgt.sock') 00:06:06.594 17:27:37 json_config_extra_key -- json_config/json_config_extra_key.sh@18 -- # declare -A app_socket 00:06:06.594 17:27:37 json_config_extra_key -- json_config/json_config_extra_key.sh@19 -- # app_params=(['target']='-m 0x1 -s 1024') 00:06:06.594 17:27:37 json_config_extra_key -- json_config/json_config_extra_key.sh@19 -- # declare -A app_params 00:06:06.594 17:27:37 json_config_extra_key -- json_config/json_config_extra_key.sh@20 -- # configs_path=(['target']='/home/vagrant/spdk_repo/spdk/test/json_config/extra_key.json') 00:06:06.594 17:27:37 json_config_extra_key -- json_config/json_config_extra_key.sh@20 -- # declare -A configs_path 00:06:06.594 17:27:37 json_config_extra_key -- json_config/json_config_extra_key.sh@22 -- # trap 'on_error_exit "${FUNCNAME}" "${LINENO}"' ERR 00:06:06.594 17:27:37 json_config_extra_key -- json_config/json_config_extra_key.sh@24 -- # echo 'INFO: launching applications...' 00:06:06.594 INFO: launching applications... 00:06:06.594 17:27:37 json_config_extra_key -- json_config/json_config_extra_key.sh@25 -- # json_config_test_start_app target --json /home/vagrant/spdk_repo/spdk/test/json_config/extra_key.json 00:06:06.594 17:27:37 json_config_extra_key -- json_config/common.sh@9 -- # local app=target 00:06:06.594 17:27:37 json_config_extra_key -- json_config/common.sh@10 -- # shift 00:06:06.594 17:27:37 json_config_extra_key -- json_config/common.sh@12 -- # [[ -n 22 ]] 00:06:06.594 17:27:37 json_config_extra_key -- json_config/common.sh@13 -- # [[ -z '' ]] 00:06:06.594 17:27:37 json_config_extra_key -- json_config/common.sh@15 -- # local app_extra_params= 00:06:06.594 17:27:37 json_config_extra_key -- json_config/common.sh@16 -- # [[ 0 -eq 1 ]] 00:06:06.594 17:27:37 json_config_extra_key -- json_config/common.sh@16 -- # [[ 0 -eq 1 ]] 00:06:06.594 17:27:37 json_config_extra_key -- json_config/common.sh@22 -- # app_pid["$app"]=69458 00:06:06.594 17:27:37 json_config_extra_key -- json_config/common.sh@24 -- # echo 'Waiting for target to run...' 00:06:06.594 Waiting for target to run... 00:06:06.594 17:27:37 json_config_extra_key -- json_config/common.sh@25 -- # waitforlisten 69458 /var/tmp/spdk_tgt.sock 00:06:06.594 17:27:37 json_config_extra_key -- json_config/common.sh@21 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 -s 1024 -r /var/tmp/spdk_tgt.sock --json /home/vagrant/spdk_repo/spdk/test/json_config/extra_key.json 00:06:06.594 17:27:37 json_config_extra_key -- common/autotest_common.sh@831 -- # '[' -z 69458 ']' 00:06:06.594 17:27:37 json_config_extra_key -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk_tgt.sock 00:06:06.594 17:27:37 json_config_extra_key -- common/autotest_common.sh@836 -- # local max_retries=100 00:06:06.594 17:27:37 json_config_extra_key -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk_tgt.sock...' 00:06:06.594 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk_tgt.sock... 00:06:06.594 17:27:37 json_config_extra_key -- common/autotest_common.sh@840 -- # xtrace_disable 00:06:06.594 17:27:37 json_config_extra_key -- common/autotest_common.sh@10 -- # set +x 00:06:06.594 [2024-11-27 17:27:37.706229] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:06:06.594 [2024-11-27 17:27:37.706459] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 -m 1024 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid69458 ] 00:06:07.164 [2024-11-27 17:27:38.067846] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:07.164 [2024-11-27 17:27:38.108316] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:06:07.425 17:27:38 json_config_extra_key -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:06:07.425 17:27:38 json_config_extra_key -- common/autotest_common.sh@864 -- # return 0 00:06:07.425 00:06:07.425 INFO: shutting down applications... 00:06:07.425 17:27:38 json_config_extra_key -- json_config/common.sh@26 -- # echo '' 00:06:07.425 17:27:38 json_config_extra_key -- json_config/json_config_extra_key.sh@27 -- # echo 'INFO: shutting down applications...' 00:06:07.425 17:27:38 json_config_extra_key -- json_config/json_config_extra_key.sh@28 -- # json_config_test_shutdown_app target 00:06:07.425 17:27:38 json_config_extra_key -- json_config/common.sh@31 -- # local app=target 00:06:07.425 17:27:38 json_config_extra_key -- json_config/common.sh@34 -- # [[ -n 22 ]] 00:06:07.425 17:27:38 json_config_extra_key -- json_config/common.sh@35 -- # [[ -n 69458 ]] 00:06:07.425 17:27:38 json_config_extra_key -- json_config/common.sh@38 -- # kill -SIGINT 69458 00:06:07.425 17:27:38 json_config_extra_key -- json_config/common.sh@40 -- # (( i = 0 )) 00:06:07.425 17:27:38 json_config_extra_key -- json_config/common.sh@40 -- # (( i < 30 )) 00:06:07.425 17:27:38 json_config_extra_key -- json_config/common.sh@41 -- # kill -0 69458 00:06:07.425 17:27:38 json_config_extra_key -- json_config/common.sh@45 -- # sleep 0.5 00:06:08.000 17:27:39 json_config_extra_key -- json_config/common.sh@40 -- # (( i++ )) 00:06:08.000 17:27:39 json_config_extra_key -- json_config/common.sh@40 -- # (( i < 30 )) 00:06:08.000 17:27:39 json_config_extra_key -- json_config/common.sh@41 -- # kill -0 69458 00:06:08.000 17:27:39 json_config_extra_key -- json_config/common.sh@45 -- # sleep 0.5 00:06:08.568 17:27:39 json_config_extra_key -- json_config/common.sh@40 -- # (( i++ )) 00:06:08.568 17:27:39 json_config_extra_key -- json_config/common.sh@40 -- # (( i < 30 )) 00:06:08.568 17:27:39 json_config_extra_key -- json_config/common.sh@41 -- # kill -0 69458 00:06:08.568 17:27:39 json_config_extra_key -- json_config/common.sh@42 -- # app_pid["$app"]= 00:06:08.568 17:27:39 json_config_extra_key -- json_config/common.sh@43 -- # break 00:06:08.568 17:27:39 json_config_extra_key -- json_config/common.sh@48 -- # [[ -n '' ]] 00:06:08.568 17:27:39 json_config_extra_key -- json_config/common.sh@53 -- # echo 'SPDK target shutdown done' 00:06:08.568 SPDK target shutdown done 00:06:08.568 Success 00:06:08.568 17:27:39 json_config_extra_key -- json_config/json_config_extra_key.sh@30 -- # echo Success 00:06:08.568 00:06:08.568 real 0m2.154s 00:06:08.568 user 0m1.649s 00:06:08.568 sys 0m0.475s 00:06:08.568 17:27:39 json_config_extra_key -- common/autotest_common.sh@1126 -- # xtrace_disable 00:06:08.568 17:27:39 json_config_extra_key -- common/autotest_common.sh@10 -- # set +x 00:06:08.568 ************************************ 00:06:08.568 END TEST json_config_extra_key 00:06:08.568 ************************************ 00:06:08.568 17:27:39 -- spdk/autotest.sh@161 -- # run_test alias_rpc /home/vagrant/spdk_repo/spdk/test/json_config/alias_rpc/alias_rpc.sh 00:06:08.568 17:27:39 -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:06:08.568 17:27:39 -- common/autotest_common.sh@1107 -- # xtrace_disable 00:06:08.568 17:27:39 -- common/autotest_common.sh@10 -- # set +x 00:06:08.568 ************************************ 00:06:08.568 START TEST alias_rpc 00:06:08.568 ************************************ 00:06:08.568 17:27:39 alias_rpc -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/json_config/alias_rpc/alias_rpc.sh 00:06:08.568 * Looking for test storage... 00:06:08.568 * Found test storage at /home/vagrant/spdk_repo/spdk/test/json_config/alias_rpc 00:06:08.568 17:27:39 alias_rpc -- common/autotest_common.sh@1680 -- # [[ y == y ]] 00:06:08.568 17:27:39 alias_rpc -- common/autotest_common.sh@1681 -- # lcov --version 00:06:08.568 17:27:39 alias_rpc -- common/autotest_common.sh@1681 -- # awk '{print $NF}' 00:06:08.828 17:27:39 alias_rpc -- common/autotest_common.sh@1681 -- # lt 1.15 2 00:06:08.828 17:27:39 alias_rpc -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:06:08.828 17:27:39 alias_rpc -- scripts/common.sh@333 -- # local ver1 ver1_l 00:06:08.828 17:27:39 alias_rpc -- scripts/common.sh@334 -- # local ver2 ver2_l 00:06:08.828 17:27:39 alias_rpc -- scripts/common.sh@336 -- # IFS=.-: 00:06:08.828 17:27:39 alias_rpc -- scripts/common.sh@336 -- # read -ra ver1 00:06:08.828 17:27:39 alias_rpc -- scripts/common.sh@337 -- # IFS=.-: 00:06:08.828 17:27:39 alias_rpc -- scripts/common.sh@337 -- # read -ra ver2 00:06:08.828 17:27:39 alias_rpc -- scripts/common.sh@338 -- # local 'op=<' 00:06:08.828 17:27:39 alias_rpc -- scripts/common.sh@340 -- # ver1_l=2 00:06:08.828 17:27:39 alias_rpc -- scripts/common.sh@341 -- # ver2_l=1 00:06:08.829 17:27:39 alias_rpc -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:06:08.829 17:27:39 alias_rpc -- scripts/common.sh@344 -- # case "$op" in 00:06:08.829 17:27:39 alias_rpc -- scripts/common.sh@345 -- # : 1 00:06:08.829 17:27:39 alias_rpc -- scripts/common.sh@364 -- # (( v = 0 )) 00:06:08.829 17:27:39 alias_rpc -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:06:08.829 17:27:39 alias_rpc -- scripts/common.sh@365 -- # decimal 1 00:06:08.829 17:27:39 alias_rpc -- scripts/common.sh@353 -- # local d=1 00:06:08.829 17:27:39 alias_rpc -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:06:08.829 17:27:39 alias_rpc -- scripts/common.sh@355 -- # echo 1 00:06:08.829 17:27:39 alias_rpc -- scripts/common.sh@365 -- # ver1[v]=1 00:06:08.829 17:27:39 alias_rpc -- scripts/common.sh@366 -- # decimal 2 00:06:08.829 17:27:39 alias_rpc -- scripts/common.sh@353 -- # local d=2 00:06:08.829 17:27:39 alias_rpc -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:06:08.829 17:27:39 alias_rpc -- scripts/common.sh@355 -- # echo 2 00:06:08.829 17:27:39 alias_rpc -- scripts/common.sh@366 -- # ver2[v]=2 00:06:08.829 17:27:39 alias_rpc -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:06:08.829 17:27:39 alias_rpc -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:06:08.829 17:27:39 alias_rpc -- scripts/common.sh@368 -- # return 0 00:06:08.829 17:27:39 alias_rpc -- common/autotest_common.sh@1682 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:06:08.829 17:27:39 alias_rpc -- common/autotest_common.sh@1694 -- # export 'LCOV_OPTS= 00:06:08.829 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:08.829 --rc genhtml_branch_coverage=1 00:06:08.829 --rc genhtml_function_coverage=1 00:06:08.829 --rc genhtml_legend=1 00:06:08.829 --rc geninfo_all_blocks=1 00:06:08.829 --rc geninfo_unexecuted_blocks=1 00:06:08.829 00:06:08.829 ' 00:06:08.829 17:27:39 alias_rpc -- common/autotest_common.sh@1694 -- # LCOV_OPTS=' 00:06:08.829 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:08.829 --rc genhtml_branch_coverage=1 00:06:08.829 --rc genhtml_function_coverage=1 00:06:08.829 --rc genhtml_legend=1 00:06:08.829 --rc geninfo_all_blocks=1 00:06:08.829 --rc geninfo_unexecuted_blocks=1 00:06:08.829 00:06:08.829 ' 00:06:08.829 17:27:39 alias_rpc -- common/autotest_common.sh@1695 -- # export 'LCOV=lcov 00:06:08.829 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:08.829 --rc genhtml_branch_coverage=1 00:06:08.829 --rc genhtml_function_coverage=1 00:06:08.829 --rc genhtml_legend=1 00:06:08.829 --rc geninfo_all_blocks=1 00:06:08.829 --rc geninfo_unexecuted_blocks=1 00:06:08.829 00:06:08.829 ' 00:06:08.829 17:27:39 alias_rpc -- common/autotest_common.sh@1695 -- # LCOV='lcov 00:06:08.829 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:08.829 --rc genhtml_branch_coverage=1 00:06:08.829 --rc genhtml_function_coverage=1 00:06:08.829 --rc genhtml_legend=1 00:06:08.829 --rc geninfo_all_blocks=1 00:06:08.829 --rc geninfo_unexecuted_blocks=1 00:06:08.829 00:06:08.829 ' 00:06:08.829 17:27:39 alias_rpc -- alias_rpc/alias_rpc.sh@10 -- # trap 'killprocess $spdk_tgt_pid; exit 1' ERR 00:06:08.829 17:27:39 alias_rpc -- alias_rpc/alias_rpc.sh@13 -- # spdk_tgt_pid=69538 00:06:08.829 17:27:39 alias_rpc -- alias_rpc/alias_rpc.sh@12 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:06:08.829 17:27:39 alias_rpc -- alias_rpc/alias_rpc.sh@14 -- # waitforlisten 69538 00:06:08.829 17:27:39 alias_rpc -- common/autotest_common.sh@831 -- # '[' -z 69538 ']' 00:06:08.829 17:27:39 alias_rpc -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:08.829 17:27:39 alias_rpc -- common/autotest_common.sh@836 -- # local max_retries=100 00:06:08.829 17:27:39 alias_rpc -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:08.829 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:08.829 17:27:39 alias_rpc -- common/autotest_common.sh@840 -- # xtrace_disable 00:06:08.829 17:27:39 alias_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:08.829 [2024-11-27 17:27:39.944985] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:06:08.829 [2024-11-27 17:27:39.945213] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid69538 ] 00:06:09.089 [2024-11-27 17:27:40.090300] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:09.089 [2024-11-27 17:27:40.160444] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:06:09.659 17:27:40 alias_rpc -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:06:09.659 17:27:40 alias_rpc -- common/autotest_common.sh@864 -- # return 0 00:06:09.659 17:27:40 alias_rpc -- alias_rpc/alias_rpc.sh@17 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py load_config -i 00:06:09.919 17:27:40 alias_rpc -- alias_rpc/alias_rpc.sh@19 -- # killprocess 69538 00:06:09.919 17:27:40 alias_rpc -- common/autotest_common.sh@950 -- # '[' -z 69538 ']' 00:06:09.919 17:27:40 alias_rpc -- common/autotest_common.sh@954 -- # kill -0 69538 00:06:09.919 17:27:40 alias_rpc -- common/autotest_common.sh@955 -- # uname 00:06:09.919 17:27:40 alias_rpc -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:06:09.919 17:27:40 alias_rpc -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 69538 00:06:09.919 killing process with pid 69538 00:06:09.919 17:27:41 alias_rpc -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:06:09.919 17:27:41 alias_rpc -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:06:09.919 17:27:41 alias_rpc -- common/autotest_common.sh@968 -- # echo 'killing process with pid 69538' 00:06:09.919 17:27:41 alias_rpc -- common/autotest_common.sh@969 -- # kill 69538 00:06:09.919 17:27:41 alias_rpc -- common/autotest_common.sh@974 -- # wait 69538 00:06:10.860 ************************************ 00:06:10.860 END TEST alias_rpc 00:06:10.860 ************************************ 00:06:10.860 00:06:10.860 real 0m2.066s 00:06:10.860 user 0m1.921s 00:06:10.860 sys 0m0.676s 00:06:10.860 17:27:41 alias_rpc -- common/autotest_common.sh@1126 -- # xtrace_disable 00:06:10.860 17:27:41 alias_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:10.860 17:27:41 -- spdk/autotest.sh@163 -- # [[ 0 -eq 0 ]] 00:06:10.860 17:27:41 -- spdk/autotest.sh@164 -- # run_test spdkcli_tcp /home/vagrant/spdk_repo/spdk/test/spdkcli/tcp.sh 00:06:10.860 17:27:41 -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:06:10.860 17:27:41 -- common/autotest_common.sh@1107 -- # xtrace_disable 00:06:10.860 17:27:41 -- common/autotest_common.sh@10 -- # set +x 00:06:10.860 ************************************ 00:06:10.860 START TEST spdkcli_tcp 00:06:10.860 ************************************ 00:06:10.860 17:27:41 spdkcli_tcp -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/spdkcli/tcp.sh 00:06:10.860 * Looking for test storage... 00:06:10.860 * Found test storage at /home/vagrant/spdk_repo/spdk/test/spdkcli 00:06:10.860 17:27:41 spdkcli_tcp -- common/autotest_common.sh@1680 -- # [[ y == y ]] 00:06:10.860 17:27:41 spdkcli_tcp -- common/autotest_common.sh@1681 -- # lcov --version 00:06:10.860 17:27:41 spdkcli_tcp -- common/autotest_common.sh@1681 -- # awk '{print $NF}' 00:06:10.860 17:27:41 spdkcli_tcp -- common/autotest_common.sh@1681 -- # lt 1.15 2 00:06:10.860 17:27:41 spdkcli_tcp -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:06:10.860 17:27:41 spdkcli_tcp -- scripts/common.sh@333 -- # local ver1 ver1_l 00:06:10.860 17:27:41 spdkcli_tcp -- scripts/common.sh@334 -- # local ver2 ver2_l 00:06:10.860 17:27:41 spdkcli_tcp -- scripts/common.sh@336 -- # IFS=.-: 00:06:10.860 17:27:41 spdkcli_tcp -- scripts/common.sh@336 -- # read -ra ver1 00:06:10.860 17:27:41 spdkcli_tcp -- scripts/common.sh@337 -- # IFS=.-: 00:06:10.860 17:27:41 spdkcli_tcp -- scripts/common.sh@337 -- # read -ra ver2 00:06:10.860 17:27:41 spdkcli_tcp -- scripts/common.sh@338 -- # local 'op=<' 00:06:10.860 17:27:41 spdkcli_tcp -- scripts/common.sh@340 -- # ver1_l=2 00:06:10.860 17:27:41 spdkcli_tcp -- scripts/common.sh@341 -- # ver2_l=1 00:06:10.860 17:27:41 spdkcli_tcp -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:06:10.860 17:27:41 spdkcli_tcp -- scripts/common.sh@344 -- # case "$op" in 00:06:10.860 17:27:41 spdkcli_tcp -- scripts/common.sh@345 -- # : 1 00:06:10.861 17:27:41 spdkcli_tcp -- scripts/common.sh@364 -- # (( v = 0 )) 00:06:10.861 17:27:41 spdkcli_tcp -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:06:10.861 17:27:41 spdkcli_tcp -- scripts/common.sh@365 -- # decimal 1 00:06:10.861 17:27:41 spdkcli_tcp -- scripts/common.sh@353 -- # local d=1 00:06:10.861 17:27:41 spdkcli_tcp -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:06:10.861 17:27:41 spdkcli_tcp -- scripts/common.sh@355 -- # echo 1 00:06:10.861 17:27:41 spdkcli_tcp -- scripts/common.sh@365 -- # ver1[v]=1 00:06:10.861 17:27:41 spdkcli_tcp -- scripts/common.sh@366 -- # decimal 2 00:06:10.861 17:27:41 spdkcli_tcp -- scripts/common.sh@353 -- # local d=2 00:06:10.861 17:27:41 spdkcli_tcp -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:06:10.861 17:27:41 spdkcli_tcp -- scripts/common.sh@355 -- # echo 2 00:06:10.861 17:27:41 spdkcli_tcp -- scripts/common.sh@366 -- # ver2[v]=2 00:06:10.861 17:27:41 spdkcli_tcp -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:06:10.861 17:27:41 spdkcli_tcp -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:06:10.861 17:27:41 spdkcli_tcp -- scripts/common.sh@368 -- # return 0 00:06:10.861 17:27:41 spdkcli_tcp -- common/autotest_common.sh@1682 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:06:10.861 17:27:41 spdkcli_tcp -- common/autotest_common.sh@1694 -- # export 'LCOV_OPTS= 00:06:10.861 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:10.861 --rc genhtml_branch_coverage=1 00:06:10.861 --rc genhtml_function_coverage=1 00:06:10.861 --rc genhtml_legend=1 00:06:10.861 --rc geninfo_all_blocks=1 00:06:10.861 --rc geninfo_unexecuted_blocks=1 00:06:10.861 00:06:10.861 ' 00:06:10.861 17:27:41 spdkcli_tcp -- common/autotest_common.sh@1694 -- # LCOV_OPTS=' 00:06:10.861 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:10.861 --rc genhtml_branch_coverage=1 00:06:10.861 --rc genhtml_function_coverage=1 00:06:10.861 --rc genhtml_legend=1 00:06:10.861 --rc geninfo_all_blocks=1 00:06:10.861 --rc geninfo_unexecuted_blocks=1 00:06:10.861 00:06:10.861 ' 00:06:10.861 17:27:41 spdkcli_tcp -- common/autotest_common.sh@1695 -- # export 'LCOV=lcov 00:06:10.861 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:10.861 --rc genhtml_branch_coverage=1 00:06:10.861 --rc genhtml_function_coverage=1 00:06:10.861 --rc genhtml_legend=1 00:06:10.861 --rc geninfo_all_blocks=1 00:06:10.861 --rc geninfo_unexecuted_blocks=1 00:06:10.861 00:06:10.861 ' 00:06:10.861 17:27:41 spdkcli_tcp -- common/autotest_common.sh@1695 -- # LCOV='lcov 00:06:10.861 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:10.861 --rc genhtml_branch_coverage=1 00:06:10.861 --rc genhtml_function_coverage=1 00:06:10.861 --rc genhtml_legend=1 00:06:10.861 --rc geninfo_all_blocks=1 00:06:10.861 --rc geninfo_unexecuted_blocks=1 00:06:10.861 00:06:10.861 ' 00:06:10.861 17:27:41 spdkcli_tcp -- spdkcli/tcp.sh@9 -- # source /home/vagrant/spdk_repo/spdk/test/spdkcli/common.sh 00:06:10.861 17:27:41 spdkcli_tcp -- spdkcli/common.sh@6 -- # spdkcli_job=/home/vagrant/spdk_repo/spdk/test/spdkcli/spdkcli_job.py 00:06:10.861 17:27:41 spdkcli_tcp -- spdkcli/common.sh@7 -- # spdk_clear_config_py=/home/vagrant/spdk_repo/spdk/test/json_config/clear_config.py 00:06:10.861 17:27:41 spdkcli_tcp -- spdkcli/tcp.sh@18 -- # IP_ADDRESS=127.0.0.1 00:06:10.861 17:27:41 spdkcli_tcp -- spdkcli/tcp.sh@19 -- # PORT=9998 00:06:10.861 17:27:41 spdkcli_tcp -- spdkcli/tcp.sh@21 -- # trap 'err_cleanup; exit 1' SIGINT SIGTERM EXIT 00:06:10.861 17:27:41 spdkcli_tcp -- spdkcli/tcp.sh@23 -- # timing_enter run_spdk_tgt_tcp 00:06:10.861 17:27:41 spdkcli_tcp -- common/autotest_common.sh@724 -- # xtrace_disable 00:06:10.861 17:27:41 spdkcli_tcp -- common/autotest_common.sh@10 -- # set +x 00:06:10.861 17:27:41 spdkcli_tcp -- spdkcli/tcp.sh@25 -- # spdk_tgt_pid=69623 00:06:10.861 17:27:41 spdkcli_tcp -- spdkcli/tcp.sh@24 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x3 -p 0 00:06:10.861 17:27:41 spdkcli_tcp -- spdkcli/tcp.sh@27 -- # waitforlisten 69623 00:06:10.861 17:27:41 spdkcli_tcp -- common/autotest_common.sh@831 -- # '[' -z 69623 ']' 00:06:10.861 17:27:41 spdkcli_tcp -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:10.861 17:27:41 spdkcli_tcp -- common/autotest_common.sh@836 -- # local max_retries=100 00:06:10.861 17:27:41 spdkcli_tcp -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:10.861 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:10.861 17:27:41 spdkcli_tcp -- common/autotest_common.sh@840 -- # xtrace_disable 00:06:10.861 17:27:41 spdkcli_tcp -- common/autotest_common.sh@10 -- # set +x 00:06:11.121 [2024-11-27 17:27:42.091889] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:06:11.121 [2024-11-27 17:27:42.092098] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x3 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid69623 ] 00:06:11.121 [2024-11-27 17:27:42.238943] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 2 00:06:11.381 [2024-11-27 17:27:42.311618] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:06:11.381 [2024-11-27 17:27:42.311731] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 1 00:06:11.952 17:27:42 spdkcli_tcp -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:06:11.952 17:27:42 spdkcli_tcp -- common/autotest_common.sh@864 -- # return 0 00:06:11.952 17:27:42 spdkcli_tcp -- spdkcli/tcp.sh@31 -- # socat_pid=69640 00:06:11.952 17:27:42 spdkcli_tcp -- spdkcli/tcp.sh@30 -- # socat TCP-LISTEN:9998 UNIX-CONNECT:/var/tmp/spdk.sock 00:06:11.952 17:27:42 spdkcli_tcp -- spdkcli/tcp.sh@33 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -r 100 -t 2 -s 127.0.0.1 -p 9998 rpc_get_methods 00:06:11.952 [ 00:06:11.952 "bdev_malloc_delete", 00:06:11.952 "bdev_malloc_create", 00:06:11.952 "bdev_null_resize", 00:06:11.952 "bdev_null_delete", 00:06:11.952 "bdev_null_create", 00:06:11.952 "bdev_nvme_cuse_unregister", 00:06:11.952 "bdev_nvme_cuse_register", 00:06:11.952 "bdev_opal_new_user", 00:06:11.952 "bdev_opal_set_lock_state", 00:06:11.952 "bdev_opal_delete", 00:06:11.952 "bdev_opal_get_info", 00:06:11.952 "bdev_opal_create", 00:06:11.952 "bdev_nvme_opal_revert", 00:06:11.952 "bdev_nvme_opal_init", 00:06:11.952 "bdev_nvme_send_cmd", 00:06:11.952 "bdev_nvme_set_keys", 00:06:11.952 "bdev_nvme_get_path_iostat", 00:06:11.952 "bdev_nvme_get_mdns_discovery_info", 00:06:11.952 "bdev_nvme_stop_mdns_discovery", 00:06:11.952 "bdev_nvme_start_mdns_discovery", 00:06:11.952 "bdev_nvme_set_multipath_policy", 00:06:11.952 "bdev_nvme_set_preferred_path", 00:06:11.952 "bdev_nvme_get_io_paths", 00:06:11.952 "bdev_nvme_remove_error_injection", 00:06:11.952 "bdev_nvme_add_error_injection", 00:06:11.952 "bdev_nvme_get_discovery_info", 00:06:11.952 "bdev_nvme_stop_discovery", 00:06:11.952 "bdev_nvme_start_discovery", 00:06:11.952 "bdev_nvme_get_controller_health_info", 00:06:11.952 "bdev_nvme_disable_controller", 00:06:11.952 "bdev_nvme_enable_controller", 00:06:11.952 "bdev_nvme_reset_controller", 00:06:11.952 "bdev_nvme_get_transport_statistics", 00:06:11.952 "bdev_nvme_apply_firmware", 00:06:11.952 "bdev_nvme_detach_controller", 00:06:11.952 "bdev_nvme_get_controllers", 00:06:11.952 "bdev_nvme_attach_controller", 00:06:11.952 "bdev_nvme_set_hotplug", 00:06:11.952 "bdev_nvme_set_options", 00:06:11.952 "bdev_passthru_delete", 00:06:11.952 "bdev_passthru_create", 00:06:11.952 "bdev_lvol_set_parent_bdev", 00:06:11.952 "bdev_lvol_set_parent", 00:06:11.952 "bdev_lvol_check_shallow_copy", 00:06:11.952 "bdev_lvol_start_shallow_copy", 00:06:11.952 "bdev_lvol_grow_lvstore", 00:06:11.952 "bdev_lvol_get_lvols", 00:06:11.952 "bdev_lvol_get_lvstores", 00:06:11.952 "bdev_lvol_delete", 00:06:11.952 "bdev_lvol_set_read_only", 00:06:11.952 "bdev_lvol_resize", 00:06:11.952 "bdev_lvol_decouple_parent", 00:06:11.952 "bdev_lvol_inflate", 00:06:11.952 "bdev_lvol_rename", 00:06:11.952 "bdev_lvol_clone_bdev", 00:06:11.952 "bdev_lvol_clone", 00:06:11.952 "bdev_lvol_snapshot", 00:06:11.952 "bdev_lvol_create", 00:06:11.952 "bdev_lvol_delete_lvstore", 00:06:11.952 "bdev_lvol_rename_lvstore", 00:06:11.952 "bdev_lvol_create_lvstore", 00:06:11.952 "bdev_raid_set_options", 00:06:11.952 "bdev_raid_remove_base_bdev", 00:06:11.952 "bdev_raid_add_base_bdev", 00:06:11.952 "bdev_raid_delete", 00:06:11.952 "bdev_raid_create", 00:06:11.952 "bdev_raid_get_bdevs", 00:06:11.952 "bdev_error_inject_error", 00:06:11.952 "bdev_error_delete", 00:06:11.952 "bdev_error_create", 00:06:11.952 "bdev_split_delete", 00:06:11.952 "bdev_split_create", 00:06:11.952 "bdev_delay_delete", 00:06:11.952 "bdev_delay_create", 00:06:11.952 "bdev_delay_update_latency", 00:06:11.952 "bdev_zone_block_delete", 00:06:11.952 "bdev_zone_block_create", 00:06:11.952 "blobfs_create", 00:06:11.952 "blobfs_detect", 00:06:11.952 "blobfs_set_cache_size", 00:06:11.952 "bdev_aio_delete", 00:06:11.952 "bdev_aio_rescan", 00:06:11.952 "bdev_aio_create", 00:06:11.952 "bdev_ftl_set_property", 00:06:11.952 "bdev_ftl_get_properties", 00:06:11.952 "bdev_ftl_get_stats", 00:06:11.952 "bdev_ftl_unmap", 00:06:11.952 "bdev_ftl_unload", 00:06:11.952 "bdev_ftl_delete", 00:06:11.952 "bdev_ftl_load", 00:06:11.952 "bdev_ftl_create", 00:06:11.952 "bdev_virtio_attach_controller", 00:06:11.952 "bdev_virtio_scsi_get_devices", 00:06:11.952 "bdev_virtio_detach_controller", 00:06:11.952 "bdev_virtio_blk_set_hotplug", 00:06:11.952 "bdev_iscsi_delete", 00:06:11.952 "bdev_iscsi_create", 00:06:11.952 "bdev_iscsi_set_options", 00:06:11.952 "accel_error_inject_error", 00:06:11.952 "ioat_scan_accel_module", 00:06:11.952 "dsa_scan_accel_module", 00:06:11.952 "iaa_scan_accel_module", 00:06:11.952 "keyring_file_remove_key", 00:06:11.952 "keyring_file_add_key", 00:06:11.952 "keyring_linux_set_options", 00:06:11.952 "fsdev_aio_delete", 00:06:11.952 "fsdev_aio_create", 00:06:11.952 "iscsi_get_histogram", 00:06:11.952 "iscsi_enable_histogram", 00:06:11.952 "iscsi_set_options", 00:06:11.952 "iscsi_get_auth_groups", 00:06:11.952 "iscsi_auth_group_remove_secret", 00:06:11.952 "iscsi_auth_group_add_secret", 00:06:11.952 "iscsi_delete_auth_group", 00:06:11.952 "iscsi_create_auth_group", 00:06:11.952 "iscsi_set_discovery_auth", 00:06:11.952 "iscsi_get_options", 00:06:11.952 "iscsi_target_node_request_logout", 00:06:11.952 "iscsi_target_node_set_redirect", 00:06:11.952 "iscsi_target_node_set_auth", 00:06:11.952 "iscsi_target_node_add_lun", 00:06:11.952 "iscsi_get_stats", 00:06:11.952 "iscsi_get_connections", 00:06:11.952 "iscsi_portal_group_set_auth", 00:06:11.952 "iscsi_start_portal_group", 00:06:11.952 "iscsi_delete_portal_group", 00:06:11.952 "iscsi_create_portal_group", 00:06:11.952 "iscsi_get_portal_groups", 00:06:11.952 "iscsi_delete_target_node", 00:06:11.952 "iscsi_target_node_remove_pg_ig_maps", 00:06:11.952 "iscsi_target_node_add_pg_ig_maps", 00:06:11.952 "iscsi_create_target_node", 00:06:11.952 "iscsi_get_target_nodes", 00:06:11.952 "iscsi_delete_initiator_group", 00:06:11.952 "iscsi_initiator_group_remove_initiators", 00:06:11.952 "iscsi_initiator_group_add_initiators", 00:06:11.952 "iscsi_create_initiator_group", 00:06:11.952 "iscsi_get_initiator_groups", 00:06:11.952 "nvmf_set_crdt", 00:06:11.952 "nvmf_set_config", 00:06:11.952 "nvmf_set_max_subsystems", 00:06:11.952 "nvmf_stop_mdns_prr", 00:06:11.952 "nvmf_publish_mdns_prr", 00:06:11.952 "nvmf_subsystem_get_listeners", 00:06:11.952 "nvmf_subsystem_get_qpairs", 00:06:11.952 "nvmf_subsystem_get_controllers", 00:06:11.952 "nvmf_get_stats", 00:06:11.952 "nvmf_get_transports", 00:06:11.952 "nvmf_create_transport", 00:06:11.952 "nvmf_get_targets", 00:06:11.952 "nvmf_delete_target", 00:06:11.952 "nvmf_create_target", 00:06:11.952 "nvmf_subsystem_allow_any_host", 00:06:11.952 "nvmf_subsystem_set_keys", 00:06:11.952 "nvmf_subsystem_remove_host", 00:06:11.952 "nvmf_subsystem_add_host", 00:06:11.952 "nvmf_ns_remove_host", 00:06:11.952 "nvmf_ns_add_host", 00:06:11.952 "nvmf_subsystem_remove_ns", 00:06:11.952 "nvmf_subsystem_set_ns_ana_group", 00:06:11.952 "nvmf_subsystem_add_ns", 00:06:11.952 "nvmf_subsystem_listener_set_ana_state", 00:06:11.952 "nvmf_discovery_get_referrals", 00:06:11.952 "nvmf_discovery_remove_referral", 00:06:11.952 "nvmf_discovery_add_referral", 00:06:11.953 "nvmf_subsystem_remove_listener", 00:06:11.953 "nvmf_subsystem_add_listener", 00:06:11.953 "nvmf_delete_subsystem", 00:06:11.953 "nvmf_create_subsystem", 00:06:11.953 "nvmf_get_subsystems", 00:06:11.953 "env_dpdk_get_mem_stats", 00:06:11.953 "nbd_get_disks", 00:06:11.953 "nbd_stop_disk", 00:06:11.953 "nbd_start_disk", 00:06:11.953 "ublk_recover_disk", 00:06:11.953 "ublk_get_disks", 00:06:11.953 "ublk_stop_disk", 00:06:11.953 "ublk_start_disk", 00:06:11.953 "ublk_destroy_target", 00:06:11.953 "ublk_create_target", 00:06:11.953 "virtio_blk_create_transport", 00:06:11.953 "virtio_blk_get_transports", 00:06:11.953 "vhost_controller_set_coalescing", 00:06:11.953 "vhost_get_controllers", 00:06:11.953 "vhost_delete_controller", 00:06:11.953 "vhost_create_blk_controller", 00:06:11.953 "vhost_scsi_controller_remove_target", 00:06:11.953 "vhost_scsi_controller_add_target", 00:06:11.953 "vhost_start_scsi_controller", 00:06:11.953 "vhost_create_scsi_controller", 00:06:11.953 "thread_set_cpumask", 00:06:11.953 "scheduler_set_options", 00:06:11.953 "framework_get_governor", 00:06:11.953 "framework_get_scheduler", 00:06:11.953 "framework_set_scheduler", 00:06:11.953 "framework_get_reactors", 00:06:11.953 "thread_get_io_channels", 00:06:11.953 "thread_get_pollers", 00:06:11.953 "thread_get_stats", 00:06:11.953 "framework_monitor_context_switch", 00:06:11.953 "spdk_kill_instance", 00:06:11.953 "log_enable_timestamps", 00:06:11.953 "log_get_flags", 00:06:11.953 "log_clear_flag", 00:06:11.953 "log_set_flag", 00:06:11.953 "log_get_level", 00:06:11.953 "log_set_level", 00:06:11.953 "log_get_print_level", 00:06:11.953 "log_set_print_level", 00:06:11.953 "framework_enable_cpumask_locks", 00:06:11.953 "framework_disable_cpumask_locks", 00:06:11.953 "framework_wait_init", 00:06:11.953 "framework_start_init", 00:06:11.953 "scsi_get_devices", 00:06:11.953 "bdev_get_histogram", 00:06:11.953 "bdev_enable_histogram", 00:06:11.953 "bdev_set_qos_limit", 00:06:11.953 "bdev_set_qd_sampling_period", 00:06:11.953 "bdev_get_bdevs", 00:06:11.953 "bdev_reset_iostat", 00:06:11.953 "bdev_get_iostat", 00:06:11.953 "bdev_examine", 00:06:11.953 "bdev_wait_for_examine", 00:06:11.953 "bdev_set_options", 00:06:11.953 "accel_get_stats", 00:06:11.953 "accel_set_options", 00:06:11.953 "accel_set_driver", 00:06:11.953 "accel_crypto_key_destroy", 00:06:11.953 "accel_crypto_keys_get", 00:06:11.953 "accel_crypto_key_create", 00:06:11.953 "accel_assign_opc", 00:06:11.953 "accel_get_module_info", 00:06:11.953 "accel_get_opc_assignments", 00:06:11.953 "vmd_rescan", 00:06:11.953 "vmd_remove_device", 00:06:11.953 "vmd_enable", 00:06:11.953 "sock_get_default_impl", 00:06:11.953 "sock_set_default_impl", 00:06:11.953 "sock_impl_set_options", 00:06:11.953 "sock_impl_get_options", 00:06:11.953 "iobuf_get_stats", 00:06:11.953 "iobuf_set_options", 00:06:11.953 "keyring_get_keys", 00:06:11.953 "framework_get_pci_devices", 00:06:11.953 "framework_get_config", 00:06:11.953 "framework_get_subsystems", 00:06:11.953 "fsdev_set_opts", 00:06:11.953 "fsdev_get_opts", 00:06:11.953 "trace_get_info", 00:06:11.953 "trace_get_tpoint_group_mask", 00:06:11.953 "trace_disable_tpoint_group", 00:06:11.953 "trace_enable_tpoint_group", 00:06:11.953 "trace_clear_tpoint_mask", 00:06:11.953 "trace_set_tpoint_mask", 00:06:11.953 "notify_get_notifications", 00:06:11.953 "notify_get_types", 00:06:11.953 "spdk_get_version", 00:06:11.953 "rpc_get_methods" 00:06:11.953 ] 00:06:11.953 17:27:43 spdkcli_tcp -- spdkcli/tcp.sh@35 -- # timing_exit run_spdk_tgt_tcp 00:06:11.953 17:27:43 spdkcli_tcp -- common/autotest_common.sh@730 -- # xtrace_disable 00:06:11.953 17:27:43 spdkcli_tcp -- common/autotest_common.sh@10 -- # set +x 00:06:11.953 17:27:43 spdkcli_tcp -- spdkcli/tcp.sh@37 -- # trap - SIGINT SIGTERM EXIT 00:06:11.953 17:27:43 spdkcli_tcp -- spdkcli/tcp.sh@38 -- # killprocess 69623 00:06:11.953 17:27:43 spdkcli_tcp -- common/autotest_common.sh@950 -- # '[' -z 69623 ']' 00:06:11.953 17:27:43 spdkcli_tcp -- common/autotest_common.sh@954 -- # kill -0 69623 00:06:11.953 17:27:43 spdkcli_tcp -- common/autotest_common.sh@955 -- # uname 00:06:11.953 17:27:43 spdkcli_tcp -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:06:12.213 17:27:43 spdkcli_tcp -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 69623 00:06:12.213 killing process with pid 69623 00:06:12.213 17:27:43 spdkcli_tcp -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:06:12.213 17:27:43 spdkcli_tcp -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:06:12.213 17:27:43 spdkcli_tcp -- common/autotest_common.sh@968 -- # echo 'killing process with pid 69623' 00:06:12.213 17:27:43 spdkcli_tcp -- common/autotest_common.sh@969 -- # kill 69623 00:06:12.213 17:27:43 spdkcli_tcp -- common/autotest_common.sh@974 -- # wait 69623 00:06:12.784 ************************************ 00:06:12.784 END TEST spdkcli_tcp 00:06:12.784 ************************************ 00:06:12.784 00:06:12.784 real 0m2.073s 00:06:12.784 user 0m3.245s 00:06:12.784 sys 0m0.726s 00:06:12.784 17:27:43 spdkcli_tcp -- common/autotest_common.sh@1126 -- # xtrace_disable 00:06:12.784 17:27:43 spdkcli_tcp -- common/autotest_common.sh@10 -- # set +x 00:06:12.784 17:27:43 -- spdk/autotest.sh@167 -- # run_test dpdk_mem_utility /home/vagrant/spdk_repo/spdk/test/dpdk_memory_utility/test_dpdk_mem_info.sh 00:06:12.784 17:27:43 -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:06:12.784 17:27:43 -- common/autotest_common.sh@1107 -- # xtrace_disable 00:06:12.784 17:27:43 -- common/autotest_common.sh@10 -- # set +x 00:06:12.784 ************************************ 00:06:12.784 START TEST dpdk_mem_utility 00:06:12.784 ************************************ 00:06:12.784 17:27:43 dpdk_mem_utility -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/dpdk_memory_utility/test_dpdk_mem_info.sh 00:06:13.045 * Looking for test storage... 00:06:13.045 * Found test storage at /home/vagrant/spdk_repo/spdk/test/dpdk_memory_utility 00:06:13.045 17:27:44 dpdk_mem_utility -- common/autotest_common.sh@1680 -- # [[ y == y ]] 00:06:13.045 17:27:44 dpdk_mem_utility -- common/autotest_common.sh@1681 -- # lcov --version 00:06:13.045 17:27:44 dpdk_mem_utility -- common/autotest_common.sh@1681 -- # awk '{print $NF}' 00:06:13.045 17:27:44 dpdk_mem_utility -- common/autotest_common.sh@1681 -- # lt 1.15 2 00:06:13.045 17:27:44 dpdk_mem_utility -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:06:13.045 17:27:44 dpdk_mem_utility -- scripts/common.sh@333 -- # local ver1 ver1_l 00:06:13.045 17:27:44 dpdk_mem_utility -- scripts/common.sh@334 -- # local ver2 ver2_l 00:06:13.045 17:27:44 dpdk_mem_utility -- scripts/common.sh@336 -- # IFS=.-: 00:06:13.045 17:27:44 dpdk_mem_utility -- scripts/common.sh@336 -- # read -ra ver1 00:06:13.045 17:27:44 dpdk_mem_utility -- scripts/common.sh@337 -- # IFS=.-: 00:06:13.045 17:27:44 dpdk_mem_utility -- scripts/common.sh@337 -- # read -ra ver2 00:06:13.045 17:27:44 dpdk_mem_utility -- scripts/common.sh@338 -- # local 'op=<' 00:06:13.045 17:27:44 dpdk_mem_utility -- scripts/common.sh@340 -- # ver1_l=2 00:06:13.045 17:27:44 dpdk_mem_utility -- scripts/common.sh@341 -- # ver2_l=1 00:06:13.045 17:27:44 dpdk_mem_utility -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:06:13.045 17:27:44 dpdk_mem_utility -- scripts/common.sh@344 -- # case "$op" in 00:06:13.045 17:27:44 dpdk_mem_utility -- scripts/common.sh@345 -- # : 1 00:06:13.045 17:27:44 dpdk_mem_utility -- scripts/common.sh@364 -- # (( v = 0 )) 00:06:13.045 17:27:44 dpdk_mem_utility -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:06:13.045 17:27:44 dpdk_mem_utility -- scripts/common.sh@365 -- # decimal 1 00:06:13.045 17:27:44 dpdk_mem_utility -- scripts/common.sh@353 -- # local d=1 00:06:13.045 17:27:44 dpdk_mem_utility -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:06:13.045 17:27:44 dpdk_mem_utility -- scripts/common.sh@355 -- # echo 1 00:06:13.045 17:27:44 dpdk_mem_utility -- scripts/common.sh@365 -- # ver1[v]=1 00:06:13.045 17:27:44 dpdk_mem_utility -- scripts/common.sh@366 -- # decimal 2 00:06:13.045 17:27:44 dpdk_mem_utility -- scripts/common.sh@353 -- # local d=2 00:06:13.045 17:27:44 dpdk_mem_utility -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:06:13.045 17:27:44 dpdk_mem_utility -- scripts/common.sh@355 -- # echo 2 00:06:13.045 17:27:44 dpdk_mem_utility -- scripts/common.sh@366 -- # ver2[v]=2 00:06:13.045 17:27:44 dpdk_mem_utility -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:06:13.045 17:27:44 dpdk_mem_utility -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:06:13.045 17:27:44 dpdk_mem_utility -- scripts/common.sh@368 -- # return 0 00:06:13.045 17:27:44 dpdk_mem_utility -- common/autotest_common.sh@1682 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:06:13.045 17:27:44 dpdk_mem_utility -- common/autotest_common.sh@1694 -- # export 'LCOV_OPTS= 00:06:13.045 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:13.045 --rc genhtml_branch_coverage=1 00:06:13.045 --rc genhtml_function_coverage=1 00:06:13.045 --rc genhtml_legend=1 00:06:13.045 --rc geninfo_all_blocks=1 00:06:13.045 --rc geninfo_unexecuted_blocks=1 00:06:13.045 00:06:13.045 ' 00:06:13.045 17:27:44 dpdk_mem_utility -- common/autotest_common.sh@1694 -- # LCOV_OPTS=' 00:06:13.045 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:13.045 --rc genhtml_branch_coverage=1 00:06:13.045 --rc genhtml_function_coverage=1 00:06:13.045 --rc genhtml_legend=1 00:06:13.045 --rc geninfo_all_blocks=1 00:06:13.045 --rc geninfo_unexecuted_blocks=1 00:06:13.045 00:06:13.045 ' 00:06:13.045 17:27:44 dpdk_mem_utility -- common/autotest_common.sh@1695 -- # export 'LCOV=lcov 00:06:13.045 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:13.045 --rc genhtml_branch_coverage=1 00:06:13.045 --rc genhtml_function_coverage=1 00:06:13.045 --rc genhtml_legend=1 00:06:13.045 --rc geninfo_all_blocks=1 00:06:13.045 --rc geninfo_unexecuted_blocks=1 00:06:13.045 00:06:13.045 ' 00:06:13.045 17:27:44 dpdk_mem_utility -- common/autotest_common.sh@1695 -- # LCOV='lcov 00:06:13.045 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:13.045 --rc genhtml_branch_coverage=1 00:06:13.045 --rc genhtml_function_coverage=1 00:06:13.045 --rc genhtml_legend=1 00:06:13.045 --rc geninfo_all_blocks=1 00:06:13.045 --rc geninfo_unexecuted_blocks=1 00:06:13.045 00:06:13.045 ' 00:06:13.045 17:27:44 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@10 -- # MEM_SCRIPT=/home/vagrant/spdk_repo/spdk/scripts/dpdk_mem_info.py 00:06:13.045 17:27:44 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@13 -- # spdkpid=69723 00:06:13.045 17:27:44 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@12 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt 00:06:13.045 17:27:44 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@15 -- # waitforlisten 69723 00:06:13.045 17:27:44 dpdk_mem_utility -- common/autotest_common.sh@831 -- # '[' -z 69723 ']' 00:06:13.045 17:27:44 dpdk_mem_utility -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:13.045 17:27:44 dpdk_mem_utility -- common/autotest_common.sh@836 -- # local max_retries=100 00:06:13.045 17:27:44 dpdk_mem_utility -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:13.045 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:13.045 17:27:44 dpdk_mem_utility -- common/autotest_common.sh@840 -- # xtrace_disable 00:06:13.045 17:27:44 dpdk_mem_utility -- common/autotest_common.sh@10 -- # set +x 00:06:13.305 [2024-11-27 17:27:44.234219] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:06:13.306 [2024-11-27 17:27:44.234369] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid69723 ] 00:06:13.306 [2024-11-27 17:27:44.379035] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:13.306 [2024-11-27 17:27:44.447866] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:06:13.875 17:27:45 dpdk_mem_utility -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:06:13.875 17:27:45 dpdk_mem_utility -- common/autotest_common.sh@864 -- # return 0 00:06:13.875 17:27:45 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@17 -- # trap 'killprocess $spdkpid' SIGINT SIGTERM EXIT 00:06:13.875 17:27:45 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@19 -- # rpc_cmd env_dpdk_get_mem_stats 00:06:13.875 17:27:45 dpdk_mem_utility -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:13.875 17:27:45 dpdk_mem_utility -- common/autotest_common.sh@10 -- # set +x 00:06:13.875 { 00:06:13.875 "filename": "/tmp/spdk_mem_dump.txt" 00:06:13.875 } 00:06:13.875 17:27:45 dpdk_mem_utility -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:13.875 17:27:45 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@21 -- # /home/vagrant/spdk_repo/spdk/scripts/dpdk_mem_info.py 00:06:14.136 DPDK memory size 860.000000 MiB in 1 heap(s) 00:06:14.136 1 heaps totaling size 860.000000 MiB 00:06:14.136 size: 860.000000 MiB heap id: 0 00:06:14.136 end heaps---------- 00:06:14.136 9 mempools totaling size 642.649841 MiB 00:06:14.136 size: 212.674988 MiB name: PDU_immediate_data_Pool 00:06:14.136 size: 158.602051 MiB name: PDU_data_out_Pool 00:06:14.136 size: 92.545471 MiB name: bdev_io_69723 00:06:14.136 size: 51.011292 MiB name: evtpool_69723 00:06:14.136 size: 50.003479 MiB name: msgpool_69723 00:06:14.136 size: 36.509338 MiB name: fsdev_io_69723 00:06:14.136 size: 21.763794 MiB name: PDU_Pool 00:06:14.136 size: 19.513306 MiB name: SCSI_TASK_Pool 00:06:14.136 size: 0.026123 MiB name: Session_Pool 00:06:14.136 end mempools------- 00:06:14.136 6 memzones totaling size 4.142822 MiB 00:06:14.136 size: 1.000366 MiB name: RG_ring_0_69723 00:06:14.136 size: 1.000366 MiB name: RG_ring_1_69723 00:06:14.136 size: 1.000366 MiB name: RG_ring_4_69723 00:06:14.136 size: 1.000366 MiB name: RG_ring_5_69723 00:06:14.136 size: 0.125366 MiB name: RG_ring_2_69723 00:06:14.136 size: 0.015991 MiB name: RG_ring_3_69723 00:06:14.136 end memzones------- 00:06:14.136 17:27:45 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@23 -- # /home/vagrant/spdk_repo/spdk/scripts/dpdk_mem_info.py -m 0 00:06:14.136 heap id: 0 total size: 860.000000 MiB number of busy elements: 312 number of free elements: 16 00:06:14.136 list of free elements. size: 13.935608 MiB 00:06:14.136 element at address: 0x200000400000 with size: 1.999512 MiB 00:06:14.137 element at address: 0x200000800000 with size: 1.996948 MiB 00:06:14.137 element at address: 0x20001bc00000 with size: 0.999878 MiB 00:06:14.137 element at address: 0x20001be00000 with size: 0.999878 MiB 00:06:14.137 element at address: 0x200034a00000 with size: 0.994446 MiB 00:06:14.137 element at address: 0x200009600000 with size: 0.959839 MiB 00:06:14.137 element at address: 0x200015e00000 with size: 0.954285 MiB 00:06:14.137 element at address: 0x20001c000000 with size: 0.936584 MiB 00:06:14.137 element at address: 0x200000200000 with size: 0.835022 MiB 00:06:14.137 element at address: 0x20001d800000 with size: 0.567505 MiB 00:06:14.137 element at address: 0x20000d800000 with size: 0.489258 MiB 00:06:14.137 element at address: 0x200003e00000 with size: 0.487549 MiB 00:06:14.137 element at address: 0x20001c200000 with size: 0.485657 MiB 00:06:14.137 element at address: 0x200007000000 with size: 0.480286 MiB 00:06:14.137 element at address: 0x20002ac00000 with size: 0.395752 MiB 00:06:14.137 element at address: 0x200003a00000 with size: 0.353210 MiB 00:06:14.137 list of standard malloc elements. size: 199.267700 MiB 00:06:14.137 element at address: 0x20000d9fff80 with size: 132.000122 MiB 00:06:14.137 element at address: 0x2000097fff80 with size: 64.000122 MiB 00:06:14.137 element at address: 0x20001bcfff80 with size: 1.000122 MiB 00:06:14.137 element at address: 0x20001befff80 with size: 1.000122 MiB 00:06:14.137 element at address: 0x20001c0fff80 with size: 1.000122 MiB 00:06:14.137 element at address: 0x2000003d9f00 with size: 0.140747 MiB 00:06:14.137 element at address: 0x20001c0eff00 with size: 0.062622 MiB 00:06:14.137 element at address: 0x2000003fdf80 with size: 0.007935 MiB 00:06:14.137 element at address: 0x20001c0efdc0 with size: 0.000305 MiB 00:06:14.137 element at address: 0x2000002d5c40 with size: 0.000183 MiB 00:06:14.137 element at address: 0x2000002d5d00 with size: 0.000183 MiB 00:06:14.137 element at address: 0x2000002d5dc0 with size: 0.000183 MiB 00:06:14.137 element at address: 0x2000002d5e80 with size: 0.000183 MiB 00:06:14.137 element at address: 0x2000002d5f40 with size: 0.000183 MiB 00:06:14.137 element at address: 0x2000002d6000 with size: 0.000183 MiB 00:06:14.137 element at address: 0x2000002d60c0 with size: 0.000183 MiB 00:06:14.137 element at address: 0x2000002d6180 with size: 0.000183 MiB 00:06:14.137 element at address: 0x2000002d6240 with size: 0.000183 MiB 00:06:14.137 element at address: 0x2000002d6300 with size: 0.000183 MiB 00:06:14.137 element at address: 0x2000002d63c0 with size: 0.000183 MiB 00:06:14.137 element at address: 0x2000002d6480 with size: 0.000183 MiB 00:06:14.137 element at address: 0x2000002d6540 with size: 0.000183 MiB 00:06:14.137 element at address: 0x2000002d6600 with size: 0.000183 MiB 00:06:14.137 element at address: 0x2000002d66c0 with size: 0.000183 MiB 00:06:14.137 element at address: 0x2000002d68c0 with size: 0.000183 MiB 00:06:14.137 element at address: 0x2000002d6980 with size: 0.000183 MiB 00:06:14.137 element at address: 0x2000002d6a40 with size: 0.000183 MiB 00:06:14.137 element at address: 0x2000002d6b00 with size: 0.000183 MiB 00:06:14.137 element at address: 0x2000002d6bc0 with size: 0.000183 MiB 00:06:14.137 element at address: 0x2000002d6c80 with size: 0.000183 MiB 00:06:14.137 element at address: 0x2000002d6d40 with size: 0.000183 MiB 00:06:14.137 element at address: 0x2000002d6e00 with size: 0.000183 MiB 00:06:14.137 element at address: 0x2000002d6ec0 with size: 0.000183 MiB 00:06:14.137 element at address: 0x2000002d6f80 with size: 0.000183 MiB 00:06:14.137 element at address: 0x2000002d7040 with size: 0.000183 MiB 00:06:14.137 element at address: 0x2000002d7100 with size: 0.000183 MiB 00:06:14.137 element at address: 0x2000002d71c0 with size: 0.000183 MiB 00:06:14.137 element at address: 0x2000002d7280 with size: 0.000183 MiB 00:06:14.137 element at address: 0x2000002d7340 with size: 0.000183 MiB 00:06:14.137 element at address: 0x2000002d7400 with size: 0.000183 MiB 00:06:14.137 element at address: 0x2000002d74c0 with size: 0.000183 MiB 00:06:14.137 element at address: 0x2000002d7580 with size: 0.000183 MiB 00:06:14.137 element at address: 0x2000002d7640 with size: 0.000183 MiB 00:06:14.137 element at address: 0x2000002d7700 with size: 0.000183 MiB 00:06:14.137 element at address: 0x2000002d77c0 with size: 0.000183 MiB 00:06:14.137 element at address: 0x2000002d7880 with size: 0.000183 MiB 00:06:14.137 element at address: 0x2000002d7940 with size: 0.000183 MiB 00:06:14.137 element at address: 0x2000002d7a00 with size: 0.000183 MiB 00:06:14.137 element at address: 0x2000002d7ac0 with size: 0.000183 MiB 00:06:14.137 element at address: 0x2000002d7b80 with size: 0.000183 MiB 00:06:14.137 element at address: 0x2000002d7c40 with size: 0.000183 MiB 00:06:14.137 element at address: 0x2000003d9e40 with size: 0.000183 MiB 00:06:14.137 element at address: 0x200003a5a6c0 with size: 0.000183 MiB 00:06:14.137 element at address: 0x200003a5a8c0 with size: 0.000183 MiB 00:06:14.137 element at address: 0x200003a5eb80 with size: 0.000183 MiB 00:06:14.137 element at address: 0x200003a7ee40 with size: 0.000183 MiB 00:06:14.137 element at address: 0x200003a7ef00 with size: 0.000183 MiB 00:06:14.137 element at address: 0x200003a7efc0 with size: 0.000183 MiB 00:06:14.137 element at address: 0x200003a7f080 with size: 0.000183 MiB 00:06:14.137 element at address: 0x200003a7f140 with size: 0.000183 MiB 00:06:14.137 element at address: 0x200003a7f200 with size: 0.000183 MiB 00:06:14.137 element at address: 0x200003a7f2c0 with size: 0.000183 MiB 00:06:14.137 element at address: 0x200003a7f380 with size: 0.000183 MiB 00:06:14.137 element at address: 0x200003a7f440 with size: 0.000183 MiB 00:06:14.137 element at address: 0x200003a7f500 with size: 0.000183 MiB 00:06:14.137 element at address: 0x200003a7f5c0 with size: 0.000183 MiB 00:06:14.137 element at address: 0x200003aff880 with size: 0.000183 MiB 00:06:14.137 element at address: 0x200003affa80 with size: 0.000183 MiB 00:06:14.137 element at address: 0x200003affb40 with size: 0.000183 MiB 00:06:14.137 element at address: 0x200003e7cd00 with size: 0.000183 MiB 00:06:14.137 element at address: 0x200003e7cdc0 with size: 0.000183 MiB 00:06:14.137 element at address: 0x200003e7ce80 with size: 0.000183 MiB 00:06:14.137 element at address: 0x200003e7cf40 with size: 0.000183 MiB 00:06:14.137 element at address: 0x200003e7d000 with size: 0.000183 MiB 00:06:14.137 element at address: 0x200003e7d0c0 with size: 0.000183 MiB 00:06:14.137 element at address: 0x200003e7d180 with size: 0.000183 MiB 00:06:14.137 element at address: 0x200003e7d240 with size: 0.000183 MiB 00:06:14.137 element at address: 0x200003e7d300 with size: 0.000183 MiB 00:06:14.137 element at address: 0x200003e7d3c0 with size: 0.000183 MiB 00:06:14.137 element at address: 0x200003e7d480 with size: 0.000183 MiB 00:06:14.137 element at address: 0x200003e7d540 with size: 0.000183 MiB 00:06:14.137 element at address: 0x200003e7d600 with size: 0.000183 MiB 00:06:14.137 element at address: 0x200003e7d6c0 with size: 0.000183 MiB 00:06:14.137 element at address: 0x200003e7d780 with size: 0.000183 MiB 00:06:14.137 element at address: 0x200003e7d840 with size: 0.000183 MiB 00:06:14.137 element at address: 0x200003e7d900 with size: 0.000183 MiB 00:06:14.137 element at address: 0x200003e7d9c0 with size: 0.000183 MiB 00:06:14.137 element at address: 0x200003e7da80 with size: 0.000183 MiB 00:06:14.137 element at address: 0x200003e7db40 with size: 0.000183 MiB 00:06:14.137 element at address: 0x200003e7dc00 with size: 0.000183 MiB 00:06:14.137 element at address: 0x200003e7dcc0 with size: 0.000183 MiB 00:06:14.137 element at address: 0x200003e7dd80 with size: 0.000183 MiB 00:06:14.137 element at address: 0x200003e7de40 with size: 0.000183 MiB 00:06:14.137 element at address: 0x200003e7df00 with size: 0.000183 MiB 00:06:14.137 element at address: 0x200003e7dfc0 with size: 0.000183 MiB 00:06:14.137 element at address: 0x200003e7e080 with size: 0.000183 MiB 00:06:14.137 element at address: 0x200003e7e140 with size: 0.000183 MiB 00:06:14.137 element at address: 0x200003e7e200 with size: 0.000183 MiB 00:06:14.137 element at address: 0x200003e7e2c0 with size: 0.000183 MiB 00:06:14.137 element at address: 0x200003e7e380 with size: 0.000183 MiB 00:06:14.137 element at address: 0x200003e7e440 with size: 0.000183 MiB 00:06:14.137 element at address: 0x200003e7e500 with size: 0.000183 MiB 00:06:14.137 element at address: 0x200003e7e5c0 with size: 0.000183 MiB 00:06:14.137 element at address: 0x200003e7e680 with size: 0.000183 MiB 00:06:14.137 element at address: 0x200003e7e740 with size: 0.000183 MiB 00:06:14.137 element at address: 0x200003e7e800 with size: 0.000183 MiB 00:06:14.137 element at address: 0x200003e7e8c0 with size: 0.000183 MiB 00:06:14.137 element at address: 0x200003e7e980 with size: 0.000183 MiB 00:06:14.137 element at address: 0x200003e7ea40 with size: 0.000183 MiB 00:06:14.137 element at address: 0x200003e7eb00 with size: 0.000183 MiB 00:06:14.137 element at address: 0x200003e7ebc0 with size: 0.000183 MiB 00:06:14.137 element at address: 0x200003e7ec80 with size: 0.000183 MiB 00:06:14.137 element at address: 0x200003e7ed40 with size: 0.000183 MiB 00:06:14.137 element at address: 0x200003e7ee00 with size: 0.000183 MiB 00:06:14.137 element at address: 0x200003eff0c0 with size: 0.000183 MiB 00:06:14.137 element at address: 0x20000707af40 with size: 0.000183 MiB 00:06:14.137 element at address: 0x20000707b000 with size: 0.000183 MiB 00:06:14.137 element at address: 0x20000707b0c0 with size: 0.000183 MiB 00:06:14.137 element at address: 0x20000707b180 with size: 0.000183 MiB 00:06:14.137 element at address: 0x20000707b240 with size: 0.000183 MiB 00:06:14.137 element at address: 0x20000707b300 with size: 0.000183 MiB 00:06:14.137 element at address: 0x20000707b3c0 with size: 0.000183 MiB 00:06:14.137 element at address: 0x20000707b480 with size: 0.000183 MiB 00:06:14.137 element at address: 0x20000707b540 with size: 0.000183 MiB 00:06:14.137 element at address: 0x20000707b600 with size: 0.000183 MiB 00:06:14.137 element at address: 0x20000707b6c0 with size: 0.000183 MiB 00:06:14.137 element at address: 0x2000070fb980 with size: 0.000183 MiB 00:06:14.137 element at address: 0x2000096fdd80 with size: 0.000183 MiB 00:06:14.137 element at address: 0x20000d87d400 with size: 0.000183 MiB 00:06:14.137 element at address: 0x20000d87d4c0 with size: 0.000183 MiB 00:06:14.137 element at address: 0x20000d87d580 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20000d87d640 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20000d87d700 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20000d87d7c0 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20000d87d880 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20000d87d940 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20000d87da00 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20000d87dac0 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20000d8fdd80 with size: 0.000183 MiB 00:06:14.138 element at address: 0x200015ef44c0 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20001c0efc40 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20001c0efd00 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20001c2bc740 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20001d891480 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20001d891540 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20001d891600 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20001d8916c0 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20001d891780 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20001d891840 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20001d891900 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20001d8919c0 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20001d891a80 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20001d891b40 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20001d891c00 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20001d891cc0 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20001d891d80 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20001d891e40 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20001d891f00 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20001d891fc0 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20001d892080 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20001d892140 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20001d892200 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20001d8922c0 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20001d892380 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20001d892440 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20001d892500 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20001d8925c0 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20001d892680 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20001d892740 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20001d892800 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20001d8928c0 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20001d892980 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20001d892a40 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20001d892b00 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20001d892bc0 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20001d892c80 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20001d892d40 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20001d892e00 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20001d892ec0 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20001d892f80 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20001d893040 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20001d893100 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20001d8931c0 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20001d893280 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20001d893340 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20001d893400 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20001d8934c0 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20001d893580 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20001d893640 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20001d893700 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20001d8937c0 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20001d893880 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20001d893940 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20001d893a00 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20001d893ac0 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20001d893b80 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20001d893c40 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20001d893d00 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20001d893dc0 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20001d893e80 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20001d893f40 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20001d894000 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20001d8940c0 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20001d894180 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20001d894240 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20001d894300 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20001d8943c0 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20001d894480 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20001d894540 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20001d894600 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20001d8946c0 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20001d894780 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20001d894840 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20001d894900 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20001d8949c0 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20001d894a80 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20001d894b40 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20001d894c00 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20001d894cc0 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20001d894d80 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20001d894e40 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20001d894f00 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20001d894fc0 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20001d895080 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20001d895140 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20001d895200 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20001d8952c0 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20001d895380 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20001d895440 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20002ac65500 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20002ac655c0 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20002ac6c1c0 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20002ac6c3c0 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20002ac6c480 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20002ac6c540 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20002ac6c600 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20002ac6c6c0 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20002ac6c780 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20002ac6c840 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20002ac6c900 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20002ac6c9c0 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20002ac6ca80 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20002ac6cb40 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20002ac6cc00 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20002ac6ccc0 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20002ac6cd80 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20002ac6ce40 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20002ac6cf00 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20002ac6cfc0 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20002ac6d080 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20002ac6d140 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20002ac6d200 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20002ac6d2c0 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20002ac6d380 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20002ac6d440 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20002ac6d500 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20002ac6d5c0 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20002ac6d680 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20002ac6d740 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20002ac6d800 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20002ac6d8c0 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20002ac6d980 with size: 0.000183 MiB 00:06:14.138 element at address: 0x20002ac6da40 with size: 0.000183 MiB 00:06:14.139 element at address: 0x20002ac6db00 with size: 0.000183 MiB 00:06:14.139 element at address: 0x20002ac6dbc0 with size: 0.000183 MiB 00:06:14.139 element at address: 0x20002ac6dc80 with size: 0.000183 MiB 00:06:14.139 element at address: 0x20002ac6dd40 with size: 0.000183 MiB 00:06:14.139 element at address: 0x20002ac6de00 with size: 0.000183 MiB 00:06:14.139 element at address: 0x20002ac6dec0 with size: 0.000183 MiB 00:06:14.139 element at address: 0x20002ac6df80 with size: 0.000183 MiB 00:06:14.139 element at address: 0x20002ac6e040 with size: 0.000183 MiB 00:06:14.139 element at address: 0x20002ac6e100 with size: 0.000183 MiB 00:06:14.139 element at address: 0x20002ac6e1c0 with size: 0.000183 MiB 00:06:14.139 element at address: 0x20002ac6e280 with size: 0.000183 MiB 00:06:14.139 element at address: 0x20002ac6e340 with size: 0.000183 MiB 00:06:14.139 element at address: 0x20002ac6e400 with size: 0.000183 MiB 00:06:14.139 element at address: 0x20002ac6e4c0 with size: 0.000183 MiB 00:06:14.139 element at address: 0x20002ac6e580 with size: 0.000183 MiB 00:06:14.139 element at address: 0x20002ac6e640 with size: 0.000183 MiB 00:06:14.139 element at address: 0x20002ac6e700 with size: 0.000183 MiB 00:06:14.139 element at address: 0x20002ac6e7c0 with size: 0.000183 MiB 00:06:14.139 element at address: 0x20002ac6e880 with size: 0.000183 MiB 00:06:14.139 element at address: 0x20002ac6e940 with size: 0.000183 MiB 00:06:14.139 element at address: 0x20002ac6ea00 with size: 0.000183 MiB 00:06:14.139 element at address: 0x20002ac6eac0 with size: 0.000183 MiB 00:06:14.139 element at address: 0x20002ac6eb80 with size: 0.000183 MiB 00:06:14.139 element at address: 0x20002ac6ec40 with size: 0.000183 MiB 00:06:14.139 element at address: 0x20002ac6ed00 with size: 0.000183 MiB 00:06:14.139 element at address: 0x20002ac6edc0 with size: 0.000183 MiB 00:06:14.139 element at address: 0x20002ac6ee80 with size: 0.000183 MiB 00:06:14.139 element at address: 0x20002ac6ef40 with size: 0.000183 MiB 00:06:14.139 element at address: 0x20002ac6f000 with size: 0.000183 MiB 00:06:14.139 element at address: 0x20002ac6f0c0 with size: 0.000183 MiB 00:06:14.139 element at address: 0x20002ac6f180 with size: 0.000183 MiB 00:06:14.139 element at address: 0x20002ac6f240 with size: 0.000183 MiB 00:06:14.139 element at address: 0x20002ac6f300 with size: 0.000183 MiB 00:06:14.139 element at address: 0x20002ac6f3c0 with size: 0.000183 MiB 00:06:14.139 element at address: 0x20002ac6f480 with size: 0.000183 MiB 00:06:14.139 element at address: 0x20002ac6f540 with size: 0.000183 MiB 00:06:14.139 element at address: 0x20002ac6f600 with size: 0.000183 MiB 00:06:14.139 element at address: 0x20002ac6f6c0 with size: 0.000183 MiB 00:06:14.139 element at address: 0x20002ac6f780 with size: 0.000183 MiB 00:06:14.139 element at address: 0x20002ac6f840 with size: 0.000183 MiB 00:06:14.139 element at address: 0x20002ac6f900 with size: 0.000183 MiB 00:06:14.139 element at address: 0x20002ac6f9c0 with size: 0.000183 MiB 00:06:14.139 element at address: 0x20002ac6fa80 with size: 0.000183 MiB 00:06:14.139 element at address: 0x20002ac6fb40 with size: 0.000183 MiB 00:06:14.139 element at address: 0x20002ac6fc00 with size: 0.000183 MiB 00:06:14.139 element at address: 0x20002ac6fcc0 with size: 0.000183 MiB 00:06:14.139 element at address: 0x20002ac6fd80 with size: 0.000183 MiB 00:06:14.139 element at address: 0x20002ac6fe40 with size: 0.000183 MiB 00:06:14.139 element at address: 0x20002ac6ff00 with size: 0.000183 MiB 00:06:14.139 list of memzone associated elements. size: 646.796692 MiB 00:06:14.139 element at address: 0x20001d895500 with size: 211.416748 MiB 00:06:14.139 associated memzone info: size: 211.416626 MiB name: MP_PDU_immediate_data_Pool_0 00:06:14.139 element at address: 0x20002ac6ffc0 with size: 157.562561 MiB 00:06:14.139 associated memzone info: size: 157.562439 MiB name: MP_PDU_data_out_Pool_0 00:06:14.139 element at address: 0x200015ff4780 with size: 92.045044 MiB 00:06:14.139 associated memzone info: size: 92.044922 MiB name: MP_bdev_io_69723_0 00:06:14.139 element at address: 0x2000009ff380 with size: 48.003052 MiB 00:06:14.139 associated memzone info: size: 48.002930 MiB name: MP_evtpool_69723_0 00:06:14.139 element at address: 0x200003fff380 with size: 48.003052 MiB 00:06:14.139 associated memzone info: size: 48.002930 MiB name: MP_msgpool_69723_0 00:06:14.139 element at address: 0x2000071fdb80 with size: 36.008911 MiB 00:06:14.139 associated memzone info: size: 36.008789 MiB name: MP_fsdev_io_69723_0 00:06:14.139 element at address: 0x20001c3be940 with size: 20.255554 MiB 00:06:14.139 associated memzone info: size: 20.255432 MiB name: MP_PDU_Pool_0 00:06:14.139 element at address: 0x200034bfeb40 with size: 18.005066 MiB 00:06:14.139 associated memzone info: size: 18.004944 MiB name: MP_SCSI_TASK_Pool_0 00:06:14.139 element at address: 0x2000005ffe00 with size: 2.000488 MiB 00:06:14.139 associated memzone info: size: 2.000366 MiB name: RG_MP_evtpool_69723 00:06:14.139 element at address: 0x200003bffe00 with size: 2.000488 MiB 00:06:14.139 associated memzone info: size: 2.000366 MiB name: RG_MP_msgpool_69723 00:06:14.139 element at address: 0x2000002d7d00 with size: 1.008118 MiB 00:06:14.139 associated memzone info: size: 1.007996 MiB name: MP_evtpool_69723 00:06:14.139 element at address: 0x20000d8fde40 with size: 1.008118 MiB 00:06:14.139 associated memzone info: size: 1.007996 MiB name: MP_PDU_Pool 00:06:14.139 element at address: 0x20001c2bc800 with size: 1.008118 MiB 00:06:14.139 associated memzone info: size: 1.007996 MiB name: MP_PDU_immediate_data_Pool 00:06:14.139 element at address: 0x2000096fde40 with size: 1.008118 MiB 00:06:14.139 associated memzone info: size: 1.007996 MiB name: MP_PDU_data_out_Pool 00:06:14.139 element at address: 0x2000070fba40 with size: 1.008118 MiB 00:06:14.139 associated memzone info: size: 1.007996 MiB name: MP_SCSI_TASK_Pool 00:06:14.139 element at address: 0x200003eff180 with size: 1.000488 MiB 00:06:14.139 associated memzone info: size: 1.000366 MiB name: RG_ring_0_69723 00:06:14.139 element at address: 0x200003affc00 with size: 1.000488 MiB 00:06:14.139 associated memzone info: size: 1.000366 MiB name: RG_ring_1_69723 00:06:14.139 element at address: 0x200015ef4580 with size: 1.000488 MiB 00:06:14.139 associated memzone info: size: 1.000366 MiB name: RG_ring_4_69723 00:06:14.139 element at address: 0x200034afe940 with size: 1.000488 MiB 00:06:14.139 associated memzone info: size: 1.000366 MiB name: RG_ring_5_69723 00:06:14.139 element at address: 0x200003a7f680 with size: 0.500488 MiB 00:06:14.139 associated memzone info: size: 0.500366 MiB name: RG_MP_fsdev_io_69723 00:06:14.139 element at address: 0x200003e7eec0 with size: 0.500488 MiB 00:06:14.139 associated memzone info: size: 0.500366 MiB name: RG_MP_bdev_io_69723 00:06:14.139 element at address: 0x20000d87db80 with size: 0.500488 MiB 00:06:14.139 associated memzone info: size: 0.500366 MiB name: RG_MP_PDU_Pool 00:06:14.139 element at address: 0x20000707b780 with size: 0.500488 MiB 00:06:14.139 associated memzone info: size: 0.500366 MiB name: RG_MP_SCSI_TASK_Pool 00:06:14.139 element at address: 0x20001c27c540 with size: 0.250488 MiB 00:06:14.139 associated memzone info: size: 0.250366 MiB name: RG_MP_PDU_immediate_data_Pool 00:06:14.139 element at address: 0x200003a5ec40 with size: 0.125488 MiB 00:06:14.139 associated memzone info: size: 0.125366 MiB name: RG_ring_2_69723 00:06:14.139 element at address: 0x2000096f5b80 with size: 0.031738 MiB 00:06:14.139 associated memzone info: size: 0.031616 MiB name: RG_MP_PDU_data_out_Pool 00:06:14.139 element at address: 0x20002ac65680 with size: 0.023743 MiB 00:06:14.139 associated memzone info: size: 0.023621 MiB name: MP_Session_Pool_0 00:06:14.139 element at address: 0x200003a5a980 with size: 0.016113 MiB 00:06:14.139 associated memzone info: size: 0.015991 MiB name: RG_ring_3_69723 00:06:14.139 element at address: 0x20002ac6b7c0 with size: 0.002441 MiB 00:06:14.139 associated memzone info: size: 0.002319 MiB name: RG_MP_Session_Pool 00:06:14.139 element at address: 0x2000002d6780 with size: 0.000305 MiB 00:06:14.139 associated memzone info: size: 0.000183 MiB name: MP_msgpool_69723 00:06:14.139 element at address: 0x200003aff940 with size: 0.000305 MiB 00:06:14.139 associated memzone info: size: 0.000183 MiB name: MP_fsdev_io_69723 00:06:14.139 element at address: 0x200003a5a780 with size: 0.000305 MiB 00:06:14.139 associated memzone info: size: 0.000183 MiB name: MP_bdev_io_69723 00:06:14.139 element at address: 0x20002ac6c280 with size: 0.000305 MiB 00:06:14.139 associated memzone info: size: 0.000183 MiB name: MP_Session_Pool 00:06:14.139 17:27:45 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@25 -- # trap - SIGINT SIGTERM EXIT 00:06:14.139 17:27:45 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@26 -- # killprocess 69723 00:06:14.139 17:27:45 dpdk_mem_utility -- common/autotest_common.sh@950 -- # '[' -z 69723 ']' 00:06:14.139 17:27:45 dpdk_mem_utility -- common/autotest_common.sh@954 -- # kill -0 69723 00:06:14.139 17:27:45 dpdk_mem_utility -- common/autotest_common.sh@955 -- # uname 00:06:14.139 17:27:45 dpdk_mem_utility -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:06:14.139 17:27:45 dpdk_mem_utility -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 69723 00:06:14.139 17:27:45 dpdk_mem_utility -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:06:14.139 17:27:45 dpdk_mem_utility -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:06:14.139 17:27:45 dpdk_mem_utility -- common/autotest_common.sh@968 -- # echo 'killing process with pid 69723' 00:06:14.139 killing process with pid 69723 00:06:14.139 17:27:45 dpdk_mem_utility -- common/autotest_common.sh@969 -- # kill 69723 00:06:14.139 17:27:45 dpdk_mem_utility -- common/autotest_common.sh@974 -- # wait 69723 00:06:14.710 00:06:14.710 real 0m1.988s 00:06:14.710 user 0m1.782s 00:06:14.710 sys 0m0.683s 00:06:14.710 17:27:45 dpdk_mem_utility -- common/autotest_common.sh@1126 -- # xtrace_disable 00:06:14.710 17:27:45 dpdk_mem_utility -- common/autotest_common.sh@10 -- # set +x 00:06:14.710 ************************************ 00:06:14.710 END TEST dpdk_mem_utility 00:06:14.710 ************************************ 00:06:14.971 17:27:45 -- spdk/autotest.sh@168 -- # run_test event /home/vagrant/spdk_repo/spdk/test/event/event.sh 00:06:14.971 17:27:45 -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:06:14.971 17:27:45 -- common/autotest_common.sh@1107 -- # xtrace_disable 00:06:14.971 17:27:45 -- common/autotest_common.sh@10 -- # set +x 00:06:14.971 ************************************ 00:06:14.971 START TEST event 00:06:14.971 ************************************ 00:06:14.971 17:27:45 event -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/event/event.sh 00:06:14.971 * Looking for test storage... 00:06:14.971 * Found test storage at /home/vagrant/spdk_repo/spdk/test/event 00:06:14.971 17:27:46 event -- common/autotest_common.sh@1680 -- # [[ y == y ]] 00:06:14.971 17:27:46 event -- common/autotest_common.sh@1681 -- # lcov --version 00:06:14.971 17:27:46 event -- common/autotest_common.sh@1681 -- # awk '{print $NF}' 00:06:15.231 17:27:46 event -- common/autotest_common.sh@1681 -- # lt 1.15 2 00:06:15.231 17:27:46 event -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:06:15.231 17:27:46 event -- scripts/common.sh@333 -- # local ver1 ver1_l 00:06:15.231 17:27:46 event -- scripts/common.sh@334 -- # local ver2 ver2_l 00:06:15.231 17:27:46 event -- scripts/common.sh@336 -- # IFS=.-: 00:06:15.231 17:27:46 event -- scripts/common.sh@336 -- # read -ra ver1 00:06:15.231 17:27:46 event -- scripts/common.sh@337 -- # IFS=.-: 00:06:15.231 17:27:46 event -- scripts/common.sh@337 -- # read -ra ver2 00:06:15.231 17:27:46 event -- scripts/common.sh@338 -- # local 'op=<' 00:06:15.231 17:27:46 event -- scripts/common.sh@340 -- # ver1_l=2 00:06:15.231 17:27:46 event -- scripts/common.sh@341 -- # ver2_l=1 00:06:15.231 17:27:46 event -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:06:15.231 17:27:46 event -- scripts/common.sh@344 -- # case "$op" in 00:06:15.231 17:27:46 event -- scripts/common.sh@345 -- # : 1 00:06:15.231 17:27:46 event -- scripts/common.sh@364 -- # (( v = 0 )) 00:06:15.231 17:27:46 event -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:06:15.231 17:27:46 event -- scripts/common.sh@365 -- # decimal 1 00:06:15.231 17:27:46 event -- scripts/common.sh@353 -- # local d=1 00:06:15.231 17:27:46 event -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:06:15.231 17:27:46 event -- scripts/common.sh@355 -- # echo 1 00:06:15.231 17:27:46 event -- scripts/common.sh@365 -- # ver1[v]=1 00:06:15.231 17:27:46 event -- scripts/common.sh@366 -- # decimal 2 00:06:15.231 17:27:46 event -- scripts/common.sh@353 -- # local d=2 00:06:15.231 17:27:46 event -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:06:15.231 17:27:46 event -- scripts/common.sh@355 -- # echo 2 00:06:15.231 17:27:46 event -- scripts/common.sh@366 -- # ver2[v]=2 00:06:15.231 17:27:46 event -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:06:15.232 17:27:46 event -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:06:15.232 17:27:46 event -- scripts/common.sh@368 -- # return 0 00:06:15.232 17:27:46 event -- common/autotest_common.sh@1682 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:06:15.232 17:27:46 event -- common/autotest_common.sh@1694 -- # export 'LCOV_OPTS= 00:06:15.232 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:15.232 --rc genhtml_branch_coverage=1 00:06:15.232 --rc genhtml_function_coverage=1 00:06:15.232 --rc genhtml_legend=1 00:06:15.232 --rc geninfo_all_blocks=1 00:06:15.232 --rc geninfo_unexecuted_blocks=1 00:06:15.232 00:06:15.232 ' 00:06:15.232 17:27:46 event -- common/autotest_common.sh@1694 -- # LCOV_OPTS=' 00:06:15.232 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:15.232 --rc genhtml_branch_coverage=1 00:06:15.232 --rc genhtml_function_coverage=1 00:06:15.232 --rc genhtml_legend=1 00:06:15.232 --rc geninfo_all_blocks=1 00:06:15.232 --rc geninfo_unexecuted_blocks=1 00:06:15.232 00:06:15.232 ' 00:06:15.232 17:27:46 event -- common/autotest_common.sh@1695 -- # export 'LCOV=lcov 00:06:15.232 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:15.232 --rc genhtml_branch_coverage=1 00:06:15.232 --rc genhtml_function_coverage=1 00:06:15.232 --rc genhtml_legend=1 00:06:15.232 --rc geninfo_all_blocks=1 00:06:15.232 --rc geninfo_unexecuted_blocks=1 00:06:15.232 00:06:15.232 ' 00:06:15.232 17:27:46 event -- common/autotest_common.sh@1695 -- # LCOV='lcov 00:06:15.232 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:15.232 --rc genhtml_branch_coverage=1 00:06:15.232 --rc genhtml_function_coverage=1 00:06:15.232 --rc genhtml_legend=1 00:06:15.232 --rc geninfo_all_blocks=1 00:06:15.232 --rc geninfo_unexecuted_blocks=1 00:06:15.232 00:06:15.232 ' 00:06:15.232 17:27:46 event -- event/event.sh@9 -- # source /home/vagrant/spdk_repo/spdk/test/bdev/nbd_common.sh 00:06:15.232 17:27:46 event -- bdev/nbd_common.sh@6 -- # set -e 00:06:15.232 17:27:46 event -- event/event.sh@45 -- # run_test event_perf /home/vagrant/spdk_repo/spdk/test/event/event_perf/event_perf -m 0xF -t 1 00:06:15.232 17:27:46 event -- common/autotest_common.sh@1101 -- # '[' 6 -le 1 ']' 00:06:15.232 17:27:46 event -- common/autotest_common.sh@1107 -- # xtrace_disable 00:06:15.232 17:27:46 event -- common/autotest_common.sh@10 -- # set +x 00:06:15.232 ************************************ 00:06:15.232 START TEST event_perf 00:06:15.232 ************************************ 00:06:15.232 17:27:46 event.event_perf -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/event/event_perf/event_perf -m 0xF -t 1 00:06:15.232 Running I/O for 1 seconds...[2024-11-27 17:27:46.242271] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:06:15.232 [2024-11-27 17:27:46.242468] [ DPDK EAL parameters: event_perf --no-shconf -c 0xF --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid69809 ] 00:06:15.232 [2024-11-27 17:27:46.389193] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 4 00:06:15.491 [2024-11-27 17:27:46.472016] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 1 00:06:15.491 Running I/O for 1 seconds...[2024-11-27 17:27:46.472275] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 2 00:06:15.491 [2024-11-27 17:27:46.472906] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:06:15.491 [2024-11-27 17:27:46.473045] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 3 00:06:16.430 00:06:16.430 lcore 0: 80197 00:06:16.430 lcore 1: 80200 00:06:16.430 lcore 2: 80198 00:06:16.430 lcore 3: 80194 00:06:16.430 done. 00:06:16.430 00:06:16.430 real 0m1.409s 00:06:16.430 user 0m4.152s 00:06:16.430 sys 0m0.134s 00:06:16.430 17:27:47 event.event_perf -- common/autotest_common.sh@1126 -- # xtrace_disable 00:06:16.430 17:27:47 event.event_perf -- common/autotest_common.sh@10 -- # set +x 00:06:16.430 ************************************ 00:06:16.430 END TEST event_perf 00:06:16.430 ************************************ 00:06:16.691 17:27:47 event -- event/event.sh@46 -- # run_test event_reactor /home/vagrant/spdk_repo/spdk/test/event/reactor/reactor -t 1 00:06:16.691 17:27:47 event -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:06:16.691 17:27:47 event -- common/autotest_common.sh@1107 -- # xtrace_disable 00:06:16.691 17:27:47 event -- common/autotest_common.sh@10 -- # set +x 00:06:16.691 ************************************ 00:06:16.691 START TEST event_reactor 00:06:16.691 ************************************ 00:06:16.691 17:27:47 event.event_reactor -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/event/reactor/reactor -t 1 00:06:16.691 [2024-11-27 17:27:47.732630] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:06:16.691 [2024-11-27 17:27:47.732951] [ DPDK EAL parameters: reactor --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid69852 ] 00:06:16.691 [2024-11-27 17:27:47.878314] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:16.951 [2024-11-27 17:27:47.963373] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:06:17.915 test_start 00:06:17.915 oneshot 00:06:17.915 tick 100 00:06:17.915 tick 100 00:06:17.915 tick 250 00:06:17.915 tick 100 00:06:17.915 tick 100 00:06:17.915 tick 100 00:06:17.915 tick 250 00:06:17.915 tick 500 00:06:17.915 tick 100 00:06:17.915 tick 100 00:06:17.915 tick 250 00:06:17.915 tick 100 00:06:17.915 tick 100 00:06:17.915 test_end 00:06:17.915 00:06:17.915 real 0m1.404s 00:06:17.915 user 0m1.183s 00:06:17.915 sys 0m0.114s 00:06:17.915 17:27:49 event.event_reactor -- common/autotest_common.sh@1126 -- # xtrace_disable 00:06:17.915 17:27:49 event.event_reactor -- common/autotest_common.sh@10 -- # set +x 00:06:17.915 ************************************ 00:06:17.915 END TEST event_reactor 00:06:17.915 ************************************ 00:06:18.179 17:27:49 event -- event/event.sh@47 -- # run_test event_reactor_perf /home/vagrant/spdk_repo/spdk/test/event/reactor_perf/reactor_perf -t 1 00:06:18.179 17:27:49 event -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:06:18.179 17:27:49 event -- common/autotest_common.sh@1107 -- # xtrace_disable 00:06:18.179 17:27:49 event -- common/autotest_common.sh@10 -- # set +x 00:06:18.179 ************************************ 00:06:18.179 START TEST event_reactor_perf 00:06:18.179 ************************************ 00:06:18.179 17:27:49 event.event_reactor_perf -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/event/reactor_perf/reactor_perf -t 1 00:06:18.179 [2024-11-27 17:27:49.206522] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:06:18.179 [2024-11-27 17:27:49.206764] [ DPDK EAL parameters: reactor_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid69885 ] 00:06:18.179 [2024-11-27 17:27:49.353186] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:18.441 [2024-11-27 17:27:49.422766] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:06:19.380 test_start 00:06:19.380 test_end 00:06:19.380 Performance: 408715 events per second 00:06:19.380 00:06:19.380 real 0m1.391s 00:06:19.380 user 0m1.173s 00:06:19.380 sys 0m0.110s 00:06:19.380 17:27:50 event.event_reactor_perf -- common/autotest_common.sh@1126 -- # xtrace_disable 00:06:19.380 17:27:50 event.event_reactor_perf -- common/autotest_common.sh@10 -- # set +x 00:06:19.380 ************************************ 00:06:19.380 END TEST event_reactor_perf 00:06:19.380 ************************************ 00:06:19.640 17:27:50 event -- event/event.sh@49 -- # uname -s 00:06:19.640 17:27:50 event -- event/event.sh@49 -- # '[' Linux = Linux ']' 00:06:19.640 17:27:50 event -- event/event.sh@50 -- # run_test event_scheduler /home/vagrant/spdk_repo/spdk/test/event/scheduler/scheduler.sh 00:06:19.640 17:27:50 event -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:06:19.640 17:27:50 event -- common/autotest_common.sh@1107 -- # xtrace_disable 00:06:19.640 17:27:50 event -- common/autotest_common.sh@10 -- # set +x 00:06:19.640 ************************************ 00:06:19.640 START TEST event_scheduler 00:06:19.640 ************************************ 00:06:19.640 17:27:50 event.event_scheduler -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/event/scheduler/scheduler.sh 00:06:19.640 * Looking for test storage... 00:06:19.640 * Found test storage at /home/vagrant/spdk_repo/spdk/test/event/scheduler 00:06:19.640 17:27:50 event.event_scheduler -- common/autotest_common.sh@1680 -- # [[ y == y ]] 00:06:19.640 17:27:50 event.event_scheduler -- common/autotest_common.sh@1681 -- # lcov --version 00:06:19.640 17:27:50 event.event_scheduler -- common/autotest_common.sh@1681 -- # awk '{print $NF}' 00:06:19.901 17:27:50 event.event_scheduler -- common/autotest_common.sh@1681 -- # lt 1.15 2 00:06:19.901 17:27:50 event.event_scheduler -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:06:19.901 17:27:50 event.event_scheduler -- scripts/common.sh@333 -- # local ver1 ver1_l 00:06:19.901 17:27:50 event.event_scheduler -- scripts/common.sh@334 -- # local ver2 ver2_l 00:06:19.901 17:27:50 event.event_scheduler -- scripts/common.sh@336 -- # IFS=.-: 00:06:19.901 17:27:50 event.event_scheduler -- scripts/common.sh@336 -- # read -ra ver1 00:06:19.901 17:27:50 event.event_scheduler -- scripts/common.sh@337 -- # IFS=.-: 00:06:19.901 17:27:50 event.event_scheduler -- scripts/common.sh@337 -- # read -ra ver2 00:06:19.901 17:27:50 event.event_scheduler -- scripts/common.sh@338 -- # local 'op=<' 00:06:19.901 17:27:50 event.event_scheduler -- scripts/common.sh@340 -- # ver1_l=2 00:06:19.901 17:27:50 event.event_scheduler -- scripts/common.sh@341 -- # ver2_l=1 00:06:19.901 17:27:50 event.event_scheduler -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:06:19.901 17:27:50 event.event_scheduler -- scripts/common.sh@344 -- # case "$op" in 00:06:19.901 17:27:50 event.event_scheduler -- scripts/common.sh@345 -- # : 1 00:06:19.901 17:27:50 event.event_scheduler -- scripts/common.sh@364 -- # (( v = 0 )) 00:06:19.901 17:27:50 event.event_scheduler -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:06:19.901 17:27:50 event.event_scheduler -- scripts/common.sh@365 -- # decimal 1 00:06:19.901 17:27:50 event.event_scheduler -- scripts/common.sh@353 -- # local d=1 00:06:19.901 17:27:50 event.event_scheduler -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:06:19.901 17:27:50 event.event_scheduler -- scripts/common.sh@355 -- # echo 1 00:06:19.901 17:27:50 event.event_scheduler -- scripts/common.sh@365 -- # ver1[v]=1 00:06:19.901 17:27:50 event.event_scheduler -- scripts/common.sh@366 -- # decimal 2 00:06:19.901 17:27:50 event.event_scheduler -- scripts/common.sh@353 -- # local d=2 00:06:19.901 17:27:50 event.event_scheduler -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:06:19.901 17:27:50 event.event_scheduler -- scripts/common.sh@355 -- # echo 2 00:06:19.901 17:27:50 event.event_scheduler -- scripts/common.sh@366 -- # ver2[v]=2 00:06:19.901 17:27:50 event.event_scheduler -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:06:19.901 17:27:50 event.event_scheduler -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:06:19.901 17:27:50 event.event_scheduler -- scripts/common.sh@368 -- # return 0 00:06:19.901 17:27:50 event.event_scheduler -- common/autotest_common.sh@1682 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:06:19.901 17:27:50 event.event_scheduler -- common/autotest_common.sh@1694 -- # export 'LCOV_OPTS= 00:06:19.901 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:19.901 --rc genhtml_branch_coverage=1 00:06:19.901 --rc genhtml_function_coverage=1 00:06:19.901 --rc genhtml_legend=1 00:06:19.901 --rc geninfo_all_blocks=1 00:06:19.901 --rc geninfo_unexecuted_blocks=1 00:06:19.901 00:06:19.901 ' 00:06:19.901 17:27:50 event.event_scheduler -- common/autotest_common.sh@1694 -- # LCOV_OPTS=' 00:06:19.901 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:19.901 --rc genhtml_branch_coverage=1 00:06:19.901 --rc genhtml_function_coverage=1 00:06:19.901 --rc genhtml_legend=1 00:06:19.901 --rc geninfo_all_blocks=1 00:06:19.901 --rc geninfo_unexecuted_blocks=1 00:06:19.901 00:06:19.901 ' 00:06:19.901 17:27:50 event.event_scheduler -- common/autotest_common.sh@1695 -- # export 'LCOV=lcov 00:06:19.901 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:19.901 --rc genhtml_branch_coverage=1 00:06:19.901 --rc genhtml_function_coverage=1 00:06:19.901 --rc genhtml_legend=1 00:06:19.901 --rc geninfo_all_blocks=1 00:06:19.901 --rc geninfo_unexecuted_blocks=1 00:06:19.901 00:06:19.901 ' 00:06:19.901 17:27:50 event.event_scheduler -- common/autotest_common.sh@1695 -- # LCOV='lcov 00:06:19.901 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:19.901 --rc genhtml_branch_coverage=1 00:06:19.901 --rc genhtml_function_coverage=1 00:06:19.901 --rc genhtml_legend=1 00:06:19.901 --rc geninfo_all_blocks=1 00:06:19.901 --rc geninfo_unexecuted_blocks=1 00:06:19.901 00:06:19.901 ' 00:06:19.901 17:27:50 event.event_scheduler -- scheduler/scheduler.sh@29 -- # rpc=rpc_cmd 00:06:19.901 17:27:50 event.event_scheduler -- scheduler/scheduler.sh@35 -- # scheduler_pid=69960 00:06:19.901 17:27:50 event.event_scheduler -- scheduler/scheduler.sh@34 -- # /home/vagrant/spdk_repo/spdk/test/event/scheduler/scheduler -m 0xF -p 0x2 --wait-for-rpc -f 00:06:19.901 17:27:50 event.event_scheduler -- scheduler/scheduler.sh@36 -- # trap 'killprocess $scheduler_pid; exit 1' SIGINT SIGTERM EXIT 00:06:19.901 17:27:50 event.event_scheduler -- scheduler/scheduler.sh@37 -- # waitforlisten 69960 00:06:19.901 17:27:50 event.event_scheduler -- common/autotest_common.sh@831 -- # '[' -z 69960 ']' 00:06:19.901 17:27:50 event.event_scheduler -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:19.901 17:27:50 event.event_scheduler -- common/autotest_common.sh@836 -- # local max_retries=100 00:06:19.901 17:27:50 event.event_scheduler -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:19.901 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:19.901 17:27:50 event.event_scheduler -- common/autotest_common.sh@840 -- # xtrace_disable 00:06:19.901 17:27:50 event.event_scheduler -- common/autotest_common.sh@10 -- # set +x 00:06:19.901 [2024-11-27 17:27:50.955119] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:06:19.901 [2024-11-27 17:27:50.955753] [ DPDK EAL parameters: scheduler --no-shconf -c 0xF --main-lcore=2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid69960 ] 00:06:20.161 [2024-11-27 17:27:51.102192] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 4 00:06:20.161 [2024-11-27 17:27:51.154370] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:06:20.161 [2024-11-27 17:27:51.154660] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 2 00:06:20.161 [2024-11-27 17:27:51.154623] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 1 00:06:20.161 [2024-11-27 17:27:51.154827] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 3 00:06:20.731 17:27:51 event.event_scheduler -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:06:20.731 17:27:51 event.event_scheduler -- common/autotest_common.sh@864 -- # return 0 00:06:20.731 17:27:51 event.event_scheduler -- scheduler/scheduler.sh@39 -- # rpc_cmd framework_set_scheduler dynamic 00:06:20.731 17:27:51 event.event_scheduler -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:20.731 17:27:51 event.event_scheduler -- common/autotest_common.sh@10 -- # set +x 00:06:20.731 POWER: failed to open /sys/devices/system/cpu/cpu%u/cpufreq/scaling_governor 00:06:20.731 POWER: Cannot set governor of lcore 0 to userspace 00:06:20.731 POWER: failed to open /sys/devices/system/cpu/cpu%u/cpufreq/scaling_governor 00:06:20.731 POWER: Cannot set governor of lcore 0 to performance 00:06:20.731 POWER: failed to open /sys/devices/system/cpu/cpu%u/cpufreq/scaling_governor 00:06:20.731 POWER: Cannot set governor of lcore 0 to userspace 00:06:20.731 GUEST_CHANNEL: Unable to to connect to '/dev/virtio-ports/virtio.serial.port.poweragent.0' with error No such file or directory 00:06:20.731 POWER: Unable to set Power Management Environment for lcore 0 00:06:20.731 [2024-11-27 17:27:51.775581] dpdk_governor.c: 130:_init_core: *ERROR*: Failed to initialize on core0 00:06:20.731 [2024-11-27 17:27:51.775620] dpdk_governor.c: 191:_init: *ERROR*: Failed to initialize on core0 00:06:20.731 [2024-11-27 17:27:51.775645] scheduler_dynamic.c: 280:init: *NOTICE*: Unable to initialize dpdk governor 00:06:20.731 [2024-11-27 17:27:51.775676] scheduler_dynamic.c: 427:set_opts: *NOTICE*: Setting scheduler load limit to 20 00:06:20.731 [2024-11-27 17:27:51.775697] scheduler_dynamic.c: 429:set_opts: *NOTICE*: Setting scheduler core limit to 80 00:06:20.731 [2024-11-27 17:27:51.775707] scheduler_dynamic.c: 431:set_opts: *NOTICE*: Setting scheduler core busy to 95 00:06:20.731 17:27:51 event.event_scheduler -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:20.731 17:27:51 event.event_scheduler -- scheduler/scheduler.sh@40 -- # rpc_cmd framework_start_init 00:06:20.731 17:27:51 event.event_scheduler -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:20.731 17:27:51 event.event_scheduler -- common/autotest_common.sh@10 -- # set +x 00:06:20.731 [2024-11-27 17:27:51.853132] scheduler.c: 382:test_start: *NOTICE*: Scheduler test application started. 00:06:20.731 17:27:51 event.event_scheduler -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:20.731 17:27:51 event.event_scheduler -- scheduler/scheduler.sh@43 -- # run_test scheduler_create_thread scheduler_create_thread 00:06:20.731 17:27:51 event.event_scheduler -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:06:20.731 17:27:51 event.event_scheduler -- common/autotest_common.sh@1107 -- # xtrace_disable 00:06:20.731 17:27:51 event.event_scheduler -- common/autotest_common.sh@10 -- # set +x 00:06:20.731 ************************************ 00:06:20.731 START TEST scheduler_create_thread 00:06:20.731 ************************************ 00:06:20.731 17:27:51 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@1125 -- # scheduler_create_thread 00:06:20.731 17:27:51 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@12 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n active_pinned -m 0x1 -a 100 00:06:20.731 17:27:51 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:20.731 17:27:51 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:06:20.731 2 00:06:20.731 17:27:51 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:20.731 17:27:51 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@13 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n active_pinned -m 0x2 -a 100 00:06:20.731 17:27:51 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:20.731 17:27:51 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:06:20.731 3 00:06:20.731 17:27:51 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:20.731 17:27:51 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@14 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n active_pinned -m 0x4 -a 100 00:06:20.731 17:27:51 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:20.731 17:27:51 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:06:20.731 4 00:06:20.731 17:27:51 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:20.731 17:27:51 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@15 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n active_pinned -m 0x8 -a 100 00:06:20.731 17:27:51 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:20.731 17:27:51 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:06:20.991 5 00:06:20.991 17:27:51 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:20.991 17:27:51 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@16 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n idle_pinned -m 0x1 -a 0 00:06:20.991 17:27:51 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:20.991 17:27:51 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:06:20.991 6 00:06:20.991 17:27:51 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:20.991 17:27:51 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@17 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n idle_pinned -m 0x2 -a 0 00:06:20.991 17:27:51 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:20.991 17:27:51 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:06:20.991 7 00:06:20.991 17:27:51 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:20.991 17:27:51 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@18 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n idle_pinned -m 0x4 -a 0 00:06:20.991 17:27:51 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:20.991 17:27:51 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:06:20.991 8 00:06:20.991 17:27:51 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:20.991 17:27:51 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@19 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n idle_pinned -m 0x8 -a 0 00:06:20.991 17:27:51 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:20.991 17:27:51 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:06:20.991 9 00:06:20.991 17:27:51 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:20.991 17:27:51 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@21 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n one_third_active -a 30 00:06:20.991 17:27:51 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:20.991 17:27:51 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:06:22.373 10 00:06:22.373 17:27:53 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:22.373 17:27:53 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@22 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n half_active -a 0 00:06:22.373 17:27:53 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:22.373 17:27:53 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:06:22.943 17:27:54 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:22.943 17:27:54 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@22 -- # thread_id=11 00:06:22.943 17:27:54 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@23 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_set_active 11 50 00:06:22.943 17:27:54 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:22.943 17:27:54 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:06:23.882 17:27:54 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:23.882 17:27:54 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@25 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n deleted -a 100 00:06:23.882 17:27:54 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:23.882 17:27:54 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:06:24.450 17:27:55 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:24.450 17:27:55 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@25 -- # thread_id=12 00:06:24.450 17:27:55 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@26 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_delete 12 00:06:24.450 17:27:55 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:24.450 17:27:55 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:06:25.019 ************************************ 00:06:25.019 END TEST scheduler_create_thread 00:06:25.019 ************************************ 00:06:25.019 17:27:56 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:25.019 00:06:25.019 real 0m4.201s 00:06:25.019 user 0m0.029s 00:06:25.019 sys 0m0.008s 00:06:25.019 17:27:56 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@1126 -- # xtrace_disable 00:06:25.019 17:27:56 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:06:25.019 17:27:56 event.event_scheduler -- scheduler/scheduler.sh@45 -- # trap - SIGINT SIGTERM EXIT 00:06:25.019 17:27:56 event.event_scheduler -- scheduler/scheduler.sh@46 -- # killprocess 69960 00:06:25.019 17:27:56 event.event_scheduler -- common/autotest_common.sh@950 -- # '[' -z 69960 ']' 00:06:25.019 17:27:56 event.event_scheduler -- common/autotest_common.sh@954 -- # kill -0 69960 00:06:25.019 17:27:56 event.event_scheduler -- common/autotest_common.sh@955 -- # uname 00:06:25.019 17:27:56 event.event_scheduler -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:06:25.019 17:27:56 event.event_scheduler -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 69960 00:06:25.019 killing process with pid 69960 00:06:25.020 17:27:56 event.event_scheduler -- common/autotest_common.sh@956 -- # process_name=reactor_2 00:06:25.020 17:27:56 event.event_scheduler -- common/autotest_common.sh@960 -- # '[' reactor_2 = sudo ']' 00:06:25.020 17:27:56 event.event_scheduler -- common/autotest_common.sh@968 -- # echo 'killing process with pid 69960' 00:06:25.020 17:27:56 event.event_scheduler -- common/autotest_common.sh@969 -- # kill 69960 00:06:25.020 17:27:56 event.event_scheduler -- common/autotest_common.sh@974 -- # wait 69960 00:06:25.279 [2024-11-27 17:27:56.348417] scheduler.c: 360:test_shutdown: *NOTICE*: Scheduler test application stopped. 00:06:25.538 ************************************ 00:06:25.538 END TEST event_scheduler 00:06:25.538 00:06:25.538 real 0m6.018s 00:06:25.538 user 0m13.373s 00:06:25.538 sys 0m0.516s 00:06:25.538 17:27:56 event.event_scheduler -- common/autotest_common.sh@1126 -- # xtrace_disable 00:06:25.538 17:27:56 event.event_scheduler -- common/autotest_common.sh@10 -- # set +x 00:06:25.538 ************************************ 00:06:25.538 17:27:56 event -- event/event.sh@51 -- # modprobe -n nbd 00:06:25.538 17:27:56 event -- event/event.sh@52 -- # run_test app_repeat app_repeat_test 00:06:25.538 17:27:56 event -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:06:25.538 17:27:56 event -- common/autotest_common.sh@1107 -- # xtrace_disable 00:06:25.538 17:27:56 event -- common/autotest_common.sh@10 -- # set +x 00:06:25.538 ************************************ 00:06:25.538 START TEST app_repeat 00:06:25.538 ************************************ 00:06:25.538 17:27:56 event.app_repeat -- common/autotest_common.sh@1125 -- # app_repeat_test 00:06:25.799 17:27:56 event.app_repeat -- event/event.sh@12 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:25.799 17:27:56 event.app_repeat -- event/event.sh@13 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:25.799 17:27:56 event.app_repeat -- event/event.sh@13 -- # local nbd_list 00:06:25.799 17:27:56 event.app_repeat -- event/event.sh@14 -- # bdev_list=('Malloc0' 'Malloc1') 00:06:25.799 17:27:56 event.app_repeat -- event/event.sh@14 -- # local bdev_list 00:06:25.799 17:27:56 event.app_repeat -- event/event.sh@15 -- # local repeat_times=4 00:06:25.799 17:27:56 event.app_repeat -- event/event.sh@17 -- # modprobe nbd 00:06:25.799 17:27:56 event.app_repeat -- event/event.sh@19 -- # repeat_pid=70067 00:06:25.799 17:27:56 event.app_repeat -- event/event.sh@18 -- # /home/vagrant/spdk_repo/spdk/test/event/app_repeat/app_repeat -r /var/tmp/spdk-nbd.sock -m 0x3 -t 4 00:06:25.799 17:27:56 event.app_repeat -- event/event.sh@20 -- # trap 'killprocess $repeat_pid; exit 1' SIGINT SIGTERM EXIT 00:06:25.799 17:27:56 event.app_repeat -- event/event.sh@21 -- # echo 'Process app_repeat pid: 70067' 00:06:25.799 Process app_repeat pid: 70067 00:06:25.799 17:27:56 event.app_repeat -- event/event.sh@23 -- # for i in {0..2} 00:06:25.799 17:27:56 event.app_repeat -- event/event.sh@24 -- # echo 'spdk_app_start Round 0' 00:06:25.799 spdk_app_start Round 0 00:06:25.799 17:27:56 event.app_repeat -- event/event.sh@25 -- # waitforlisten 70067 /var/tmp/spdk-nbd.sock 00:06:25.799 17:27:56 event.app_repeat -- common/autotest_common.sh@831 -- # '[' -z 70067 ']' 00:06:25.799 17:27:56 event.app_repeat -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk-nbd.sock 00:06:25.799 17:27:56 event.app_repeat -- common/autotest_common.sh@836 -- # local max_retries=100 00:06:25.799 17:27:56 event.app_repeat -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock...' 00:06:25.799 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock... 00:06:25.799 17:27:56 event.app_repeat -- common/autotest_common.sh@840 -- # xtrace_disable 00:06:25.799 17:27:56 event.app_repeat -- common/autotest_common.sh@10 -- # set +x 00:06:25.799 [2024-11-27 17:27:56.787696] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:06:25.799 [2024-11-27 17:27:56.787866] [ DPDK EAL parameters: app_repeat --no-shconf -c 0x3 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70067 ] 00:06:25.799 [2024-11-27 17:27:56.930789] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 2 00:06:26.059 [2024-11-27 17:27:57.005957] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:06:26.059 [2024-11-27 17:27:57.006058] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 1 00:06:26.628 17:27:57 event.app_repeat -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:06:26.628 17:27:57 event.app_repeat -- common/autotest_common.sh@864 -- # return 0 00:06:26.628 17:27:57 event.app_repeat -- event/event.sh@27 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:06:26.888 Malloc0 00:06:26.888 17:27:57 event.app_repeat -- event/event.sh@28 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:06:26.888 Malloc1 00:06:26.888 17:27:58 event.app_repeat -- event/event.sh@30 -- # nbd_rpc_data_verify /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:06:26.888 17:27:58 event.app_repeat -- bdev/nbd_common.sh@90 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:26.888 17:27:58 event.app_repeat -- bdev/nbd_common.sh@91 -- # bdev_list=('Malloc0' 'Malloc1') 00:06:26.888 17:27:58 event.app_repeat -- bdev/nbd_common.sh@91 -- # local bdev_list 00:06:26.888 17:27:58 event.app_repeat -- bdev/nbd_common.sh@92 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:26.888 17:27:58 event.app_repeat -- bdev/nbd_common.sh@92 -- # local nbd_list 00:06:26.888 17:27:58 event.app_repeat -- bdev/nbd_common.sh@94 -- # nbd_start_disks /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:06:26.888 17:27:58 event.app_repeat -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:26.888 17:27:58 event.app_repeat -- bdev/nbd_common.sh@10 -- # bdev_list=('Malloc0' 'Malloc1') 00:06:26.888 17:27:58 event.app_repeat -- bdev/nbd_common.sh@10 -- # local bdev_list 00:06:26.888 17:27:58 event.app_repeat -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:26.888 17:27:58 event.app_repeat -- bdev/nbd_common.sh@11 -- # local nbd_list 00:06:26.888 17:27:58 event.app_repeat -- bdev/nbd_common.sh@12 -- # local i 00:06:26.888 17:27:58 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:06:26.888 17:27:58 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:06:26.888 17:27:58 event.app_repeat -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc0 /dev/nbd0 00:06:27.149 /dev/nbd0 00:06:27.149 17:27:58 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:06:27.149 17:27:58 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:06:27.149 17:27:58 event.app_repeat -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:06:27.149 17:27:58 event.app_repeat -- common/autotest_common.sh@869 -- # local i 00:06:27.149 17:27:58 event.app_repeat -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:06:27.149 17:27:58 event.app_repeat -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:06:27.149 17:27:58 event.app_repeat -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:06:27.149 17:27:58 event.app_repeat -- common/autotest_common.sh@873 -- # break 00:06:27.149 17:27:58 event.app_repeat -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:06:27.149 17:27:58 event.app_repeat -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:06:27.149 17:27:58 event.app_repeat -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:06:27.149 1+0 records in 00:06:27.149 1+0 records out 00:06:27.149 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000222018 s, 18.4 MB/s 00:06:27.149 17:27:58 event.app_repeat -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:06:27.149 17:27:58 event.app_repeat -- common/autotest_common.sh@886 -- # size=4096 00:06:27.149 17:27:58 event.app_repeat -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:06:27.149 17:27:58 event.app_repeat -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:06:27.149 17:27:58 event.app_repeat -- common/autotest_common.sh@889 -- # return 0 00:06:27.149 17:27:58 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:06:27.149 17:27:58 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:06:27.149 17:27:58 event.app_repeat -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc1 /dev/nbd1 00:06:27.409 /dev/nbd1 00:06:27.409 17:27:58 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:06:27.409 17:27:58 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:06:27.409 17:27:58 event.app_repeat -- common/autotest_common.sh@868 -- # local nbd_name=nbd1 00:06:27.409 17:27:58 event.app_repeat -- common/autotest_common.sh@869 -- # local i 00:06:27.409 17:27:58 event.app_repeat -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:06:27.409 17:27:58 event.app_repeat -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:06:27.409 17:27:58 event.app_repeat -- common/autotest_common.sh@872 -- # grep -q -w nbd1 /proc/partitions 00:06:27.409 17:27:58 event.app_repeat -- common/autotest_common.sh@873 -- # break 00:06:27.409 17:27:58 event.app_repeat -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:06:27.409 17:27:58 event.app_repeat -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:06:27.409 17:27:58 event.app_repeat -- common/autotest_common.sh@885 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:06:27.409 1+0 records in 00:06:27.409 1+0 records out 00:06:27.409 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.00037955 s, 10.8 MB/s 00:06:27.409 17:27:58 event.app_repeat -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:06:27.409 17:27:58 event.app_repeat -- common/autotest_common.sh@886 -- # size=4096 00:06:27.409 17:27:58 event.app_repeat -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:06:27.409 17:27:58 event.app_repeat -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:06:27.409 17:27:58 event.app_repeat -- common/autotest_common.sh@889 -- # return 0 00:06:27.409 17:27:58 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:06:27.409 17:27:58 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:06:27.409 17:27:58 event.app_repeat -- bdev/nbd_common.sh@95 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:06:27.409 17:27:58 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:27.410 17:27:58 event.app_repeat -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:06:27.670 17:27:58 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[ 00:06:27.670 { 00:06:27.670 "nbd_device": "/dev/nbd0", 00:06:27.670 "bdev_name": "Malloc0" 00:06:27.670 }, 00:06:27.670 { 00:06:27.670 "nbd_device": "/dev/nbd1", 00:06:27.670 "bdev_name": "Malloc1" 00:06:27.670 } 00:06:27.670 ]' 00:06:27.670 17:27:58 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[ 00:06:27.670 { 00:06:27.670 "nbd_device": "/dev/nbd0", 00:06:27.670 "bdev_name": "Malloc0" 00:06:27.670 }, 00:06:27.670 { 00:06:27.670 "nbd_device": "/dev/nbd1", 00:06:27.670 "bdev_name": "Malloc1" 00:06:27.670 } 00:06:27.670 ]' 00:06:27.670 17:27:58 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:06:27.670 17:27:58 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name='/dev/nbd0 00:06:27.670 /dev/nbd1' 00:06:27.670 17:27:58 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:06:27.670 17:27:58 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '/dev/nbd0 00:06:27.670 /dev/nbd1' 00:06:27.670 17:27:58 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=2 00:06:27.670 17:27:58 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 2 00:06:27.670 17:27:58 event.app_repeat -- bdev/nbd_common.sh@95 -- # count=2 00:06:27.670 17:27:58 event.app_repeat -- bdev/nbd_common.sh@96 -- # '[' 2 -ne 2 ']' 00:06:27.670 17:27:58 event.app_repeat -- bdev/nbd_common.sh@100 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' write 00:06:27.670 17:27:58 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:27.670 17:27:58 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:06:27.670 17:27:58 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=write 00:06:27.670 17:27:58 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:06:27.670 17:27:58 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' write = write ']' 00:06:27.670 17:27:58 event.app_repeat -- bdev/nbd_common.sh@76 -- # dd if=/dev/urandom of=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest bs=4096 count=256 00:06:27.670 256+0 records in 00:06:27.670 256+0 records out 00:06:27.670 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0138395 s, 75.8 MB/s 00:06:27.670 17:27:58 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:06:27.670 17:27:58 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest of=/dev/nbd0 bs=4096 count=256 oflag=direct 00:06:27.930 256+0 records in 00:06:27.930 256+0 records out 00:06:27.930 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0254295 s, 41.2 MB/s 00:06:27.930 17:27:58 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:06:27.930 17:27:58 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest of=/dev/nbd1 bs=4096 count=256 oflag=direct 00:06:27.930 256+0 records in 00:06:27.930 256+0 records out 00:06:27.930 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0198108 s, 52.9 MB/s 00:06:27.930 17:27:58 event.app_repeat -- bdev/nbd_common.sh@101 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' verify 00:06:27.930 17:27:58 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:27.930 17:27:58 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:06:27.930 17:27:58 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=verify 00:06:27.930 17:27:58 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:06:27.930 17:27:58 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' verify = write ']' 00:06:27.930 17:27:58 event.app_repeat -- bdev/nbd_common.sh@80 -- # '[' verify = verify ']' 00:06:27.930 17:27:58 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:06:27.930 17:27:58 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest /dev/nbd0 00:06:27.930 17:27:58 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:06:27.930 17:27:58 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest /dev/nbd1 00:06:27.930 17:27:58 event.app_repeat -- bdev/nbd_common.sh@85 -- # rm /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:06:27.930 17:27:58 event.app_repeat -- bdev/nbd_common.sh@103 -- # nbd_stop_disks /var/tmp/spdk-nbd.sock '/dev/nbd0 /dev/nbd1' 00:06:27.930 17:27:58 event.app_repeat -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:27.930 17:27:58 event.app_repeat -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:27.930 17:27:58 event.app_repeat -- bdev/nbd_common.sh@50 -- # local nbd_list 00:06:27.930 17:27:58 event.app_repeat -- bdev/nbd_common.sh@51 -- # local i 00:06:27.930 17:27:58 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:06:27.930 17:27:58 event.app_repeat -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd0 00:06:28.190 17:27:59 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:06:28.190 17:27:59 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:06:28.190 17:27:59 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:06:28.190 17:27:59 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:06:28.190 17:27:59 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:06:28.190 17:27:59 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:06:28.190 17:27:59 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:06:28.190 17:27:59 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:06:28.190 17:27:59 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:06:28.190 17:27:59 event.app_repeat -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd1 00:06:28.190 17:27:59 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:06:28.190 17:27:59 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:06:28.190 17:27:59 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:06:28.190 17:27:59 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:06:28.190 17:27:59 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:06:28.190 17:27:59 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:06:28.190 17:27:59 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:06:28.190 17:27:59 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:06:28.190 17:27:59 event.app_repeat -- bdev/nbd_common.sh@104 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:06:28.190 17:27:59 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:28.190 17:27:59 event.app_repeat -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:06:28.450 17:27:59 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:06:28.450 17:27:59 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:06:28.450 17:27:59 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[]' 00:06:28.450 17:27:59 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:06:28.450 17:27:59 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '' 00:06:28.450 17:27:59 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:06:28.450 17:27:59 event.app_repeat -- bdev/nbd_common.sh@65 -- # true 00:06:28.450 17:27:59 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=0 00:06:28.450 17:27:59 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 0 00:06:28.450 17:27:59 event.app_repeat -- bdev/nbd_common.sh@104 -- # count=0 00:06:28.450 17:27:59 event.app_repeat -- bdev/nbd_common.sh@105 -- # '[' 0 -ne 0 ']' 00:06:28.450 17:27:59 event.app_repeat -- bdev/nbd_common.sh@109 -- # return 0 00:06:28.450 17:27:59 event.app_repeat -- event/event.sh@34 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock spdk_kill_instance SIGTERM 00:06:28.710 17:27:59 event.app_repeat -- event/event.sh@35 -- # sleep 3 00:06:28.970 [2024-11-27 17:28:00.137983] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 2 00:06:29.230 [2024-11-27 17:28:00.211981] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 1 00:06:29.230 [2024-11-27 17:28:00.211986] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:06:29.230 [2024-11-27 17:28:00.287707] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_register' already registered. 00:06:29.231 [2024-11-27 17:28:00.287772] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_unregister' already registered. 00:06:31.797 spdk_app_start Round 1 00:06:31.797 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock... 00:06:31.797 17:28:02 event.app_repeat -- event/event.sh@23 -- # for i in {0..2} 00:06:31.797 17:28:02 event.app_repeat -- event/event.sh@24 -- # echo 'spdk_app_start Round 1' 00:06:31.797 17:28:02 event.app_repeat -- event/event.sh@25 -- # waitforlisten 70067 /var/tmp/spdk-nbd.sock 00:06:31.797 17:28:02 event.app_repeat -- common/autotest_common.sh@831 -- # '[' -z 70067 ']' 00:06:31.797 17:28:02 event.app_repeat -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk-nbd.sock 00:06:31.797 17:28:02 event.app_repeat -- common/autotest_common.sh@836 -- # local max_retries=100 00:06:31.797 17:28:02 event.app_repeat -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock...' 00:06:31.797 17:28:02 event.app_repeat -- common/autotest_common.sh@840 -- # xtrace_disable 00:06:31.797 17:28:02 event.app_repeat -- common/autotest_common.sh@10 -- # set +x 00:06:32.057 17:28:03 event.app_repeat -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:06:32.057 17:28:03 event.app_repeat -- common/autotest_common.sh@864 -- # return 0 00:06:32.057 17:28:03 event.app_repeat -- event/event.sh@27 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:06:32.057 Malloc0 00:06:32.318 17:28:03 event.app_repeat -- event/event.sh@28 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:06:32.318 Malloc1 00:06:32.318 17:28:03 event.app_repeat -- event/event.sh@30 -- # nbd_rpc_data_verify /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:06:32.318 17:28:03 event.app_repeat -- bdev/nbd_common.sh@90 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:32.318 17:28:03 event.app_repeat -- bdev/nbd_common.sh@91 -- # bdev_list=('Malloc0' 'Malloc1') 00:06:32.318 17:28:03 event.app_repeat -- bdev/nbd_common.sh@91 -- # local bdev_list 00:06:32.318 17:28:03 event.app_repeat -- bdev/nbd_common.sh@92 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:32.318 17:28:03 event.app_repeat -- bdev/nbd_common.sh@92 -- # local nbd_list 00:06:32.318 17:28:03 event.app_repeat -- bdev/nbd_common.sh@94 -- # nbd_start_disks /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:06:32.318 17:28:03 event.app_repeat -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:32.318 17:28:03 event.app_repeat -- bdev/nbd_common.sh@10 -- # bdev_list=('Malloc0' 'Malloc1') 00:06:32.318 17:28:03 event.app_repeat -- bdev/nbd_common.sh@10 -- # local bdev_list 00:06:32.318 17:28:03 event.app_repeat -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:32.318 17:28:03 event.app_repeat -- bdev/nbd_common.sh@11 -- # local nbd_list 00:06:32.318 17:28:03 event.app_repeat -- bdev/nbd_common.sh@12 -- # local i 00:06:32.318 17:28:03 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:06:32.318 17:28:03 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:06:32.318 17:28:03 event.app_repeat -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc0 /dev/nbd0 00:06:32.579 /dev/nbd0 00:06:32.579 17:28:03 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:06:32.579 17:28:03 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:06:32.579 17:28:03 event.app_repeat -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:06:32.579 17:28:03 event.app_repeat -- common/autotest_common.sh@869 -- # local i 00:06:32.579 17:28:03 event.app_repeat -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:06:32.579 17:28:03 event.app_repeat -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:06:32.579 17:28:03 event.app_repeat -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:06:32.579 17:28:03 event.app_repeat -- common/autotest_common.sh@873 -- # break 00:06:32.579 17:28:03 event.app_repeat -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:06:32.579 17:28:03 event.app_repeat -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:06:32.579 17:28:03 event.app_repeat -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:06:32.579 1+0 records in 00:06:32.579 1+0 records out 00:06:32.579 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000444374 s, 9.2 MB/s 00:06:32.579 17:28:03 event.app_repeat -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:06:32.579 17:28:03 event.app_repeat -- common/autotest_common.sh@886 -- # size=4096 00:06:32.579 17:28:03 event.app_repeat -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:06:32.579 17:28:03 event.app_repeat -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:06:32.579 17:28:03 event.app_repeat -- common/autotest_common.sh@889 -- # return 0 00:06:32.579 17:28:03 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:06:32.579 17:28:03 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:06:32.579 17:28:03 event.app_repeat -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc1 /dev/nbd1 00:06:32.840 /dev/nbd1 00:06:32.840 17:28:03 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:06:32.840 17:28:03 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:06:32.840 17:28:03 event.app_repeat -- common/autotest_common.sh@868 -- # local nbd_name=nbd1 00:06:32.840 17:28:03 event.app_repeat -- common/autotest_common.sh@869 -- # local i 00:06:32.840 17:28:03 event.app_repeat -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:06:32.840 17:28:03 event.app_repeat -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:06:32.840 17:28:03 event.app_repeat -- common/autotest_common.sh@872 -- # grep -q -w nbd1 /proc/partitions 00:06:32.840 17:28:03 event.app_repeat -- common/autotest_common.sh@873 -- # break 00:06:32.840 17:28:03 event.app_repeat -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:06:32.840 17:28:03 event.app_repeat -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:06:32.840 17:28:03 event.app_repeat -- common/autotest_common.sh@885 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:06:32.840 1+0 records in 00:06:32.840 1+0 records out 00:06:32.840 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.00046928 s, 8.7 MB/s 00:06:32.840 17:28:03 event.app_repeat -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:06:32.840 17:28:03 event.app_repeat -- common/autotest_common.sh@886 -- # size=4096 00:06:32.840 17:28:03 event.app_repeat -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:06:32.840 17:28:03 event.app_repeat -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:06:32.840 17:28:03 event.app_repeat -- common/autotest_common.sh@889 -- # return 0 00:06:32.840 17:28:03 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:06:32.840 17:28:03 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:06:32.840 17:28:03 event.app_repeat -- bdev/nbd_common.sh@95 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:06:32.840 17:28:03 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:32.840 17:28:03 event.app_repeat -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:06:33.100 17:28:04 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[ 00:06:33.100 { 00:06:33.100 "nbd_device": "/dev/nbd0", 00:06:33.100 "bdev_name": "Malloc0" 00:06:33.100 }, 00:06:33.100 { 00:06:33.100 "nbd_device": "/dev/nbd1", 00:06:33.100 "bdev_name": "Malloc1" 00:06:33.100 } 00:06:33.100 ]' 00:06:33.100 17:28:04 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[ 00:06:33.100 { 00:06:33.100 "nbd_device": "/dev/nbd0", 00:06:33.100 "bdev_name": "Malloc0" 00:06:33.100 }, 00:06:33.100 { 00:06:33.101 "nbd_device": "/dev/nbd1", 00:06:33.101 "bdev_name": "Malloc1" 00:06:33.101 } 00:06:33.101 ]' 00:06:33.101 17:28:04 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:06:33.101 17:28:04 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name='/dev/nbd0 00:06:33.101 /dev/nbd1' 00:06:33.101 17:28:04 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '/dev/nbd0 00:06:33.101 /dev/nbd1' 00:06:33.101 17:28:04 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:06:33.101 17:28:04 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=2 00:06:33.101 17:28:04 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 2 00:06:33.101 17:28:04 event.app_repeat -- bdev/nbd_common.sh@95 -- # count=2 00:06:33.101 17:28:04 event.app_repeat -- bdev/nbd_common.sh@96 -- # '[' 2 -ne 2 ']' 00:06:33.101 17:28:04 event.app_repeat -- bdev/nbd_common.sh@100 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' write 00:06:33.101 17:28:04 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:33.101 17:28:04 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:06:33.101 17:28:04 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=write 00:06:33.101 17:28:04 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:06:33.101 17:28:04 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' write = write ']' 00:06:33.101 17:28:04 event.app_repeat -- bdev/nbd_common.sh@76 -- # dd if=/dev/urandom of=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest bs=4096 count=256 00:06:33.101 256+0 records in 00:06:33.101 256+0 records out 00:06:33.101 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.00560604 s, 187 MB/s 00:06:33.101 17:28:04 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:06:33.101 17:28:04 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest of=/dev/nbd0 bs=4096 count=256 oflag=direct 00:06:33.101 256+0 records in 00:06:33.101 256+0 records out 00:06:33.101 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0232552 s, 45.1 MB/s 00:06:33.101 17:28:04 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:06:33.101 17:28:04 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest of=/dev/nbd1 bs=4096 count=256 oflag=direct 00:06:33.101 256+0 records in 00:06:33.101 256+0 records out 00:06:33.101 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0258146 s, 40.6 MB/s 00:06:33.101 17:28:04 event.app_repeat -- bdev/nbd_common.sh@101 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' verify 00:06:33.101 17:28:04 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:33.101 17:28:04 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:06:33.101 17:28:04 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=verify 00:06:33.101 17:28:04 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:06:33.101 17:28:04 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' verify = write ']' 00:06:33.101 17:28:04 event.app_repeat -- bdev/nbd_common.sh@80 -- # '[' verify = verify ']' 00:06:33.101 17:28:04 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:06:33.101 17:28:04 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest /dev/nbd0 00:06:33.361 17:28:04 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:06:33.361 17:28:04 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest /dev/nbd1 00:06:33.361 17:28:04 event.app_repeat -- bdev/nbd_common.sh@85 -- # rm /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:06:33.361 17:28:04 event.app_repeat -- bdev/nbd_common.sh@103 -- # nbd_stop_disks /var/tmp/spdk-nbd.sock '/dev/nbd0 /dev/nbd1' 00:06:33.361 17:28:04 event.app_repeat -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:33.361 17:28:04 event.app_repeat -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:33.361 17:28:04 event.app_repeat -- bdev/nbd_common.sh@50 -- # local nbd_list 00:06:33.361 17:28:04 event.app_repeat -- bdev/nbd_common.sh@51 -- # local i 00:06:33.361 17:28:04 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:06:33.361 17:28:04 event.app_repeat -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd0 00:06:33.361 17:28:04 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:06:33.361 17:28:04 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:06:33.361 17:28:04 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:06:33.361 17:28:04 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:06:33.361 17:28:04 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:06:33.361 17:28:04 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:06:33.361 17:28:04 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:06:33.361 17:28:04 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:06:33.361 17:28:04 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:06:33.361 17:28:04 event.app_repeat -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd1 00:06:33.622 17:28:04 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:06:33.622 17:28:04 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:06:33.622 17:28:04 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:06:33.622 17:28:04 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:06:33.622 17:28:04 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:06:33.622 17:28:04 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:06:33.622 17:28:04 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:06:33.622 17:28:04 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:06:33.622 17:28:04 event.app_repeat -- bdev/nbd_common.sh@104 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:06:33.622 17:28:04 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:33.622 17:28:04 event.app_repeat -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:06:33.882 17:28:04 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:06:33.882 17:28:04 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[]' 00:06:33.882 17:28:04 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:06:33.882 17:28:04 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:06:33.882 17:28:04 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '' 00:06:33.882 17:28:04 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:06:33.882 17:28:04 event.app_repeat -- bdev/nbd_common.sh@65 -- # true 00:06:33.882 17:28:04 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=0 00:06:33.882 17:28:04 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 0 00:06:33.882 17:28:04 event.app_repeat -- bdev/nbd_common.sh@104 -- # count=0 00:06:33.882 17:28:04 event.app_repeat -- bdev/nbd_common.sh@105 -- # '[' 0 -ne 0 ']' 00:06:33.882 17:28:04 event.app_repeat -- bdev/nbd_common.sh@109 -- # return 0 00:06:33.882 17:28:04 event.app_repeat -- event/event.sh@34 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock spdk_kill_instance SIGTERM 00:06:34.143 17:28:05 event.app_repeat -- event/event.sh@35 -- # sleep 3 00:06:34.403 [2024-11-27 17:28:05.514862] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 2 00:06:34.403 [2024-11-27 17:28:05.576500] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:06:34.403 [2024-11-27 17:28:05.576523] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 1 00:06:34.664 [2024-11-27 17:28:05.652118] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_register' already registered. 00:06:34.664 [2024-11-27 17:28:05.652183] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_unregister' already registered. 00:06:37.204 spdk_app_start Round 2 00:06:37.204 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock... 00:06:37.204 17:28:08 event.app_repeat -- event/event.sh@23 -- # for i in {0..2} 00:06:37.204 17:28:08 event.app_repeat -- event/event.sh@24 -- # echo 'spdk_app_start Round 2' 00:06:37.204 17:28:08 event.app_repeat -- event/event.sh@25 -- # waitforlisten 70067 /var/tmp/spdk-nbd.sock 00:06:37.204 17:28:08 event.app_repeat -- common/autotest_common.sh@831 -- # '[' -z 70067 ']' 00:06:37.204 17:28:08 event.app_repeat -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk-nbd.sock 00:06:37.204 17:28:08 event.app_repeat -- common/autotest_common.sh@836 -- # local max_retries=100 00:06:37.204 17:28:08 event.app_repeat -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock...' 00:06:37.204 17:28:08 event.app_repeat -- common/autotest_common.sh@840 -- # xtrace_disable 00:06:37.204 17:28:08 event.app_repeat -- common/autotest_common.sh@10 -- # set +x 00:06:37.464 17:28:08 event.app_repeat -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:06:37.464 17:28:08 event.app_repeat -- common/autotest_common.sh@864 -- # return 0 00:06:37.464 17:28:08 event.app_repeat -- event/event.sh@27 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:06:37.464 Malloc0 00:06:37.464 17:28:08 event.app_repeat -- event/event.sh@28 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:06:37.724 Malloc1 00:06:37.724 17:28:08 event.app_repeat -- event/event.sh@30 -- # nbd_rpc_data_verify /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:06:37.724 17:28:08 event.app_repeat -- bdev/nbd_common.sh@90 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:37.724 17:28:08 event.app_repeat -- bdev/nbd_common.sh@91 -- # bdev_list=('Malloc0' 'Malloc1') 00:06:37.724 17:28:08 event.app_repeat -- bdev/nbd_common.sh@91 -- # local bdev_list 00:06:37.724 17:28:08 event.app_repeat -- bdev/nbd_common.sh@92 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:37.724 17:28:08 event.app_repeat -- bdev/nbd_common.sh@92 -- # local nbd_list 00:06:37.724 17:28:08 event.app_repeat -- bdev/nbd_common.sh@94 -- # nbd_start_disks /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:06:37.724 17:28:08 event.app_repeat -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:37.724 17:28:08 event.app_repeat -- bdev/nbd_common.sh@10 -- # bdev_list=('Malloc0' 'Malloc1') 00:06:37.724 17:28:08 event.app_repeat -- bdev/nbd_common.sh@10 -- # local bdev_list 00:06:37.724 17:28:08 event.app_repeat -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:37.724 17:28:08 event.app_repeat -- bdev/nbd_common.sh@11 -- # local nbd_list 00:06:37.724 17:28:08 event.app_repeat -- bdev/nbd_common.sh@12 -- # local i 00:06:37.724 17:28:08 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:06:37.724 17:28:08 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:06:37.724 17:28:08 event.app_repeat -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc0 /dev/nbd0 00:06:37.984 /dev/nbd0 00:06:37.984 17:28:09 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:06:37.984 17:28:09 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:06:37.984 17:28:09 event.app_repeat -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:06:37.984 17:28:09 event.app_repeat -- common/autotest_common.sh@869 -- # local i 00:06:37.984 17:28:09 event.app_repeat -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:06:37.984 17:28:09 event.app_repeat -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:06:37.984 17:28:09 event.app_repeat -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:06:37.984 17:28:09 event.app_repeat -- common/autotest_common.sh@873 -- # break 00:06:37.984 17:28:09 event.app_repeat -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:06:37.984 17:28:09 event.app_repeat -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:06:37.984 17:28:09 event.app_repeat -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:06:37.984 1+0 records in 00:06:37.984 1+0 records out 00:06:37.984 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000380692 s, 10.8 MB/s 00:06:37.984 17:28:09 event.app_repeat -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:06:37.984 17:28:09 event.app_repeat -- common/autotest_common.sh@886 -- # size=4096 00:06:37.984 17:28:09 event.app_repeat -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:06:37.984 17:28:09 event.app_repeat -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:06:37.984 17:28:09 event.app_repeat -- common/autotest_common.sh@889 -- # return 0 00:06:37.984 17:28:09 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:06:37.984 17:28:09 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:06:37.984 17:28:09 event.app_repeat -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc1 /dev/nbd1 00:06:38.243 /dev/nbd1 00:06:38.243 17:28:09 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:06:38.243 17:28:09 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:06:38.243 17:28:09 event.app_repeat -- common/autotest_common.sh@868 -- # local nbd_name=nbd1 00:06:38.243 17:28:09 event.app_repeat -- common/autotest_common.sh@869 -- # local i 00:06:38.243 17:28:09 event.app_repeat -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:06:38.243 17:28:09 event.app_repeat -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:06:38.243 17:28:09 event.app_repeat -- common/autotest_common.sh@872 -- # grep -q -w nbd1 /proc/partitions 00:06:38.243 17:28:09 event.app_repeat -- common/autotest_common.sh@873 -- # break 00:06:38.243 17:28:09 event.app_repeat -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:06:38.243 17:28:09 event.app_repeat -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:06:38.243 17:28:09 event.app_repeat -- common/autotest_common.sh@885 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:06:38.243 1+0 records in 00:06:38.243 1+0 records out 00:06:38.243 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.00035749 s, 11.5 MB/s 00:06:38.243 17:28:09 event.app_repeat -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:06:38.243 17:28:09 event.app_repeat -- common/autotest_common.sh@886 -- # size=4096 00:06:38.243 17:28:09 event.app_repeat -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/event/nbdtest 00:06:38.243 17:28:09 event.app_repeat -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:06:38.243 17:28:09 event.app_repeat -- common/autotest_common.sh@889 -- # return 0 00:06:38.243 17:28:09 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:06:38.243 17:28:09 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:06:38.243 17:28:09 event.app_repeat -- bdev/nbd_common.sh@95 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:06:38.243 17:28:09 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:38.243 17:28:09 event.app_repeat -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:06:38.501 17:28:09 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[ 00:06:38.501 { 00:06:38.501 "nbd_device": "/dev/nbd0", 00:06:38.501 "bdev_name": "Malloc0" 00:06:38.501 }, 00:06:38.501 { 00:06:38.501 "nbd_device": "/dev/nbd1", 00:06:38.501 "bdev_name": "Malloc1" 00:06:38.501 } 00:06:38.501 ]' 00:06:38.501 17:28:09 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[ 00:06:38.501 { 00:06:38.502 "nbd_device": "/dev/nbd0", 00:06:38.502 "bdev_name": "Malloc0" 00:06:38.502 }, 00:06:38.502 { 00:06:38.502 "nbd_device": "/dev/nbd1", 00:06:38.502 "bdev_name": "Malloc1" 00:06:38.502 } 00:06:38.502 ]' 00:06:38.502 17:28:09 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:06:38.502 17:28:09 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name='/dev/nbd0 00:06:38.502 /dev/nbd1' 00:06:38.502 17:28:09 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:06:38.502 17:28:09 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '/dev/nbd0 00:06:38.502 /dev/nbd1' 00:06:38.502 17:28:09 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=2 00:06:38.502 17:28:09 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 2 00:06:38.502 17:28:09 event.app_repeat -- bdev/nbd_common.sh@95 -- # count=2 00:06:38.502 17:28:09 event.app_repeat -- bdev/nbd_common.sh@96 -- # '[' 2 -ne 2 ']' 00:06:38.502 17:28:09 event.app_repeat -- bdev/nbd_common.sh@100 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' write 00:06:38.502 17:28:09 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:38.502 17:28:09 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:06:38.502 17:28:09 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=write 00:06:38.502 17:28:09 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:06:38.502 17:28:09 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' write = write ']' 00:06:38.502 17:28:09 event.app_repeat -- bdev/nbd_common.sh@76 -- # dd if=/dev/urandom of=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest bs=4096 count=256 00:06:38.502 256+0 records in 00:06:38.502 256+0 records out 00:06:38.502 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0129475 s, 81.0 MB/s 00:06:38.502 17:28:09 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:06:38.502 17:28:09 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest of=/dev/nbd0 bs=4096 count=256 oflag=direct 00:06:38.502 256+0 records in 00:06:38.502 256+0 records out 00:06:38.502 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.019468 s, 53.9 MB/s 00:06:38.502 17:28:09 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:06:38.502 17:28:09 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest of=/dev/nbd1 bs=4096 count=256 oflag=direct 00:06:38.762 256+0 records in 00:06:38.762 256+0 records out 00:06:38.762 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0203015 s, 51.7 MB/s 00:06:38.762 17:28:09 event.app_repeat -- bdev/nbd_common.sh@101 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' verify 00:06:38.762 17:28:09 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:38.762 17:28:09 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:06:38.762 17:28:09 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=verify 00:06:38.762 17:28:09 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:06:38.762 17:28:09 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' verify = write ']' 00:06:38.762 17:28:09 event.app_repeat -- bdev/nbd_common.sh@80 -- # '[' verify = verify ']' 00:06:38.762 17:28:09 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:06:38.762 17:28:09 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest /dev/nbd0 00:06:38.762 17:28:09 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:06:38.762 17:28:09 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest /dev/nbd1 00:06:38.762 17:28:09 event.app_repeat -- bdev/nbd_common.sh@85 -- # rm /home/vagrant/spdk_repo/spdk/test/event/nbdrandtest 00:06:38.762 17:28:09 event.app_repeat -- bdev/nbd_common.sh@103 -- # nbd_stop_disks /var/tmp/spdk-nbd.sock '/dev/nbd0 /dev/nbd1' 00:06:38.762 17:28:09 event.app_repeat -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:38.762 17:28:09 event.app_repeat -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:06:38.762 17:28:09 event.app_repeat -- bdev/nbd_common.sh@50 -- # local nbd_list 00:06:38.762 17:28:09 event.app_repeat -- bdev/nbd_common.sh@51 -- # local i 00:06:38.762 17:28:09 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:06:38.762 17:28:09 event.app_repeat -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd0 00:06:39.022 17:28:09 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:06:39.022 17:28:09 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:06:39.022 17:28:09 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:06:39.022 17:28:09 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:06:39.022 17:28:09 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:06:39.022 17:28:09 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:06:39.022 17:28:09 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:06:39.022 17:28:09 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:06:39.022 17:28:09 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:06:39.022 17:28:09 event.app_repeat -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd1 00:06:39.022 17:28:10 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:06:39.022 17:28:10 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:06:39.022 17:28:10 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:06:39.022 17:28:10 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:06:39.022 17:28:10 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:06:39.022 17:28:10 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:06:39.022 17:28:10 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:06:39.022 17:28:10 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:06:39.022 17:28:10 event.app_repeat -- bdev/nbd_common.sh@104 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:06:39.022 17:28:10 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:06:39.022 17:28:10 event.app_repeat -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:06:39.282 17:28:10 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:06:39.282 17:28:10 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:06:39.282 17:28:10 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[]' 00:06:39.282 17:28:10 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:06:39.282 17:28:10 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '' 00:06:39.282 17:28:10 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:06:39.282 17:28:10 event.app_repeat -- bdev/nbd_common.sh@65 -- # true 00:06:39.282 17:28:10 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=0 00:06:39.282 17:28:10 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 0 00:06:39.282 17:28:10 event.app_repeat -- bdev/nbd_common.sh@104 -- # count=0 00:06:39.282 17:28:10 event.app_repeat -- bdev/nbd_common.sh@105 -- # '[' 0 -ne 0 ']' 00:06:39.282 17:28:10 event.app_repeat -- bdev/nbd_common.sh@109 -- # return 0 00:06:39.282 17:28:10 event.app_repeat -- event/event.sh@34 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock spdk_kill_instance SIGTERM 00:06:39.543 17:28:10 event.app_repeat -- event/event.sh@35 -- # sleep 3 00:06:39.803 [2024-11-27 17:28:10.985827] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 2 00:06:40.063 [2024-11-27 17:28:11.048013] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:06:40.063 [2024-11-27 17:28:11.048019] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 1 00:06:40.063 [2024-11-27 17:28:11.123474] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_register' already registered. 00:06:40.063 [2024-11-27 17:28:11.123542] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_unregister' already registered. 00:06:42.603 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock... 00:06:42.603 17:28:13 event.app_repeat -- event/event.sh@38 -- # waitforlisten 70067 /var/tmp/spdk-nbd.sock 00:06:42.603 17:28:13 event.app_repeat -- common/autotest_common.sh@831 -- # '[' -z 70067 ']' 00:06:42.603 17:28:13 event.app_repeat -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk-nbd.sock 00:06:42.603 17:28:13 event.app_repeat -- common/autotest_common.sh@836 -- # local max_retries=100 00:06:42.603 17:28:13 event.app_repeat -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock...' 00:06:42.603 17:28:13 event.app_repeat -- common/autotest_common.sh@840 -- # xtrace_disable 00:06:42.603 17:28:13 event.app_repeat -- common/autotest_common.sh@10 -- # set +x 00:06:42.863 17:28:13 event.app_repeat -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:06:42.863 17:28:13 event.app_repeat -- common/autotest_common.sh@864 -- # return 0 00:06:42.863 17:28:13 event.app_repeat -- event/event.sh@39 -- # killprocess 70067 00:06:42.863 17:28:13 event.app_repeat -- common/autotest_common.sh@950 -- # '[' -z 70067 ']' 00:06:42.863 17:28:13 event.app_repeat -- common/autotest_common.sh@954 -- # kill -0 70067 00:06:42.863 17:28:13 event.app_repeat -- common/autotest_common.sh@955 -- # uname 00:06:42.863 17:28:13 event.app_repeat -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:06:42.863 17:28:13 event.app_repeat -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 70067 00:06:42.863 killing process with pid 70067 00:06:42.863 17:28:13 event.app_repeat -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:06:42.863 17:28:13 event.app_repeat -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:06:42.863 17:28:13 event.app_repeat -- common/autotest_common.sh@968 -- # echo 'killing process with pid 70067' 00:06:42.863 17:28:13 event.app_repeat -- common/autotest_common.sh@969 -- # kill 70067 00:06:42.863 17:28:13 event.app_repeat -- common/autotest_common.sh@974 -- # wait 70067 00:06:43.123 spdk_app_start is called in Round 0. 00:06:43.123 Shutdown signal received, stop current app iteration 00:06:43.123 Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 reinitialization... 00:06:43.123 spdk_app_start is called in Round 1. 00:06:43.123 Shutdown signal received, stop current app iteration 00:06:43.123 Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 reinitialization... 00:06:43.123 spdk_app_start is called in Round 2. 00:06:43.123 Shutdown signal received, stop current app iteration 00:06:43.123 Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 reinitialization... 00:06:43.123 spdk_app_start is called in Round 3. 00:06:43.123 Shutdown signal received, stop current app iteration 00:06:43.123 17:28:14 event.app_repeat -- event/event.sh@40 -- # trap - SIGINT SIGTERM EXIT 00:06:43.123 17:28:14 event.app_repeat -- event/event.sh@42 -- # return 0 00:06:43.123 00:06:43.123 real 0m17.552s 00:06:43.123 user 0m37.928s 00:06:43.123 sys 0m2.948s 00:06:43.123 17:28:14 event.app_repeat -- common/autotest_common.sh@1126 -- # xtrace_disable 00:06:43.123 ************************************ 00:06:43.123 END TEST app_repeat 00:06:43.123 ************************************ 00:06:43.123 17:28:14 event.app_repeat -- common/autotest_common.sh@10 -- # set +x 00:06:43.393 17:28:14 event -- event/event.sh@54 -- # (( SPDK_TEST_CRYPTO == 0 )) 00:06:43.393 17:28:14 event -- event/event.sh@55 -- # run_test cpu_locks /home/vagrant/spdk_repo/spdk/test/event/cpu_locks.sh 00:06:43.393 17:28:14 event -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:06:43.393 17:28:14 event -- common/autotest_common.sh@1107 -- # xtrace_disable 00:06:43.393 17:28:14 event -- common/autotest_common.sh@10 -- # set +x 00:06:43.393 ************************************ 00:06:43.393 START TEST cpu_locks 00:06:43.393 ************************************ 00:06:43.393 17:28:14 event.cpu_locks -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/event/cpu_locks.sh 00:06:43.393 * Looking for test storage... 00:06:43.393 * Found test storage at /home/vagrant/spdk_repo/spdk/test/event 00:06:43.393 17:28:14 event.cpu_locks -- common/autotest_common.sh@1680 -- # [[ y == y ]] 00:06:43.393 17:28:14 event.cpu_locks -- common/autotest_common.sh@1681 -- # lcov --version 00:06:43.393 17:28:14 event.cpu_locks -- common/autotest_common.sh@1681 -- # awk '{print $NF}' 00:06:43.393 17:28:14 event.cpu_locks -- common/autotest_common.sh@1681 -- # lt 1.15 2 00:06:43.393 17:28:14 event.cpu_locks -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:06:43.393 17:28:14 event.cpu_locks -- scripts/common.sh@333 -- # local ver1 ver1_l 00:06:43.393 17:28:14 event.cpu_locks -- scripts/common.sh@334 -- # local ver2 ver2_l 00:06:43.393 17:28:14 event.cpu_locks -- scripts/common.sh@336 -- # IFS=.-: 00:06:43.393 17:28:14 event.cpu_locks -- scripts/common.sh@336 -- # read -ra ver1 00:06:43.393 17:28:14 event.cpu_locks -- scripts/common.sh@337 -- # IFS=.-: 00:06:43.393 17:28:14 event.cpu_locks -- scripts/common.sh@337 -- # read -ra ver2 00:06:43.393 17:28:14 event.cpu_locks -- scripts/common.sh@338 -- # local 'op=<' 00:06:43.393 17:28:14 event.cpu_locks -- scripts/common.sh@340 -- # ver1_l=2 00:06:43.393 17:28:14 event.cpu_locks -- scripts/common.sh@341 -- # ver2_l=1 00:06:43.393 17:28:14 event.cpu_locks -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:06:43.393 17:28:14 event.cpu_locks -- scripts/common.sh@344 -- # case "$op" in 00:06:43.393 17:28:14 event.cpu_locks -- scripts/common.sh@345 -- # : 1 00:06:43.393 17:28:14 event.cpu_locks -- scripts/common.sh@364 -- # (( v = 0 )) 00:06:43.393 17:28:14 event.cpu_locks -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:06:43.393 17:28:14 event.cpu_locks -- scripts/common.sh@365 -- # decimal 1 00:06:43.393 17:28:14 event.cpu_locks -- scripts/common.sh@353 -- # local d=1 00:06:43.393 17:28:14 event.cpu_locks -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:06:43.393 17:28:14 event.cpu_locks -- scripts/common.sh@355 -- # echo 1 00:06:43.393 17:28:14 event.cpu_locks -- scripts/common.sh@365 -- # ver1[v]=1 00:06:43.393 17:28:14 event.cpu_locks -- scripts/common.sh@366 -- # decimal 2 00:06:43.393 17:28:14 event.cpu_locks -- scripts/common.sh@353 -- # local d=2 00:06:43.393 17:28:14 event.cpu_locks -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:06:43.393 17:28:14 event.cpu_locks -- scripts/common.sh@355 -- # echo 2 00:06:43.393 17:28:14 event.cpu_locks -- scripts/common.sh@366 -- # ver2[v]=2 00:06:43.393 17:28:14 event.cpu_locks -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:06:43.393 17:28:14 event.cpu_locks -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:06:43.670 17:28:14 event.cpu_locks -- scripts/common.sh@368 -- # return 0 00:06:43.670 17:28:14 event.cpu_locks -- common/autotest_common.sh@1682 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:06:43.670 17:28:14 event.cpu_locks -- common/autotest_common.sh@1694 -- # export 'LCOV_OPTS= 00:06:43.670 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:43.670 --rc genhtml_branch_coverage=1 00:06:43.670 --rc genhtml_function_coverage=1 00:06:43.670 --rc genhtml_legend=1 00:06:43.670 --rc geninfo_all_blocks=1 00:06:43.670 --rc geninfo_unexecuted_blocks=1 00:06:43.670 00:06:43.670 ' 00:06:43.670 17:28:14 event.cpu_locks -- common/autotest_common.sh@1694 -- # LCOV_OPTS=' 00:06:43.670 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:43.670 --rc genhtml_branch_coverage=1 00:06:43.670 --rc genhtml_function_coverage=1 00:06:43.670 --rc genhtml_legend=1 00:06:43.670 --rc geninfo_all_blocks=1 00:06:43.670 --rc geninfo_unexecuted_blocks=1 00:06:43.670 00:06:43.670 ' 00:06:43.670 17:28:14 event.cpu_locks -- common/autotest_common.sh@1695 -- # export 'LCOV=lcov 00:06:43.670 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:43.670 --rc genhtml_branch_coverage=1 00:06:43.670 --rc genhtml_function_coverage=1 00:06:43.670 --rc genhtml_legend=1 00:06:43.670 --rc geninfo_all_blocks=1 00:06:43.670 --rc geninfo_unexecuted_blocks=1 00:06:43.670 00:06:43.670 ' 00:06:43.670 17:28:14 event.cpu_locks -- common/autotest_common.sh@1695 -- # LCOV='lcov 00:06:43.670 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:43.670 --rc genhtml_branch_coverage=1 00:06:43.670 --rc genhtml_function_coverage=1 00:06:43.670 --rc genhtml_legend=1 00:06:43.670 --rc geninfo_all_blocks=1 00:06:43.670 --rc geninfo_unexecuted_blocks=1 00:06:43.670 00:06:43.670 ' 00:06:43.670 17:28:14 event.cpu_locks -- event/cpu_locks.sh@11 -- # rpc_sock1=/var/tmp/spdk.sock 00:06:43.670 17:28:14 event.cpu_locks -- event/cpu_locks.sh@12 -- # rpc_sock2=/var/tmp/spdk2.sock 00:06:43.670 17:28:14 event.cpu_locks -- event/cpu_locks.sh@164 -- # trap cleanup EXIT SIGTERM SIGINT 00:06:43.670 17:28:14 event.cpu_locks -- event/cpu_locks.sh@166 -- # run_test default_locks default_locks 00:06:43.670 17:28:14 event.cpu_locks -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:06:43.670 17:28:14 event.cpu_locks -- common/autotest_common.sh@1107 -- # xtrace_disable 00:06:43.670 17:28:14 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:06:43.670 ************************************ 00:06:43.670 START TEST default_locks 00:06:43.670 ************************************ 00:06:43.670 17:28:14 event.cpu_locks.default_locks -- common/autotest_common.sh@1125 -- # default_locks 00:06:43.670 17:28:14 event.cpu_locks.default_locks -- event/cpu_locks.sh@46 -- # spdk_tgt_pid=70498 00:06:43.670 17:28:14 event.cpu_locks.default_locks -- event/cpu_locks.sh@47 -- # waitforlisten 70498 00:06:43.670 17:28:14 event.cpu_locks.default_locks -- event/cpu_locks.sh@45 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 00:06:43.670 17:28:14 event.cpu_locks.default_locks -- common/autotest_common.sh@831 -- # '[' -z 70498 ']' 00:06:43.670 17:28:14 event.cpu_locks.default_locks -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:43.670 17:28:14 event.cpu_locks.default_locks -- common/autotest_common.sh@836 -- # local max_retries=100 00:06:43.670 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:43.670 17:28:14 event.cpu_locks.default_locks -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:43.670 17:28:14 event.cpu_locks.default_locks -- common/autotest_common.sh@840 -- # xtrace_disable 00:06:43.670 17:28:14 event.cpu_locks.default_locks -- common/autotest_common.sh@10 -- # set +x 00:06:43.670 [2024-11-27 17:28:14.690930] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:06:43.670 [2024-11-27 17:28:14.691062] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70498 ] 00:06:43.670 [2024-11-27 17:28:14.835920] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:43.930 [2024-11-27 17:28:14.904856] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:06:44.499 17:28:15 event.cpu_locks.default_locks -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:06:44.499 17:28:15 event.cpu_locks.default_locks -- common/autotest_common.sh@864 -- # return 0 00:06:44.499 17:28:15 event.cpu_locks.default_locks -- event/cpu_locks.sh@49 -- # locks_exist 70498 00:06:44.499 17:28:15 event.cpu_locks.default_locks -- event/cpu_locks.sh@22 -- # lslocks -p 70498 00:06:44.499 17:28:15 event.cpu_locks.default_locks -- event/cpu_locks.sh@22 -- # grep -q spdk_cpu_lock 00:06:44.758 17:28:15 event.cpu_locks.default_locks -- event/cpu_locks.sh@50 -- # killprocess 70498 00:06:44.758 17:28:15 event.cpu_locks.default_locks -- common/autotest_common.sh@950 -- # '[' -z 70498 ']' 00:06:44.758 17:28:15 event.cpu_locks.default_locks -- common/autotest_common.sh@954 -- # kill -0 70498 00:06:44.758 17:28:15 event.cpu_locks.default_locks -- common/autotest_common.sh@955 -- # uname 00:06:44.758 17:28:15 event.cpu_locks.default_locks -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:06:44.758 17:28:15 event.cpu_locks.default_locks -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 70498 00:06:44.758 17:28:15 event.cpu_locks.default_locks -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:06:44.758 17:28:15 event.cpu_locks.default_locks -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:06:44.758 killing process with pid 70498 00:06:44.758 17:28:15 event.cpu_locks.default_locks -- common/autotest_common.sh@968 -- # echo 'killing process with pid 70498' 00:06:44.758 17:28:15 event.cpu_locks.default_locks -- common/autotest_common.sh@969 -- # kill 70498 00:06:44.758 17:28:15 event.cpu_locks.default_locks -- common/autotest_common.sh@974 -- # wait 70498 00:06:45.699 17:28:16 event.cpu_locks.default_locks -- event/cpu_locks.sh@52 -- # NOT waitforlisten 70498 00:06:45.699 17:28:16 event.cpu_locks.default_locks -- common/autotest_common.sh@650 -- # local es=0 00:06:45.699 17:28:16 event.cpu_locks.default_locks -- common/autotest_common.sh@652 -- # valid_exec_arg waitforlisten 70498 00:06:45.699 17:28:16 event.cpu_locks.default_locks -- common/autotest_common.sh@638 -- # local arg=waitforlisten 00:06:45.699 17:28:16 event.cpu_locks.default_locks -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:06:45.699 17:28:16 event.cpu_locks.default_locks -- common/autotest_common.sh@642 -- # type -t waitforlisten 00:06:45.699 17:28:16 event.cpu_locks.default_locks -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:06:45.699 17:28:16 event.cpu_locks.default_locks -- common/autotest_common.sh@653 -- # waitforlisten 70498 00:06:45.699 17:28:16 event.cpu_locks.default_locks -- common/autotest_common.sh@831 -- # '[' -z 70498 ']' 00:06:45.699 17:28:16 event.cpu_locks.default_locks -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:45.699 17:28:16 event.cpu_locks.default_locks -- common/autotest_common.sh@836 -- # local max_retries=100 00:06:45.699 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:45.699 17:28:16 event.cpu_locks.default_locks -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:45.699 17:28:16 event.cpu_locks.default_locks -- common/autotest_common.sh@840 -- # xtrace_disable 00:06:45.699 17:28:16 event.cpu_locks.default_locks -- common/autotest_common.sh@10 -- # set +x 00:06:45.699 /home/vagrant/spdk_repo/spdk/test/common/autotest_common.sh: line 846: kill: (70498) - No such process 00:06:45.699 ERROR: process (pid: 70498) is no longer running 00:06:45.699 17:28:16 event.cpu_locks.default_locks -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:06:45.699 17:28:16 event.cpu_locks.default_locks -- common/autotest_common.sh@864 -- # return 1 00:06:45.699 17:28:16 event.cpu_locks.default_locks -- common/autotest_common.sh@653 -- # es=1 00:06:45.699 17:28:16 event.cpu_locks.default_locks -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:06:45.699 17:28:16 event.cpu_locks.default_locks -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:06:45.699 17:28:16 event.cpu_locks.default_locks -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:06:45.699 17:28:16 event.cpu_locks.default_locks -- event/cpu_locks.sh@54 -- # no_locks 00:06:45.699 17:28:16 event.cpu_locks.default_locks -- event/cpu_locks.sh@26 -- # lock_files=() 00:06:45.699 17:28:16 event.cpu_locks.default_locks -- event/cpu_locks.sh@26 -- # local lock_files 00:06:45.699 17:28:16 event.cpu_locks.default_locks -- event/cpu_locks.sh@27 -- # (( 0 != 0 )) 00:06:45.699 00:06:45.699 real 0m1.985s 00:06:45.699 user 0m1.790s 00:06:45.699 sys 0m0.759s 00:06:45.699 17:28:16 event.cpu_locks.default_locks -- common/autotest_common.sh@1126 -- # xtrace_disable 00:06:45.699 17:28:16 event.cpu_locks.default_locks -- common/autotest_common.sh@10 -- # set +x 00:06:45.699 ************************************ 00:06:45.699 END TEST default_locks 00:06:45.699 ************************************ 00:06:45.699 17:28:16 event.cpu_locks -- event/cpu_locks.sh@167 -- # run_test default_locks_via_rpc default_locks_via_rpc 00:06:45.699 17:28:16 event.cpu_locks -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:06:45.699 17:28:16 event.cpu_locks -- common/autotest_common.sh@1107 -- # xtrace_disable 00:06:45.699 17:28:16 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:06:45.699 ************************************ 00:06:45.699 START TEST default_locks_via_rpc 00:06:45.699 ************************************ 00:06:45.699 17:28:16 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@1125 -- # default_locks_via_rpc 00:06:45.699 17:28:16 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@62 -- # spdk_tgt_pid=70551 00:06:45.699 17:28:16 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@61 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 00:06:45.699 17:28:16 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@63 -- # waitforlisten 70551 00:06:45.699 17:28:16 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@831 -- # '[' -z 70551 ']' 00:06:45.699 17:28:16 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:45.699 17:28:16 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@836 -- # local max_retries=100 00:06:45.699 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:45.699 17:28:16 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:45.699 17:28:16 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@840 -- # xtrace_disable 00:06:45.699 17:28:16 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:45.700 [2024-11-27 17:28:16.753971] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:06:45.700 [2024-11-27 17:28:16.754119] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70551 ] 00:06:45.960 [2024-11-27 17:28:16.899415] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:45.960 [2024-11-27 17:28:16.969237] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:06:46.530 17:28:17 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:06:46.530 17:28:17 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@864 -- # return 0 00:06:46.530 17:28:17 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@65 -- # rpc_cmd framework_disable_cpumask_locks 00:06:46.530 17:28:17 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:46.530 17:28:17 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:46.530 17:28:17 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:46.530 17:28:17 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@67 -- # no_locks 00:06:46.530 17:28:17 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@26 -- # lock_files=() 00:06:46.530 17:28:17 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@26 -- # local lock_files 00:06:46.530 17:28:17 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@27 -- # (( 0 != 0 )) 00:06:46.530 17:28:17 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@69 -- # rpc_cmd framework_enable_cpumask_locks 00:06:46.530 17:28:17 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:06:46.530 17:28:17 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:46.530 17:28:17 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:06:46.530 17:28:17 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@71 -- # locks_exist 70551 00:06:46.530 17:28:17 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@22 -- # lslocks -p 70551 00:06:46.530 17:28:17 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@22 -- # grep -q spdk_cpu_lock 00:06:46.790 17:28:17 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@73 -- # killprocess 70551 00:06:46.790 17:28:17 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@950 -- # '[' -z 70551 ']' 00:06:46.790 17:28:17 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@954 -- # kill -0 70551 00:06:46.790 17:28:17 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@955 -- # uname 00:06:46.790 17:28:17 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:06:46.790 17:28:17 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 70551 00:06:46.790 17:28:17 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:06:46.790 17:28:17 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:06:46.790 17:28:17 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@968 -- # echo 'killing process with pid 70551' 00:06:46.790 killing process with pid 70551 00:06:46.790 17:28:17 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@969 -- # kill 70551 00:06:46.790 17:28:17 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@974 -- # wait 70551 00:06:47.732 00:06:47.732 real 0m1.956s 00:06:47.732 user 0m1.778s 00:06:47.732 sys 0m0.744s 00:06:47.732 17:28:18 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@1126 -- # xtrace_disable 00:06:47.732 17:28:18 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:47.732 ************************************ 00:06:47.732 END TEST default_locks_via_rpc 00:06:47.732 ************************************ 00:06:47.732 17:28:18 event.cpu_locks -- event/cpu_locks.sh@168 -- # run_test non_locking_app_on_locked_coremask non_locking_app_on_locked_coremask 00:06:47.732 17:28:18 event.cpu_locks -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:06:47.732 17:28:18 event.cpu_locks -- common/autotest_common.sh@1107 -- # xtrace_disable 00:06:47.732 17:28:18 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:06:47.732 ************************************ 00:06:47.732 START TEST non_locking_app_on_locked_coremask 00:06:47.732 ************************************ 00:06:47.732 17:28:18 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@1125 -- # non_locking_app_on_locked_coremask 00:06:47.732 17:28:18 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@80 -- # spdk_tgt_pid=70603 00:06:47.732 17:28:18 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@79 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 00:06:47.732 17:28:18 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@81 -- # waitforlisten 70603 /var/tmp/spdk.sock 00:06:47.732 17:28:18 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@831 -- # '[' -z 70603 ']' 00:06:47.732 17:28:18 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:47.732 17:28:18 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@836 -- # local max_retries=100 00:06:47.732 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:47.732 17:28:18 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:47.732 17:28:18 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@840 -- # xtrace_disable 00:06:47.732 17:28:18 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@10 -- # set +x 00:06:47.732 [2024-11-27 17:28:18.780291] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:06:47.732 [2024-11-27 17:28:18.780426] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70603 ] 00:06:47.992 [2024-11-27 17:28:18.925880] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:47.992 [2024-11-27 17:28:18.995134] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:06:48.560 17:28:19 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:06:48.560 17:28:19 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@864 -- # return 0 00:06:48.560 17:28:19 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@83 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 --disable-cpumask-locks -r /var/tmp/spdk2.sock 00:06:48.560 17:28:19 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@84 -- # spdk_tgt_pid2=70619 00:06:48.560 17:28:19 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@85 -- # waitforlisten 70619 /var/tmp/spdk2.sock 00:06:48.560 17:28:19 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@831 -- # '[' -z 70619 ']' 00:06:48.560 17:28:19 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk2.sock 00:06:48.561 17:28:19 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@836 -- # local max_retries=100 00:06:48.561 17:28:19 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:06:48.561 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:06:48.561 17:28:19 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@840 -- # xtrace_disable 00:06:48.561 17:28:19 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@10 -- # set +x 00:06:48.561 [2024-11-27 17:28:19.680116] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:06:48.561 [2024-11-27 17:28:19.680249] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70619 ] 00:06:48.821 [2024-11-27 17:28:19.813887] app.c: 914:spdk_app_start: *NOTICE*: CPU core locks deactivated. 00:06:48.821 [2024-11-27 17:28:19.813947] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:48.821 [2024-11-27 17:28:19.954013] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:06:49.760 17:28:20 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:06:49.760 17:28:20 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@864 -- # return 0 00:06:49.760 17:28:20 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@87 -- # locks_exist 70603 00:06:49.760 17:28:20 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@22 -- # lslocks -p 70603 00:06:49.760 17:28:20 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@22 -- # grep -q spdk_cpu_lock 00:06:50.700 17:28:21 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@89 -- # killprocess 70603 00:06:50.701 17:28:21 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@950 -- # '[' -z 70603 ']' 00:06:50.701 17:28:21 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@954 -- # kill -0 70603 00:06:50.701 17:28:21 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@955 -- # uname 00:06:50.701 17:28:21 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:06:50.701 17:28:21 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 70603 00:06:50.701 17:28:21 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:06:50.701 17:28:21 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:06:50.701 killing process with pid 70603 00:06:50.701 17:28:21 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@968 -- # echo 'killing process with pid 70603' 00:06:50.701 17:28:21 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@969 -- # kill 70603 00:06:50.701 17:28:21 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@974 -- # wait 70603 00:06:51.640 17:28:22 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@90 -- # killprocess 70619 00:06:51.641 17:28:22 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@950 -- # '[' -z 70619 ']' 00:06:51.641 17:28:22 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@954 -- # kill -0 70619 00:06:51.641 17:28:22 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@955 -- # uname 00:06:51.901 17:28:22 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:06:51.901 17:28:22 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 70619 00:06:51.901 17:28:22 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:06:51.901 17:28:22 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:06:51.901 killing process with pid 70619 00:06:51.901 17:28:22 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@968 -- # echo 'killing process with pid 70619' 00:06:51.901 17:28:22 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@969 -- # kill 70619 00:06:51.901 17:28:22 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@974 -- # wait 70619 00:06:52.471 00:06:52.471 real 0m4.831s 00:06:52.471 user 0m4.683s 00:06:52.471 sys 0m1.598s 00:06:52.471 17:28:23 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@1126 -- # xtrace_disable 00:06:52.471 17:28:23 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@10 -- # set +x 00:06:52.471 ************************************ 00:06:52.471 END TEST non_locking_app_on_locked_coremask 00:06:52.471 ************************************ 00:06:52.471 17:28:23 event.cpu_locks -- event/cpu_locks.sh@169 -- # run_test locking_app_on_unlocked_coremask locking_app_on_unlocked_coremask 00:06:52.471 17:28:23 event.cpu_locks -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:06:52.471 17:28:23 event.cpu_locks -- common/autotest_common.sh@1107 -- # xtrace_disable 00:06:52.471 17:28:23 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:06:52.471 ************************************ 00:06:52.471 START TEST locking_app_on_unlocked_coremask 00:06:52.471 ************************************ 00:06:52.471 17:28:23 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@1125 -- # locking_app_on_unlocked_coremask 00:06:52.471 17:28:23 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@98 -- # spdk_tgt_pid=70695 00:06:52.471 17:28:23 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@97 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 --disable-cpumask-locks 00:06:52.471 17:28:23 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@99 -- # waitforlisten 70695 /var/tmp/spdk.sock 00:06:52.471 17:28:23 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@831 -- # '[' -z 70695 ']' 00:06:52.471 17:28:23 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:52.471 17:28:23 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@836 -- # local max_retries=100 00:06:52.471 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:52.471 17:28:23 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:52.471 17:28:23 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@840 -- # xtrace_disable 00:06:52.471 17:28:23 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@10 -- # set +x 00:06:52.731 [2024-11-27 17:28:23.673725] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:06:52.731 [2024-11-27 17:28:23.674297] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70695 ] 00:06:52.731 [2024-11-27 17:28:23.817390] app.c: 914:spdk_app_start: *NOTICE*: CPU core locks deactivated. 00:06:52.731 [2024-11-27 17:28:23.817457] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:52.731 [2024-11-27 17:28:23.884124] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:06:53.305 17:28:24 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:06:53.305 17:28:24 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@864 -- # return 0 00:06:53.305 17:28:24 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@102 -- # spdk_tgt_pid2=70706 00:06:53.305 17:28:24 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@101 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 -r /var/tmp/spdk2.sock 00:06:53.305 17:28:24 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@103 -- # waitforlisten 70706 /var/tmp/spdk2.sock 00:06:53.305 17:28:24 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@831 -- # '[' -z 70706 ']' 00:06:53.305 17:28:24 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk2.sock 00:06:53.305 17:28:24 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@836 -- # local max_retries=100 00:06:53.305 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:06:53.305 17:28:24 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:06:53.305 17:28:24 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@840 -- # xtrace_disable 00:06:53.306 17:28:24 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@10 -- # set +x 00:06:53.565 [2024-11-27 17:28:24.568844] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:06:53.565 [2024-11-27 17:28:24.568970] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70706 ] 00:06:53.565 [2024-11-27 17:28:24.702889] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:53.825 [2024-11-27 17:28:24.843659] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:06:54.396 17:28:25 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:06:54.396 17:28:25 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@864 -- # return 0 00:06:54.396 17:28:25 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@105 -- # locks_exist 70706 00:06:54.396 17:28:25 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@22 -- # lslocks -p 70706 00:06:54.396 17:28:25 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@22 -- # grep -q spdk_cpu_lock 00:06:54.655 17:28:25 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@107 -- # killprocess 70695 00:06:54.656 17:28:25 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@950 -- # '[' -z 70695 ']' 00:06:54.656 17:28:25 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@954 -- # kill -0 70695 00:06:54.656 17:28:25 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@955 -- # uname 00:06:54.656 17:28:25 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:06:54.656 17:28:25 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 70695 00:06:54.656 17:28:25 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:06:54.656 17:28:25 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:06:54.656 killing process with pid 70695 00:06:54.656 17:28:25 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@968 -- # echo 'killing process with pid 70695' 00:06:54.656 17:28:25 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@969 -- # kill 70695 00:06:54.656 17:28:25 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@974 -- # wait 70695 00:06:56.062 17:28:27 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@108 -- # killprocess 70706 00:06:56.062 17:28:27 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@950 -- # '[' -z 70706 ']' 00:06:56.062 17:28:27 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@954 -- # kill -0 70706 00:06:56.062 17:28:27 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@955 -- # uname 00:06:56.062 17:28:27 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:06:56.062 17:28:27 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 70706 00:06:56.062 17:28:27 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:06:56.062 17:28:27 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:06:56.062 killing process with pid 70706 00:06:56.062 17:28:27 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@968 -- # echo 'killing process with pid 70706' 00:06:56.062 17:28:27 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@969 -- # kill 70706 00:06:56.062 17:28:27 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@974 -- # wait 70706 00:06:56.632 00:06:56.632 real 0m4.179s 00:06:56.632 user 0m4.038s 00:06:56.632 sys 0m1.283s 00:06:56.632 17:28:27 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@1126 -- # xtrace_disable 00:06:56.632 17:28:27 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@10 -- # set +x 00:06:56.632 ************************************ 00:06:56.632 END TEST locking_app_on_unlocked_coremask 00:06:56.632 ************************************ 00:06:56.632 17:28:27 event.cpu_locks -- event/cpu_locks.sh@170 -- # run_test locking_app_on_locked_coremask locking_app_on_locked_coremask 00:06:56.632 17:28:27 event.cpu_locks -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:06:56.632 17:28:27 event.cpu_locks -- common/autotest_common.sh@1107 -- # xtrace_disable 00:06:56.632 17:28:27 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:06:56.891 ************************************ 00:06:56.891 START TEST locking_app_on_locked_coremask 00:06:56.891 ************************************ 00:06:56.891 17:28:27 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@1125 -- # locking_app_on_locked_coremask 00:06:56.891 17:28:27 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@115 -- # spdk_tgt_pid=70786 00:06:56.891 17:28:27 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@114 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 00:06:56.891 17:28:27 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@116 -- # waitforlisten 70786 /var/tmp/spdk.sock 00:06:56.892 17:28:27 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@831 -- # '[' -z 70786 ']' 00:06:56.892 17:28:27 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:56.892 17:28:27 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@836 -- # local max_retries=100 00:06:56.892 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:56.892 17:28:27 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:56.892 17:28:27 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@840 -- # xtrace_disable 00:06:56.892 17:28:27 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@10 -- # set +x 00:06:56.892 [2024-11-27 17:28:27.916510] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:06:56.892 [2024-11-27 17:28:27.916668] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70786 ] 00:06:56.892 [2024-11-27 17:28:28.059878] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:57.152 [2024-11-27 17:28:28.127899] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:06:57.723 17:28:28 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:06:57.723 17:28:28 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@864 -- # return 0 00:06:57.723 17:28:28 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@118 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1 -r /var/tmp/spdk2.sock 00:06:57.723 17:28:28 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@119 -- # spdk_tgt_pid2=70791 00:06:57.723 17:28:28 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@120 -- # NOT waitforlisten 70791 /var/tmp/spdk2.sock 00:06:57.723 17:28:28 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@650 -- # local es=0 00:06:57.723 17:28:28 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@652 -- # valid_exec_arg waitforlisten 70791 /var/tmp/spdk2.sock 00:06:57.723 17:28:28 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@638 -- # local arg=waitforlisten 00:06:57.723 17:28:28 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:06:57.723 17:28:28 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@642 -- # type -t waitforlisten 00:06:57.723 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:06:57.723 17:28:28 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:06:57.723 17:28:28 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@653 -- # waitforlisten 70791 /var/tmp/spdk2.sock 00:06:57.723 17:28:28 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@831 -- # '[' -z 70791 ']' 00:06:57.723 17:28:28 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk2.sock 00:06:57.723 17:28:28 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@836 -- # local max_retries=100 00:06:57.723 17:28:28 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:06:57.723 17:28:28 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@840 -- # xtrace_disable 00:06:57.723 17:28:28 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@10 -- # set +x 00:06:57.723 [2024-11-27 17:28:28.811092] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:06:57.723 [2024-11-27 17:28:28.811233] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70791 ] 00:06:57.983 [2024-11-27 17:28:28.949974] app.c: 779:claim_cpu_cores: *ERROR*: Cannot create lock on core 0, probably process 70786 has claimed it. 00:06:57.983 [2024-11-27 17:28:28.950043] app.c: 910:spdk_app_start: *ERROR*: Unable to acquire lock on assigned core mask - exiting. 00:06:58.553 ERROR: process (pid: 70791) is no longer running 00:06:58.553 /home/vagrant/spdk_repo/spdk/test/common/autotest_common.sh: line 846: kill: (70791) - No such process 00:06:58.553 17:28:29 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:06:58.553 17:28:29 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@864 -- # return 1 00:06:58.553 17:28:29 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@653 -- # es=1 00:06:58.553 17:28:29 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:06:58.553 17:28:29 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:06:58.553 17:28:29 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:06:58.553 17:28:29 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@122 -- # locks_exist 70786 00:06:58.553 17:28:29 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@22 -- # grep -q spdk_cpu_lock 00:06:58.553 17:28:29 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@22 -- # lslocks -p 70786 00:06:58.813 17:28:29 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@124 -- # killprocess 70786 00:06:58.813 17:28:29 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@950 -- # '[' -z 70786 ']' 00:06:58.813 17:28:29 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@954 -- # kill -0 70786 00:06:58.813 17:28:29 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@955 -- # uname 00:06:58.813 17:28:29 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:06:58.813 17:28:29 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 70786 00:06:58.813 killing process with pid 70786 00:06:58.813 17:28:29 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:06:58.813 17:28:29 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:06:58.813 17:28:29 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@968 -- # echo 'killing process with pid 70786' 00:06:58.813 17:28:29 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@969 -- # kill 70786 00:06:58.813 17:28:29 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@974 -- # wait 70786 00:06:59.381 00:06:59.381 real 0m2.716s 00:06:59.381 user 0m2.742s 00:06:59.381 sys 0m0.875s 00:06:59.381 17:28:30 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@1126 -- # xtrace_disable 00:06:59.381 17:28:30 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@10 -- # set +x 00:06:59.381 ************************************ 00:06:59.381 END TEST locking_app_on_locked_coremask 00:06:59.381 ************************************ 00:06:59.641 17:28:30 event.cpu_locks -- event/cpu_locks.sh@171 -- # run_test locking_overlapped_coremask locking_overlapped_coremask 00:06:59.641 17:28:30 event.cpu_locks -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:06:59.641 17:28:30 event.cpu_locks -- common/autotest_common.sh@1107 -- # xtrace_disable 00:06:59.641 17:28:30 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:06:59.641 ************************************ 00:06:59.641 START TEST locking_overlapped_coremask 00:06:59.641 ************************************ 00:06:59.641 17:28:30 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@1125 -- # locking_overlapped_coremask 00:06:59.641 17:28:30 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@132 -- # spdk_tgt_pid=70844 00:06:59.641 17:28:30 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@131 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x7 00:06:59.641 17:28:30 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@133 -- # waitforlisten 70844 /var/tmp/spdk.sock 00:06:59.641 17:28:30 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@831 -- # '[' -z 70844 ']' 00:06:59.641 17:28:30 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:59.641 17:28:30 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@836 -- # local max_retries=100 00:06:59.641 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:59.641 17:28:30 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:59.641 17:28:30 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@840 -- # xtrace_disable 00:06:59.641 17:28:30 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@10 -- # set +x 00:06:59.641 [2024-11-27 17:28:30.712747] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:06:59.641 [2024-11-27 17:28:30.712892] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x7 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70844 ] 00:06:59.899 [2024-11-27 17:28:30.860745] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 3 00:06:59.899 [2024-11-27 17:28:30.933926] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 1 00:06:59.899 [2024-11-27 17:28:30.934026] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:06:59.899 [2024-11-27 17:28:30.934184] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 2 00:07:00.467 17:28:31 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:07:00.467 17:28:31 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@864 -- # return 0 00:07:00.467 17:28:31 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@135 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1c -r /var/tmp/spdk2.sock 00:07:00.467 17:28:31 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@136 -- # spdk_tgt_pid2=70862 00:07:00.467 17:28:31 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@137 -- # NOT waitforlisten 70862 /var/tmp/spdk2.sock 00:07:00.467 17:28:31 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@650 -- # local es=0 00:07:00.467 17:28:31 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@652 -- # valid_exec_arg waitforlisten 70862 /var/tmp/spdk2.sock 00:07:00.467 17:28:31 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@638 -- # local arg=waitforlisten 00:07:00.467 17:28:31 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:07:00.467 17:28:31 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@642 -- # type -t waitforlisten 00:07:00.467 17:28:31 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:07:00.467 17:28:31 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@653 -- # waitforlisten 70862 /var/tmp/spdk2.sock 00:07:00.467 17:28:31 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@831 -- # '[' -z 70862 ']' 00:07:00.467 17:28:31 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk2.sock 00:07:00.467 17:28:31 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@836 -- # local max_retries=100 00:07:00.467 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:07:00.467 17:28:31 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:07:00.467 17:28:31 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@840 -- # xtrace_disable 00:07:00.467 17:28:31 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@10 -- # set +x 00:07:00.467 [2024-11-27 17:28:31.579682] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:07:00.467 [2024-11-27 17:28:31.579809] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1c --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70862 ] 00:07:00.726 [2024-11-27 17:28:31.719970] app.c: 779:claim_cpu_cores: *ERROR*: Cannot create lock on core 2, probably process 70844 has claimed it. 00:07:00.726 [2024-11-27 17:28:31.720028] app.c: 910:spdk_app_start: *ERROR*: Unable to acquire lock on assigned core mask - exiting. 00:07:01.293 ERROR: process (pid: 70862) is no longer running 00:07:01.293 /home/vagrant/spdk_repo/spdk/test/common/autotest_common.sh: line 846: kill: (70862) - No such process 00:07:01.293 17:28:32 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:07:01.293 17:28:32 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@864 -- # return 1 00:07:01.293 17:28:32 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@653 -- # es=1 00:07:01.293 17:28:32 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:07:01.293 17:28:32 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:07:01.293 17:28:32 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:07:01.293 17:28:32 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@139 -- # check_remaining_locks 00:07:01.293 17:28:32 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@36 -- # locks=(/var/tmp/spdk_cpu_lock_*) 00:07:01.293 17:28:32 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@37 -- # locks_expected=(/var/tmp/spdk_cpu_lock_{000..002}) 00:07:01.293 17:28:32 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@38 -- # [[ /var/tmp/spdk_cpu_lock_000 /var/tmp/spdk_cpu_lock_001 /var/tmp/spdk_cpu_lock_002 == \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\0\ \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\1\ \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\2 ]] 00:07:01.293 17:28:32 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@141 -- # killprocess 70844 00:07:01.293 17:28:32 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@950 -- # '[' -z 70844 ']' 00:07:01.293 17:28:32 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@954 -- # kill -0 70844 00:07:01.293 17:28:32 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@955 -- # uname 00:07:01.293 17:28:32 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:07:01.293 17:28:32 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 70844 00:07:01.293 17:28:32 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:07:01.294 17:28:32 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:07:01.294 killing process with pid 70844 00:07:01.294 17:28:32 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@968 -- # echo 'killing process with pid 70844' 00:07:01.294 17:28:32 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@969 -- # kill 70844 00:07:01.294 17:28:32 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@974 -- # wait 70844 00:07:01.862 00:07:01.862 real 0m2.300s 00:07:01.862 user 0m5.846s 00:07:01.862 sys 0m0.670s 00:07:01.862 17:28:32 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@1126 -- # xtrace_disable 00:07:01.862 17:28:32 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@10 -- # set +x 00:07:01.862 ************************************ 00:07:01.862 END TEST locking_overlapped_coremask 00:07:01.862 ************************************ 00:07:01.862 17:28:32 event.cpu_locks -- event/cpu_locks.sh@172 -- # run_test locking_overlapped_coremask_via_rpc locking_overlapped_coremask_via_rpc 00:07:01.862 17:28:32 event.cpu_locks -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:07:01.862 17:28:32 event.cpu_locks -- common/autotest_common.sh@1107 -- # xtrace_disable 00:07:01.862 17:28:32 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:07:01.862 ************************************ 00:07:01.862 START TEST locking_overlapped_coremask_via_rpc 00:07:01.862 ************************************ 00:07:01.862 17:28:32 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@1125 -- # locking_overlapped_coremask_via_rpc 00:07:01.862 17:28:32 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@148 -- # spdk_tgt_pid=70910 00:07:01.862 17:28:32 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@147 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x7 --disable-cpumask-locks 00:07:01.862 17:28:32 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@149 -- # waitforlisten 70910 /var/tmp/spdk.sock 00:07:01.862 17:28:32 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@831 -- # '[' -z 70910 ']' 00:07:01.862 17:28:32 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:01.862 17:28:32 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@836 -- # local max_retries=100 00:07:01.862 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:01.862 17:28:32 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:01.862 17:28:32 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@840 -- # xtrace_disable 00:07:01.862 17:28:32 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:07:02.121 [2024-11-27 17:28:33.085578] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:07:02.121 [2024-11-27 17:28:33.085705] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x7 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70910 ] 00:07:02.121 [2024-11-27 17:28:33.232804] app.c: 914:spdk_app_start: *NOTICE*: CPU core locks deactivated. 00:07:02.121 [2024-11-27 17:28:33.232879] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 3 00:07:02.121 [2024-11-27 17:28:33.301835] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 1 00:07:02.121 [2024-11-27 17:28:33.301928] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:07:02.121 [2024-11-27 17:28:33.302040] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 2 00:07:03.060 17:28:33 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:07:03.060 17:28:33 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@864 -- # return 0 00:07:03.060 17:28:33 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@151 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x1c -r /var/tmp/spdk2.sock --disable-cpumask-locks 00:07:03.060 17:28:33 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@152 -- # spdk_tgt_pid2=70922 00:07:03.060 17:28:33 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@153 -- # waitforlisten 70922 /var/tmp/spdk2.sock 00:07:03.060 17:28:33 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@831 -- # '[' -z 70922 ']' 00:07:03.060 17:28:33 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk2.sock 00:07:03.060 17:28:33 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@836 -- # local max_retries=100 00:07:03.060 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:07:03.060 17:28:33 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:07:03.060 17:28:33 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@840 -- # xtrace_disable 00:07:03.060 17:28:33 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:07:03.061 [2024-11-27 17:28:33.967601] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:07:03.061 [2024-11-27 17:28:33.967728] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1c --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid70922 ] 00:07:03.061 [2024-11-27 17:28:34.104366] app.c: 914:spdk_app_start: *NOTICE*: CPU core locks deactivated. 00:07:03.061 [2024-11-27 17:28:34.104414] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 3 00:07:03.321 [2024-11-27 17:28:34.260641] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 3 00:07:03.321 [2024-11-27 17:28:34.264366] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 2 00:07:03.321 [2024-11-27 17:28:34.264487] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 4 00:07:03.891 17:28:34 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:07:03.891 17:28:34 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@864 -- # return 0 00:07:03.891 17:28:34 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@155 -- # rpc_cmd framework_enable_cpumask_locks 00:07:03.891 17:28:34 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:03.891 17:28:34 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:07:03.891 17:28:34 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:03.891 17:28:34 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@156 -- # NOT rpc_cmd -s /var/tmp/spdk2.sock framework_enable_cpumask_locks 00:07:03.891 17:28:34 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@650 -- # local es=0 00:07:03.891 17:28:34 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd -s /var/tmp/spdk2.sock framework_enable_cpumask_locks 00:07:03.891 17:28:34 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:07:03.891 17:28:34 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:07:03.891 17:28:34 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:07:03.891 17:28:34 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:07:03.891 17:28:34 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@653 -- # rpc_cmd -s /var/tmp/spdk2.sock framework_enable_cpumask_locks 00:07:03.891 17:28:34 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:03.891 17:28:34 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:07:03.891 [2024-11-27 17:28:34.919345] app.c: 779:claim_cpu_cores: *ERROR*: Cannot create lock on core 2, probably process 70910 has claimed it. 00:07:03.891 request: 00:07:03.891 { 00:07:03.891 "method": "framework_enable_cpumask_locks", 00:07:03.891 "req_id": 1 00:07:03.891 } 00:07:03.891 Got JSON-RPC error response 00:07:03.891 response: 00:07:03.891 { 00:07:03.891 "code": -32603, 00:07:03.891 "message": "Failed to claim CPU core: 2" 00:07:03.891 } 00:07:03.891 17:28:34 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:07:03.891 17:28:34 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@653 -- # es=1 00:07:03.891 17:28:34 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:07:03.891 17:28:34 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:07:03.891 17:28:34 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:07:03.891 17:28:34 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@158 -- # waitforlisten 70910 /var/tmp/spdk.sock 00:07:03.891 17:28:34 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@831 -- # '[' -z 70910 ']' 00:07:03.891 17:28:34 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:03.891 17:28:34 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@836 -- # local max_retries=100 00:07:03.891 17:28:34 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:03.891 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:03.891 17:28:34 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@840 -- # xtrace_disable 00:07:03.891 17:28:34 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:07:04.151 17:28:35 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:07:04.151 17:28:35 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@864 -- # return 0 00:07:04.151 17:28:35 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@159 -- # waitforlisten 70922 /var/tmp/spdk2.sock 00:07:04.151 17:28:35 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@831 -- # '[' -z 70922 ']' 00:07:04.151 17:28:35 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk2.sock 00:07:04.151 17:28:35 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@836 -- # local max_retries=100 00:07:04.151 17:28:35 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:07:04.151 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:07:04.151 17:28:35 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@840 -- # xtrace_disable 00:07:04.151 17:28:35 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:07:04.412 17:28:35 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:07:04.412 17:28:35 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@864 -- # return 0 00:07:04.412 17:28:35 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@161 -- # check_remaining_locks 00:07:04.412 17:28:35 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@36 -- # locks=(/var/tmp/spdk_cpu_lock_*) 00:07:04.412 17:28:35 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@37 -- # locks_expected=(/var/tmp/spdk_cpu_lock_{000..002}) 00:07:04.412 17:28:35 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@38 -- # [[ /var/tmp/spdk_cpu_lock_000 /var/tmp/spdk_cpu_lock_001 /var/tmp/spdk_cpu_lock_002 == \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\0\ \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\1\ \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\2 ]] 00:07:04.412 00:07:04.412 real 0m2.361s 00:07:04.412 user 0m1.031s 00:07:04.412 sys 0m0.182s 00:07:04.412 17:28:35 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@1126 -- # xtrace_disable 00:07:04.412 17:28:35 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:07:04.412 ************************************ 00:07:04.412 END TEST locking_overlapped_coremask_via_rpc 00:07:04.412 ************************************ 00:07:04.412 17:28:35 event.cpu_locks -- event/cpu_locks.sh@174 -- # cleanup 00:07:04.412 17:28:35 event.cpu_locks -- event/cpu_locks.sh@15 -- # [[ -z 70910 ]] 00:07:04.412 17:28:35 event.cpu_locks -- event/cpu_locks.sh@15 -- # killprocess 70910 00:07:04.412 17:28:35 event.cpu_locks -- common/autotest_common.sh@950 -- # '[' -z 70910 ']' 00:07:04.412 17:28:35 event.cpu_locks -- common/autotest_common.sh@954 -- # kill -0 70910 00:07:04.412 17:28:35 event.cpu_locks -- common/autotest_common.sh@955 -- # uname 00:07:04.412 17:28:35 event.cpu_locks -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:07:04.412 17:28:35 event.cpu_locks -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 70910 00:07:04.412 17:28:35 event.cpu_locks -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:07:04.412 17:28:35 event.cpu_locks -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:07:04.412 17:28:35 event.cpu_locks -- common/autotest_common.sh@968 -- # echo 'killing process with pid 70910' 00:07:04.412 killing process with pid 70910 00:07:04.412 17:28:35 event.cpu_locks -- common/autotest_common.sh@969 -- # kill 70910 00:07:04.412 17:28:35 event.cpu_locks -- common/autotest_common.sh@974 -- # wait 70910 00:07:04.983 17:28:36 event.cpu_locks -- event/cpu_locks.sh@16 -- # [[ -z 70922 ]] 00:07:04.984 17:28:36 event.cpu_locks -- event/cpu_locks.sh@16 -- # killprocess 70922 00:07:04.984 17:28:36 event.cpu_locks -- common/autotest_common.sh@950 -- # '[' -z 70922 ']' 00:07:04.984 17:28:36 event.cpu_locks -- common/autotest_common.sh@954 -- # kill -0 70922 00:07:04.984 17:28:36 event.cpu_locks -- common/autotest_common.sh@955 -- # uname 00:07:04.984 17:28:36 event.cpu_locks -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:07:04.984 17:28:36 event.cpu_locks -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 70922 00:07:04.984 17:28:36 event.cpu_locks -- common/autotest_common.sh@956 -- # process_name=reactor_2 00:07:04.984 17:28:36 event.cpu_locks -- common/autotest_common.sh@960 -- # '[' reactor_2 = sudo ']' 00:07:04.984 17:28:36 event.cpu_locks -- common/autotest_common.sh@968 -- # echo 'killing process with pid 70922' 00:07:04.984 killing process with pid 70922 00:07:04.984 17:28:36 event.cpu_locks -- common/autotest_common.sh@969 -- # kill 70922 00:07:04.984 17:28:36 event.cpu_locks -- common/autotest_common.sh@974 -- # wait 70922 00:07:05.924 17:28:36 event.cpu_locks -- event/cpu_locks.sh@18 -- # rm -f 00:07:05.924 17:28:36 event.cpu_locks -- event/cpu_locks.sh@1 -- # cleanup 00:07:05.924 17:28:36 event.cpu_locks -- event/cpu_locks.sh@15 -- # [[ -z 70910 ]] 00:07:05.925 17:28:36 event.cpu_locks -- event/cpu_locks.sh@15 -- # killprocess 70910 00:07:05.925 17:28:36 event.cpu_locks -- common/autotest_common.sh@950 -- # '[' -z 70910 ']' 00:07:05.925 17:28:36 event.cpu_locks -- common/autotest_common.sh@954 -- # kill -0 70910 00:07:05.925 /home/vagrant/spdk_repo/spdk/test/common/autotest_common.sh: line 954: kill: (70910) - No such process 00:07:05.925 Process with pid 70910 is not found 00:07:05.925 17:28:36 event.cpu_locks -- common/autotest_common.sh@977 -- # echo 'Process with pid 70910 is not found' 00:07:05.925 17:28:36 event.cpu_locks -- event/cpu_locks.sh@16 -- # [[ -z 70922 ]] 00:07:05.925 17:28:36 event.cpu_locks -- event/cpu_locks.sh@16 -- # killprocess 70922 00:07:05.925 17:28:36 event.cpu_locks -- common/autotest_common.sh@950 -- # '[' -z 70922 ']' 00:07:05.925 17:28:36 event.cpu_locks -- common/autotest_common.sh@954 -- # kill -0 70922 00:07:05.925 /home/vagrant/spdk_repo/spdk/test/common/autotest_common.sh: line 954: kill: (70922) - No such process 00:07:05.925 Process with pid 70922 is not found 00:07:05.925 17:28:36 event.cpu_locks -- common/autotest_common.sh@977 -- # echo 'Process with pid 70922 is not found' 00:07:05.925 17:28:36 event.cpu_locks -- event/cpu_locks.sh@18 -- # rm -f 00:07:05.925 00:07:05.925 real 0m22.462s 00:07:05.925 user 0m35.234s 00:07:05.925 sys 0m7.501s 00:07:05.925 17:28:36 event.cpu_locks -- common/autotest_common.sh@1126 -- # xtrace_disable 00:07:05.925 17:28:36 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:07:05.925 ************************************ 00:07:05.925 END TEST cpu_locks 00:07:05.925 ************************************ 00:07:05.925 00:07:05.925 real 0m50.913s 00:07:05.925 user 1m33.291s 00:07:05.925 sys 0m11.761s 00:07:05.925 17:28:36 event -- common/autotest_common.sh@1126 -- # xtrace_disable 00:07:05.925 17:28:36 event -- common/autotest_common.sh@10 -- # set +x 00:07:05.925 ************************************ 00:07:05.925 END TEST event 00:07:05.925 ************************************ 00:07:05.925 17:28:36 -- spdk/autotest.sh@169 -- # run_test thread /home/vagrant/spdk_repo/spdk/test/thread/thread.sh 00:07:05.925 17:28:36 -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:07:05.925 17:28:36 -- common/autotest_common.sh@1107 -- # xtrace_disable 00:07:05.925 17:28:36 -- common/autotest_common.sh@10 -- # set +x 00:07:05.925 ************************************ 00:07:05.925 START TEST thread 00:07:05.925 ************************************ 00:07:05.925 17:28:36 thread -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/thread/thread.sh 00:07:05.925 * Looking for test storage... 00:07:05.925 * Found test storage at /home/vagrant/spdk_repo/spdk/test/thread 00:07:05.925 17:28:37 thread -- common/autotest_common.sh@1680 -- # [[ y == y ]] 00:07:05.925 17:28:37 thread -- common/autotest_common.sh@1681 -- # lcov --version 00:07:05.925 17:28:37 thread -- common/autotest_common.sh@1681 -- # awk '{print $NF}' 00:07:06.185 17:28:37 thread -- common/autotest_common.sh@1681 -- # lt 1.15 2 00:07:06.185 17:28:37 thread -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:07:06.185 17:28:37 thread -- scripts/common.sh@333 -- # local ver1 ver1_l 00:07:06.185 17:28:37 thread -- scripts/common.sh@334 -- # local ver2 ver2_l 00:07:06.185 17:28:37 thread -- scripts/common.sh@336 -- # IFS=.-: 00:07:06.185 17:28:37 thread -- scripts/common.sh@336 -- # read -ra ver1 00:07:06.185 17:28:37 thread -- scripts/common.sh@337 -- # IFS=.-: 00:07:06.185 17:28:37 thread -- scripts/common.sh@337 -- # read -ra ver2 00:07:06.185 17:28:37 thread -- scripts/common.sh@338 -- # local 'op=<' 00:07:06.185 17:28:37 thread -- scripts/common.sh@340 -- # ver1_l=2 00:07:06.185 17:28:37 thread -- scripts/common.sh@341 -- # ver2_l=1 00:07:06.185 17:28:37 thread -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:07:06.185 17:28:37 thread -- scripts/common.sh@344 -- # case "$op" in 00:07:06.185 17:28:37 thread -- scripts/common.sh@345 -- # : 1 00:07:06.185 17:28:37 thread -- scripts/common.sh@364 -- # (( v = 0 )) 00:07:06.185 17:28:37 thread -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:07:06.185 17:28:37 thread -- scripts/common.sh@365 -- # decimal 1 00:07:06.185 17:28:37 thread -- scripts/common.sh@353 -- # local d=1 00:07:06.185 17:28:37 thread -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:07:06.185 17:28:37 thread -- scripts/common.sh@355 -- # echo 1 00:07:06.185 17:28:37 thread -- scripts/common.sh@365 -- # ver1[v]=1 00:07:06.185 17:28:37 thread -- scripts/common.sh@366 -- # decimal 2 00:07:06.185 17:28:37 thread -- scripts/common.sh@353 -- # local d=2 00:07:06.185 17:28:37 thread -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:07:06.185 17:28:37 thread -- scripts/common.sh@355 -- # echo 2 00:07:06.185 17:28:37 thread -- scripts/common.sh@366 -- # ver2[v]=2 00:07:06.185 17:28:37 thread -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:07:06.185 17:28:37 thread -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:07:06.185 17:28:37 thread -- scripts/common.sh@368 -- # return 0 00:07:06.185 17:28:37 thread -- common/autotest_common.sh@1682 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:07:06.185 17:28:37 thread -- common/autotest_common.sh@1694 -- # export 'LCOV_OPTS= 00:07:06.185 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:06.185 --rc genhtml_branch_coverage=1 00:07:06.185 --rc genhtml_function_coverage=1 00:07:06.185 --rc genhtml_legend=1 00:07:06.185 --rc geninfo_all_blocks=1 00:07:06.185 --rc geninfo_unexecuted_blocks=1 00:07:06.185 00:07:06.185 ' 00:07:06.185 17:28:37 thread -- common/autotest_common.sh@1694 -- # LCOV_OPTS=' 00:07:06.185 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:06.185 --rc genhtml_branch_coverage=1 00:07:06.185 --rc genhtml_function_coverage=1 00:07:06.185 --rc genhtml_legend=1 00:07:06.185 --rc geninfo_all_blocks=1 00:07:06.185 --rc geninfo_unexecuted_blocks=1 00:07:06.185 00:07:06.185 ' 00:07:06.185 17:28:37 thread -- common/autotest_common.sh@1695 -- # export 'LCOV=lcov 00:07:06.185 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:06.185 --rc genhtml_branch_coverage=1 00:07:06.185 --rc genhtml_function_coverage=1 00:07:06.185 --rc genhtml_legend=1 00:07:06.185 --rc geninfo_all_blocks=1 00:07:06.185 --rc geninfo_unexecuted_blocks=1 00:07:06.185 00:07:06.185 ' 00:07:06.185 17:28:37 thread -- common/autotest_common.sh@1695 -- # LCOV='lcov 00:07:06.185 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:06.185 --rc genhtml_branch_coverage=1 00:07:06.185 --rc genhtml_function_coverage=1 00:07:06.185 --rc genhtml_legend=1 00:07:06.185 --rc geninfo_all_blocks=1 00:07:06.185 --rc geninfo_unexecuted_blocks=1 00:07:06.185 00:07:06.185 ' 00:07:06.185 17:28:37 thread -- thread/thread.sh@11 -- # run_test thread_poller_perf /home/vagrant/spdk_repo/spdk/test/thread/poller_perf/poller_perf -b 1000 -l 1 -t 1 00:07:06.185 17:28:37 thread -- common/autotest_common.sh@1101 -- # '[' 8 -le 1 ']' 00:07:06.185 17:28:37 thread -- common/autotest_common.sh@1107 -- # xtrace_disable 00:07:06.185 17:28:37 thread -- common/autotest_common.sh@10 -- # set +x 00:07:06.185 ************************************ 00:07:06.185 START TEST thread_poller_perf 00:07:06.185 ************************************ 00:07:06.186 17:28:37 thread.thread_poller_perf -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/thread/poller_perf/poller_perf -b 1000 -l 1 -t 1 00:07:06.186 [2024-11-27 17:28:37.216648] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:07:06.186 [2024-11-27 17:28:37.216796] [ DPDK EAL parameters: poller_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid71066 ] 00:07:06.186 [2024-11-27 17:28:37.363813] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:06.446 Running 1000 pollers for 1 seconds with 1 microseconds period. 00:07:06.446 [2024-11-27 17:28:37.438508] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:07:07.393 [2024-11-27T17:28:38.585Z] ====================================== 00:07:07.393 [2024-11-27T17:28:38.585Z] busy:2297984714 (cyc) 00:07:07.393 [2024-11-27T17:28:38.585Z] total_run_count: 418000 00:07:07.393 [2024-11-27T17:28:38.585Z] tsc_hz: 2290000000 (cyc) 00:07:07.393 [2024-11-27T17:28:38.585Z] ====================================== 00:07:07.393 [2024-11-27T17:28:38.585Z] poller_cost: 5497 (cyc), 2400 (nsec) 00:07:07.393 00:07:07.393 real 0m1.404s 00:07:07.393 user 0m1.194s 00:07:07.393 sys 0m0.104s 00:07:07.393 17:28:38 thread.thread_poller_perf -- common/autotest_common.sh@1126 -- # xtrace_disable 00:07:07.393 17:28:38 thread.thread_poller_perf -- common/autotest_common.sh@10 -- # set +x 00:07:07.393 ************************************ 00:07:07.393 END TEST thread_poller_perf 00:07:07.393 ************************************ 00:07:07.671 17:28:38 thread -- thread/thread.sh@12 -- # run_test thread_poller_perf /home/vagrant/spdk_repo/spdk/test/thread/poller_perf/poller_perf -b 1000 -l 0 -t 1 00:07:07.671 17:28:38 thread -- common/autotest_common.sh@1101 -- # '[' 8 -le 1 ']' 00:07:07.671 17:28:38 thread -- common/autotest_common.sh@1107 -- # xtrace_disable 00:07:07.671 17:28:38 thread -- common/autotest_common.sh@10 -- # set +x 00:07:07.671 ************************************ 00:07:07.671 START TEST thread_poller_perf 00:07:07.671 ************************************ 00:07:07.671 17:28:38 thread.thread_poller_perf -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/thread/poller_perf/poller_perf -b 1000 -l 0 -t 1 00:07:07.671 [2024-11-27 17:28:38.682915] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:07:07.671 [2024-11-27 17:28:38.683062] [ DPDK EAL parameters: poller_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid71103 ] 00:07:07.671 [2024-11-27 17:28:38.816617] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:07.931 Running 1000 pollers for 1 seconds with 0 microseconds period. 00:07:07.931 [2024-11-27 17:28:38.890282] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:07:08.867 [2024-11-27T17:28:40.059Z] ====================================== 00:07:08.867 [2024-11-27T17:28:40.059Z] busy:2293486764 (cyc) 00:07:08.867 [2024-11-27T17:28:40.059Z] total_run_count: 5461000 00:07:08.867 [2024-11-27T17:28:40.059Z] tsc_hz: 2290000000 (cyc) 00:07:08.867 [2024-11-27T17:28:40.059Z] ====================================== 00:07:08.867 [2024-11-27T17:28:40.059Z] poller_cost: 419 (cyc), 182 (nsec) 00:07:08.867 00:07:08.867 real 0m1.384s 00:07:08.867 user 0m1.177s 00:07:08.867 sys 0m0.101s 00:07:08.867 17:28:40 thread.thread_poller_perf -- common/autotest_common.sh@1126 -- # xtrace_disable 00:07:08.867 17:28:40 thread.thread_poller_perf -- common/autotest_common.sh@10 -- # set +x 00:07:08.867 ************************************ 00:07:08.867 END TEST thread_poller_perf 00:07:08.867 ************************************ 00:07:09.127 17:28:40 thread -- thread/thread.sh@17 -- # [[ y != \y ]] 00:07:09.127 00:07:09.127 real 0m3.147s 00:07:09.127 user 0m2.532s 00:07:09.127 sys 0m0.419s 00:07:09.127 17:28:40 thread -- common/autotest_common.sh@1126 -- # xtrace_disable 00:07:09.127 17:28:40 thread -- common/autotest_common.sh@10 -- # set +x 00:07:09.127 ************************************ 00:07:09.127 END TEST thread 00:07:09.127 ************************************ 00:07:09.127 17:28:40 -- spdk/autotest.sh@171 -- # [[ 0 -eq 1 ]] 00:07:09.127 17:28:40 -- spdk/autotest.sh@176 -- # run_test app_cmdline /home/vagrant/spdk_repo/spdk/test/app/cmdline.sh 00:07:09.127 17:28:40 -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:07:09.127 17:28:40 -- common/autotest_common.sh@1107 -- # xtrace_disable 00:07:09.127 17:28:40 -- common/autotest_common.sh@10 -- # set +x 00:07:09.127 ************************************ 00:07:09.127 START TEST app_cmdline 00:07:09.127 ************************************ 00:07:09.127 17:28:40 app_cmdline -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/app/cmdline.sh 00:07:09.127 * Looking for test storage... 00:07:09.127 * Found test storage at /home/vagrant/spdk_repo/spdk/test/app 00:07:09.127 17:28:40 app_cmdline -- common/autotest_common.sh@1680 -- # [[ y == y ]] 00:07:09.127 17:28:40 app_cmdline -- common/autotest_common.sh@1681 -- # lcov --version 00:07:09.127 17:28:40 app_cmdline -- common/autotest_common.sh@1681 -- # awk '{print $NF}' 00:07:09.387 17:28:40 app_cmdline -- common/autotest_common.sh@1681 -- # lt 1.15 2 00:07:09.387 17:28:40 app_cmdline -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:07:09.387 17:28:40 app_cmdline -- scripts/common.sh@333 -- # local ver1 ver1_l 00:07:09.387 17:28:40 app_cmdline -- scripts/common.sh@334 -- # local ver2 ver2_l 00:07:09.387 17:28:40 app_cmdline -- scripts/common.sh@336 -- # IFS=.-: 00:07:09.387 17:28:40 app_cmdline -- scripts/common.sh@336 -- # read -ra ver1 00:07:09.387 17:28:40 app_cmdline -- scripts/common.sh@337 -- # IFS=.-: 00:07:09.387 17:28:40 app_cmdline -- scripts/common.sh@337 -- # read -ra ver2 00:07:09.387 17:28:40 app_cmdline -- scripts/common.sh@338 -- # local 'op=<' 00:07:09.387 17:28:40 app_cmdline -- scripts/common.sh@340 -- # ver1_l=2 00:07:09.387 17:28:40 app_cmdline -- scripts/common.sh@341 -- # ver2_l=1 00:07:09.387 17:28:40 app_cmdline -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:07:09.387 17:28:40 app_cmdline -- scripts/common.sh@344 -- # case "$op" in 00:07:09.387 17:28:40 app_cmdline -- scripts/common.sh@345 -- # : 1 00:07:09.387 17:28:40 app_cmdline -- scripts/common.sh@364 -- # (( v = 0 )) 00:07:09.387 17:28:40 app_cmdline -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:07:09.387 17:28:40 app_cmdline -- scripts/common.sh@365 -- # decimal 1 00:07:09.387 17:28:40 app_cmdline -- scripts/common.sh@353 -- # local d=1 00:07:09.387 17:28:40 app_cmdline -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:07:09.387 17:28:40 app_cmdline -- scripts/common.sh@355 -- # echo 1 00:07:09.387 17:28:40 app_cmdline -- scripts/common.sh@365 -- # ver1[v]=1 00:07:09.387 17:28:40 app_cmdline -- scripts/common.sh@366 -- # decimal 2 00:07:09.387 17:28:40 app_cmdline -- scripts/common.sh@353 -- # local d=2 00:07:09.387 17:28:40 app_cmdline -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:07:09.387 17:28:40 app_cmdline -- scripts/common.sh@355 -- # echo 2 00:07:09.387 17:28:40 app_cmdline -- scripts/common.sh@366 -- # ver2[v]=2 00:07:09.387 17:28:40 app_cmdline -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:07:09.387 17:28:40 app_cmdline -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:07:09.387 17:28:40 app_cmdline -- scripts/common.sh@368 -- # return 0 00:07:09.387 17:28:40 app_cmdline -- common/autotest_common.sh@1682 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:07:09.387 17:28:40 app_cmdline -- common/autotest_common.sh@1694 -- # export 'LCOV_OPTS= 00:07:09.387 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:09.387 --rc genhtml_branch_coverage=1 00:07:09.387 --rc genhtml_function_coverage=1 00:07:09.387 --rc genhtml_legend=1 00:07:09.387 --rc geninfo_all_blocks=1 00:07:09.387 --rc geninfo_unexecuted_blocks=1 00:07:09.387 00:07:09.387 ' 00:07:09.387 17:28:40 app_cmdline -- common/autotest_common.sh@1694 -- # LCOV_OPTS=' 00:07:09.387 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:09.387 --rc genhtml_branch_coverage=1 00:07:09.387 --rc genhtml_function_coverage=1 00:07:09.387 --rc genhtml_legend=1 00:07:09.387 --rc geninfo_all_blocks=1 00:07:09.387 --rc geninfo_unexecuted_blocks=1 00:07:09.387 00:07:09.387 ' 00:07:09.387 17:28:40 app_cmdline -- common/autotest_common.sh@1695 -- # export 'LCOV=lcov 00:07:09.387 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:09.387 --rc genhtml_branch_coverage=1 00:07:09.387 --rc genhtml_function_coverage=1 00:07:09.387 --rc genhtml_legend=1 00:07:09.387 --rc geninfo_all_blocks=1 00:07:09.387 --rc geninfo_unexecuted_blocks=1 00:07:09.387 00:07:09.387 ' 00:07:09.387 17:28:40 app_cmdline -- common/autotest_common.sh@1695 -- # LCOV='lcov 00:07:09.387 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:09.387 --rc genhtml_branch_coverage=1 00:07:09.387 --rc genhtml_function_coverage=1 00:07:09.387 --rc genhtml_legend=1 00:07:09.387 --rc geninfo_all_blocks=1 00:07:09.388 --rc geninfo_unexecuted_blocks=1 00:07:09.388 00:07:09.388 ' 00:07:09.388 17:28:40 app_cmdline -- app/cmdline.sh@14 -- # trap 'killprocess $spdk_tgt_pid' EXIT 00:07:09.388 17:28:40 app_cmdline -- app/cmdline.sh@17 -- # spdk_tgt_pid=71189 00:07:09.388 17:28:40 app_cmdline -- app/cmdline.sh@16 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt --rpcs-allowed spdk_get_version,rpc_get_methods 00:07:09.388 17:28:40 app_cmdline -- app/cmdline.sh@18 -- # waitforlisten 71189 00:07:09.388 17:28:40 app_cmdline -- common/autotest_common.sh@831 -- # '[' -z 71189 ']' 00:07:09.388 17:28:40 app_cmdline -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:09.388 17:28:40 app_cmdline -- common/autotest_common.sh@836 -- # local max_retries=100 00:07:09.388 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:09.388 17:28:40 app_cmdline -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:09.388 17:28:40 app_cmdline -- common/autotest_common.sh@840 -- # xtrace_disable 00:07:09.388 17:28:40 app_cmdline -- common/autotest_common.sh@10 -- # set +x 00:07:09.388 [2024-11-27 17:28:40.474594] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:07:09.388 [2024-11-27 17:28:40.474727] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid71189 ] 00:07:09.647 [2024-11-27 17:28:40.620362] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:09.647 [2024-11-27 17:28:40.694414] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:07:10.218 17:28:41 app_cmdline -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:07:10.218 17:28:41 app_cmdline -- common/autotest_common.sh@864 -- # return 0 00:07:10.218 17:28:41 app_cmdline -- app/cmdline.sh@20 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py spdk_get_version 00:07:10.477 { 00:07:10.477 "version": "SPDK v24.09.1-pre git sha1 b18e1bd62", 00:07:10.477 "fields": { 00:07:10.477 "major": 24, 00:07:10.477 "minor": 9, 00:07:10.477 "patch": 1, 00:07:10.477 "suffix": "-pre", 00:07:10.477 "commit": "b18e1bd62" 00:07:10.477 } 00:07:10.477 } 00:07:10.477 17:28:41 app_cmdline -- app/cmdline.sh@22 -- # expected_methods=() 00:07:10.477 17:28:41 app_cmdline -- app/cmdline.sh@23 -- # expected_methods+=("rpc_get_methods") 00:07:10.477 17:28:41 app_cmdline -- app/cmdline.sh@24 -- # expected_methods+=("spdk_get_version") 00:07:10.477 17:28:41 app_cmdline -- app/cmdline.sh@26 -- # methods=($(rpc_cmd rpc_get_methods | jq -r ".[]" | sort)) 00:07:10.477 17:28:41 app_cmdline -- app/cmdline.sh@26 -- # rpc_cmd rpc_get_methods 00:07:10.477 17:28:41 app_cmdline -- app/cmdline.sh@26 -- # jq -r '.[]' 00:07:10.477 17:28:41 app_cmdline -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:10.477 17:28:41 app_cmdline -- common/autotest_common.sh@10 -- # set +x 00:07:10.477 17:28:41 app_cmdline -- app/cmdline.sh@26 -- # sort 00:07:10.477 17:28:41 app_cmdline -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:10.477 17:28:41 app_cmdline -- app/cmdline.sh@27 -- # (( 2 == 2 )) 00:07:10.477 17:28:41 app_cmdline -- app/cmdline.sh@28 -- # [[ rpc_get_methods spdk_get_version == \r\p\c\_\g\e\t\_\m\e\t\h\o\d\s\ \s\p\d\k\_\g\e\t\_\v\e\r\s\i\o\n ]] 00:07:10.477 17:28:41 app_cmdline -- app/cmdline.sh@30 -- # NOT /home/vagrant/spdk_repo/spdk/scripts/rpc.py env_dpdk_get_mem_stats 00:07:10.477 17:28:41 app_cmdline -- common/autotest_common.sh@650 -- # local es=0 00:07:10.477 17:28:41 app_cmdline -- common/autotest_common.sh@652 -- # valid_exec_arg /home/vagrant/spdk_repo/spdk/scripts/rpc.py env_dpdk_get_mem_stats 00:07:10.477 17:28:41 app_cmdline -- common/autotest_common.sh@638 -- # local arg=/home/vagrant/spdk_repo/spdk/scripts/rpc.py 00:07:10.477 17:28:41 app_cmdline -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:07:10.477 17:28:41 app_cmdline -- common/autotest_common.sh@642 -- # type -t /home/vagrant/spdk_repo/spdk/scripts/rpc.py 00:07:10.477 17:28:41 app_cmdline -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:07:10.477 17:28:41 app_cmdline -- common/autotest_common.sh@644 -- # type -P /home/vagrant/spdk_repo/spdk/scripts/rpc.py 00:07:10.477 17:28:41 app_cmdline -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:07:10.477 17:28:41 app_cmdline -- common/autotest_common.sh@644 -- # arg=/home/vagrant/spdk_repo/spdk/scripts/rpc.py 00:07:10.477 17:28:41 app_cmdline -- common/autotest_common.sh@644 -- # [[ -x /home/vagrant/spdk_repo/spdk/scripts/rpc.py ]] 00:07:10.477 17:28:41 app_cmdline -- common/autotest_common.sh@653 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py env_dpdk_get_mem_stats 00:07:10.737 request: 00:07:10.737 { 00:07:10.737 "method": "env_dpdk_get_mem_stats", 00:07:10.737 "req_id": 1 00:07:10.737 } 00:07:10.737 Got JSON-RPC error response 00:07:10.737 response: 00:07:10.737 { 00:07:10.737 "code": -32601, 00:07:10.737 "message": "Method not found" 00:07:10.737 } 00:07:10.737 17:28:41 app_cmdline -- common/autotest_common.sh@653 -- # es=1 00:07:10.738 17:28:41 app_cmdline -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:07:10.738 17:28:41 app_cmdline -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:07:10.738 17:28:41 app_cmdline -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:07:10.738 17:28:41 app_cmdline -- app/cmdline.sh@1 -- # killprocess 71189 00:07:10.738 17:28:41 app_cmdline -- common/autotest_common.sh@950 -- # '[' -z 71189 ']' 00:07:10.738 17:28:41 app_cmdline -- common/autotest_common.sh@954 -- # kill -0 71189 00:07:10.738 17:28:41 app_cmdline -- common/autotest_common.sh@955 -- # uname 00:07:10.738 17:28:41 app_cmdline -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:07:10.738 17:28:41 app_cmdline -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 71189 00:07:10.738 17:28:41 app_cmdline -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:07:10.738 17:28:41 app_cmdline -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:07:10.738 killing process with pid 71189 00:07:10.738 17:28:41 app_cmdline -- common/autotest_common.sh@968 -- # echo 'killing process with pid 71189' 00:07:10.738 17:28:41 app_cmdline -- common/autotest_common.sh@969 -- # kill 71189 00:07:10.738 17:28:41 app_cmdline -- common/autotest_common.sh@974 -- # wait 71189 00:07:11.306 00:07:11.306 real 0m2.254s 00:07:11.306 user 0m2.269s 00:07:11.306 sys 0m0.757s 00:07:11.306 17:28:42 app_cmdline -- common/autotest_common.sh@1126 -- # xtrace_disable 00:07:11.306 17:28:42 app_cmdline -- common/autotest_common.sh@10 -- # set +x 00:07:11.306 ************************************ 00:07:11.306 END TEST app_cmdline 00:07:11.306 ************************************ 00:07:11.306 17:28:42 -- spdk/autotest.sh@177 -- # run_test version /home/vagrant/spdk_repo/spdk/test/app/version.sh 00:07:11.306 17:28:42 -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:07:11.306 17:28:42 -- common/autotest_common.sh@1107 -- # xtrace_disable 00:07:11.306 17:28:42 -- common/autotest_common.sh@10 -- # set +x 00:07:11.306 ************************************ 00:07:11.306 START TEST version 00:07:11.306 ************************************ 00:07:11.306 17:28:42 version -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/app/version.sh 00:07:11.567 * Looking for test storage... 00:07:11.567 * Found test storage at /home/vagrant/spdk_repo/spdk/test/app 00:07:11.567 17:28:42 version -- common/autotest_common.sh@1680 -- # [[ y == y ]] 00:07:11.567 17:28:42 version -- common/autotest_common.sh@1681 -- # lcov --version 00:07:11.567 17:28:42 version -- common/autotest_common.sh@1681 -- # awk '{print $NF}' 00:07:11.567 17:28:42 version -- common/autotest_common.sh@1681 -- # lt 1.15 2 00:07:11.567 17:28:42 version -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:07:11.567 17:28:42 version -- scripts/common.sh@333 -- # local ver1 ver1_l 00:07:11.567 17:28:42 version -- scripts/common.sh@334 -- # local ver2 ver2_l 00:07:11.567 17:28:42 version -- scripts/common.sh@336 -- # IFS=.-: 00:07:11.567 17:28:42 version -- scripts/common.sh@336 -- # read -ra ver1 00:07:11.567 17:28:42 version -- scripts/common.sh@337 -- # IFS=.-: 00:07:11.567 17:28:42 version -- scripts/common.sh@337 -- # read -ra ver2 00:07:11.567 17:28:42 version -- scripts/common.sh@338 -- # local 'op=<' 00:07:11.567 17:28:42 version -- scripts/common.sh@340 -- # ver1_l=2 00:07:11.567 17:28:42 version -- scripts/common.sh@341 -- # ver2_l=1 00:07:11.567 17:28:42 version -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:07:11.567 17:28:42 version -- scripts/common.sh@344 -- # case "$op" in 00:07:11.567 17:28:42 version -- scripts/common.sh@345 -- # : 1 00:07:11.567 17:28:42 version -- scripts/common.sh@364 -- # (( v = 0 )) 00:07:11.567 17:28:42 version -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:07:11.567 17:28:42 version -- scripts/common.sh@365 -- # decimal 1 00:07:11.567 17:28:42 version -- scripts/common.sh@353 -- # local d=1 00:07:11.567 17:28:42 version -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:07:11.567 17:28:42 version -- scripts/common.sh@355 -- # echo 1 00:07:11.567 17:28:42 version -- scripts/common.sh@365 -- # ver1[v]=1 00:07:11.567 17:28:42 version -- scripts/common.sh@366 -- # decimal 2 00:07:11.567 17:28:42 version -- scripts/common.sh@353 -- # local d=2 00:07:11.567 17:28:42 version -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:07:11.567 17:28:42 version -- scripts/common.sh@355 -- # echo 2 00:07:11.567 17:28:42 version -- scripts/common.sh@366 -- # ver2[v]=2 00:07:11.567 17:28:42 version -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:07:11.567 17:28:42 version -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:07:11.567 17:28:42 version -- scripts/common.sh@368 -- # return 0 00:07:11.567 17:28:42 version -- common/autotest_common.sh@1682 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:07:11.567 17:28:42 version -- common/autotest_common.sh@1694 -- # export 'LCOV_OPTS= 00:07:11.567 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:11.567 --rc genhtml_branch_coverage=1 00:07:11.567 --rc genhtml_function_coverage=1 00:07:11.567 --rc genhtml_legend=1 00:07:11.567 --rc geninfo_all_blocks=1 00:07:11.567 --rc geninfo_unexecuted_blocks=1 00:07:11.567 00:07:11.567 ' 00:07:11.567 17:28:42 version -- common/autotest_common.sh@1694 -- # LCOV_OPTS=' 00:07:11.567 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:11.567 --rc genhtml_branch_coverage=1 00:07:11.567 --rc genhtml_function_coverage=1 00:07:11.567 --rc genhtml_legend=1 00:07:11.567 --rc geninfo_all_blocks=1 00:07:11.567 --rc geninfo_unexecuted_blocks=1 00:07:11.567 00:07:11.567 ' 00:07:11.567 17:28:42 version -- common/autotest_common.sh@1695 -- # export 'LCOV=lcov 00:07:11.567 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:11.567 --rc genhtml_branch_coverage=1 00:07:11.567 --rc genhtml_function_coverage=1 00:07:11.567 --rc genhtml_legend=1 00:07:11.567 --rc geninfo_all_blocks=1 00:07:11.567 --rc geninfo_unexecuted_blocks=1 00:07:11.567 00:07:11.567 ' 00:07:11.567 17:28:42 version -- common/autotest_common.sh@1695 -- # LCOV='lcov 00:07:11.567 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:11.567 --rc genhtml_branch_coverage=1 00:07:11.567 --rc genhtml_function_coverage=1 00:07:11.567 --rc genhtml_legend=1 00:07:11.567 --rc geninfo_all_blocks=1 00:07:11.567 --rc geninfo_unexecuted_blocks=1 00:07:11.567 00:07:11.567 ' 00:07:11.567 17:28:42 version -- app/version.sh@17 -- # get_header_version major 00:07:11.567 17:28:42 version -- app/version.sh@13 -- # grep -E '^#define SPDK_VERSION_MAJOR[[:space:]]+' /home/vagrant/spdk_repo/spdk/include/spdk/version.h 00:07:11.567 17:28:42 version -- app/version.sh@14 -- # cut -f2 00:07:11.567 17:28:42 version -- app/version.sh@14 -- # tr -d '"' 00:07:11.567 17:28:42 version -- app/version.sh@17 -- # major=24 00:07:11.567 17:28:42 version -- app/version.sh@18 -- # get_header_version minor 00:07:11.567 17:28:42 version -- app/version.sh@13 -- # grep -E '^#define SPDK_VERSION_MINOR[[:space:]]+' /home/vagrant/spdk_repo/spdk/include/spdk/version.h 00:07:11.567 17:28:42 version -- app/version.sh@14 -- # cut -f2 00:07:11.567 17:28:42 version -- app/version.sh@14 -- # tr -d '"' 00:07:11.567 17:28:42 version -- app/version.sh@18 -- # minor=9 00:07:11.567 17:28:42 version -- app/version.sh@19 -- # get_header_version patch 00:07:11.567 17:28:42 version -- app/version.sh@13 -- # grep -E '^#define SPDK_VERSION_PATCH[[:space:]]+' /home/vagrant/spdk_repo/spdk/include/spdk/version.h 00:07:11.567 17:28:42 version -- app/version.sh@14 -- # cut -f2 00:07:11.567 17:28:42 version -- app/version.sh@14 -- # tr -d '"' 00:07:11.567 17:28:42 version -- app/version.sh@19 -- # patch=1 00:07:11.567 17:28:42 version -- app/version.sh@20 -- # get_header_version suffix 00:07:11.567 17:28:42 version -- app/version.sh@14 -- # cut -f2 00:07:11.567 17:28:42 version -- app/version.sh@13 -- # grep -E '^#define SPDK_VERSION_SUFFIX[[:space:]]+' /home/vagrant/spdk_repo/spdk/include/spdk/version.h 00:07:11.567 17:28:42 version -- app/version.sh@14 -- # tr -d '"' 00:07:11.567 17:28:42 version -- app/version.sh@20 -- # suffix=-pre 00:07:11.567 17:28:42 version -- app/version.sh@22 -- # version=24.9 00:07:11.567 17:28:42 version -- app/version.sh@25 -- # (( patch != 0 )) 00:07:11.567 17:28:42 version -- app/version.sh@25 -- # version=24.9.1 00:07:11.567 17:28:42 version -- app/version.sh@28 -- # version=24.9.1rc0 00:07:11.567 17:28:42 version -- app/version.sh@30 -- # PYTHONPATH=:/home/vagrant/spdk_repo/spdk/python:/home/vagrant/spdk_repo/spdk/test/rpc_plugins:/home/vagrant/spdk_repo/spdk/python:/home/vagrant/spdk_repo/spdk/test/rpc_plugins:/home/vagrant/spdk_repo/spdk/python 00:07:11.567 17:28:42 version -- app/version.sh@30 -- # python3 -c 'import spdk; print(spdk.__version__)' 00:07:11.828 17:28:42 version -- app/version.sh@30 -- # py_version=24.9.1rc0 00:07:11.828 17:28:42 version -- app/version.sh@31 -- # [[ 24.9.1rc0 == \2\4\.\9\.\1\r\c\0 ]] 00:07:11.828 00:07:11.828 real 0m0.314s 00:07:11.828 user 0m0.177s 00:07:11.828 sys 0m0.193s 00:07:11.828 17:28:42 version -- common/autotest_common.sh@1126 -- # xtrace_disable 00:07:11.828 17:28:42 version -- common/autotest_common.sh@10 -- # set +x 00:07:11.828 ************************************ 00:07:11.828 END TEST version 00:07:11.828 ************************************ 00:07:11.828 17:28:42 -- spdk/autotest.sh@179 -- # '[' 0 -eq 1 ']' 00:07:11.828 17:28:42 -- spdk/autotest.sh@188 -- # [[ 1 -eq 1 ]] 00:07:11.828 17:28:42 -- spdk/autotest.sh@189 -- # run_test bdev_raid /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh 00:07:11.828 17:28:42 -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:07:11.828 17:28:42 -- common/autotest_common.sh@1107 -- # xtrace_disable 00:07:11.828 17:28:42 -- common/autotest_common.sh@10 -- # set +x 00:07:11.828 ************************************ 00:07:11.828 START TEST bdev_raid 00:07:11.828 ************************************ 00:07:11.828 17:28:42 bdev_raid -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh 00:07:11.828 * Looking for test storage... 00:07:11.828 * Found test storage at /home/vagrant/spdk_repo/spdk/test/bdev 00:07:11.828 17:28:42 bdev_raid -- common/autotest_common.sh@1680 -- # [[ y == y ]] 00:07:11.828 17:28:42 bdev_raid -- common/autotest_common.sh@1681 -- # lcov --version 00:07:11.828 17:28:42 bdev_raid -- common/autotest_common.sh@1681 -- # awk '{print $NF}' 00:07:12.089 17:28:43 bdev_raid -- common/autotest_common.sh@1681 -- # lt 1.15 2 00:07:12.089 17:28:43 bdev_raid -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:07:12.089 17:28:43 bdev_raid -- scripts/common.sh@333 -- # local ver1 ver1_l 00:07:12.089 17:28:43 bdev_raid -- scripts/common.sh@334 -- # local ver2 ver2_l 00:07:12.089 17:28:43 bdev_raid -- scripts/common.sh@336 -- # IFS=.-: 00:07:12.089 17:28:43 bdev_raid -- scripts/common.sh@336 -- # read -ra ver1 00:07:12.089 17:28:43 bdev_raid -- scripts/common.sh@337 -- # IFS=.-: 00:07:12.089 17:28:43 bdev_raid -- scripts/common.sh@337 -- # read -ra ver2 00:07:12.089 17:28:43 bdev_raid -- scripts/common.sh@338 -- # local 'op=<' 00:07:12.089 17:28:43 bdev_raid -- scripts/common.sh@340 -- # ver1_l=2 00:07:12.089 17:28:43 bdev_raid -- scripts/common.sh@341 -- # ver2_l=1 00:07:12.089 17:28:43 bdev_raid -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:07:12.089 17:28:43 bdev_raid -- scripts/common.sh@344 -- # case "$op" in 00:07:12.089 17:28:43 bdev_raid -- scripts/common.sh@345 -- # : 1 00:07:12.089 17:28:43 bdev_raid -- scripts/common.sh@364 -- # (( v = 0 )) 00:07:12.089 17:28:43 bdev_raid -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:07:12.089 17:28:43 bdev_raid -- scripts/common.sh@365 -- # decimal 1 00:07:12.089 17:28:43 bdev_raid -- scripts/common.sh@353 -- # local d=1 00:07:12.089 17:28:43 bdev_raid -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:07:12.089 17:28:43 bdev_raid -- scripts/common.sh@355 -- # echo 1 00:07:12.089 17:28:43 bdev_raid -- scripts/common.sh@365 -- # ver1[v]=1 00:07:12.089 17:28:43 bdev_raid -- scripts/common.sh@366 -- # decimal 2 00:07:12.089 17:28:43 bdev_raid -- scripts/common.sh@353 -- # local d=2 00:07:12.089 17:28:43 bdev_raid -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:07:12.089 17:28:43 bdev_raid -- scripts/common.sh@355 -- # echo 2 00:07:12.089 17:28:43 bdev_raid -- scripts/common.sh@366 -- # ver2[v]=2 00:07:12.089 17:28:43 bdev_raid -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:07:12.089 17:28:43 bdev_raid -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:07:12.089 17:28:43 bdev_raid -- scripts/common.sh@368 -- # return 0 00:07:12.089 17:28:43 bdev_raid -- common/autotest_common.sh@1682 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:07:12.089 17:28:43 bdev_raid -- common/autotest_common.sh@1694 -- # export 'LCOV_OPTS= 00:07:12.089 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:12.089 --rc genhtml_branch_coverage=1 00:07:12.089 --rc genhtml_function_coverage=1 00:07:12.089 --rc genhtml_legend=1 00:07:12.089 --rc geninfo_all_blocks=1 00:07:12.089 --rc geninfo_unexecuted_blocks=1 00:07:12.089 00:07:12.089 ' 00:07:12.089 17:28:43 bdev_raid -- common/autotest_common.sh@1694 -- # LCOV_OPTS=' 00:07:12.089 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:12.089 --rc genhtml_branch_coverage=1 00:07:12.089 --rc genhtml_function_coverage=1 00:07:12.089 --rc genhtml_legend=1 00:07:12.089 --rc geninfo_all_blocks=1 00:07:12.089 --rc geninfo_unexecuted_blocks=1 00:07:12.089 00:07:12.089 ' 00:07:12.089 17:28:43 bdev_raid -- common/autotest_common.sh@1695 -- # export 'LCOV=lcov 00:07:12.089 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:12.089 --rc genhtml_branch_coverage=1 00:07:12.089 --rc genhtml_function_coverage=1 00:07:12.089 --rc genhtml_legend=1 00:07:12.089 --rc geninfo_all_blocks=1 00:07:12.089 --rc geninfo_unexecuted_blocks=1 00:07:12.089 00:07:12.089 ' 00:07:12.089 17:28:43 bdev_raid -- common/autotest_common.sh@1695 -- # LCOV='lcov 00:07:12.089 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:12.089 --rc genhtml_branch_coverage=1 00:07:12.089 --rc genhtml_function_coverage=1 00:07:12.089 --rc genhtml_legend=1 00:07:12.089 --rc geninfo_all_blocks=1 00:07:12.089 --rc geninfo_unexecuted_blocks=1 00:07:12.089 00:07:12.089 ' 00:07:12.089 17:28:43 bdev_raid -- bdev/bdev_raid.sh@12 -- # source /home/vagrant/spdk_repo/spdk/test/bdev/nbd_common.sh 00:07:12.089 17:28:43 bdev_raid -- bdev/nbd_common.sh@6 -- # set -e 00:07:12.089 17:28:43 bdev_raid -- bdev/bdev_raid.sh@14 -- # rpc_py=rpc_cmd 00:07:12.089 17:28:43 bdev_raid -- bdev/bdev_raid.sh@946 -- # mkdir -p /raidtest 00:07:12.089 17:28:43 bdev_raid -- bdev/bdev_raid.sh@947 -- # trap 'cleanup; exit 1' EXIT 00:07:12.089 17:28:43 bdev_raid -- bdev/bdev_raid.sh@949 -- # base_blocklen=512 00:07:12.089 17:28:43 bdev_raid -- bdev/bdev_raid.sh@951 -- # run_test raid1_resize_data_offset_test raid_resize_data_offset_test 00:07:12.089 17:28:43 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:07:12.089 17:28:43 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:07:12.089 17:28:43 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:12.089 ************************************ 00:07:12.089 START TEST raid1_resize_data_offset_test 00:07:12.089 ************************************ 00:07:12.089 17:28:43 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@1125 -- # raid_resize_data_offset_test 00:07:12.089 17:28:43 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@917 -- # raid_pid=71354 00:07:12.089 Process raid pid: 71354 00:07:12.089 17:28:43 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@918 -- # echo 'Process raid pid: 71354' 00:07:12.089 17:28:43 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@919 -- # waitforlisten 71354 00:07:12.089 17:28:43 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@916 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:07:12.089 17:28:43 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@831 -- # '[' -z 71354 ']' 00:07:12.089 17:28:43 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:12.089 17:28:43 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:07:12.089 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:12.089 17:28:43 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:12.089 17:28:43 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:07:12.089 17:28:43 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:07:12.089 [2024-11-27 17:28:43.168397] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:07:12.089 [2024-11-27 17:28:43.168519] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:07:12.349 [2024-11-27 17:28:43.315064] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:12.349 [2024-11-27 17:28:43.388829] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:07:12.349 [2024-11-27 17:28:43.466887] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:12.349 [2024-11-27 17:28:43.466928] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:12.920 17:28:43 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:07:12.920 17:28:43 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@864 -- # return 0 00:07:12.920 17:28:43 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@922 -- # rpc_cmd bdev_malloc_create -b malloc0 64 512 -o 16 00:07:12.920 17:28:43 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:12.920 17:28:43 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:07:12.920 malloc0 00:07:12.920 17:28:44 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:12.920 17:28:44 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@923 -- # rpc_cmd bdev_malloc_create -b malloc1 64 512 -o 16 00:07:12.920 17:28:44 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:12.920 17:28:44 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:07:12.920 malloc1 00:07:12.920 17:28:44 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:12.920 17:28:44 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@924 -- # rpc_cmd bdev_null_create null0 64 512 00:07:12.920 17:28:44 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:12.920 17:28:44 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:07:12.920 null0 00:07:12.920 17:28:44 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:12.920 17:28:44 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@926 -- # rpc_cmd bdev_raid_create -n Raid -r 1 -b ''\''malloc0 malloc1 null0'\''' -s 00:07:12.920 17:28:44 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:12.920 17:28:44 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:07:12.920 [2024-11-27 17:28:44.077575] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc0 is claimed 00:07:12.920 [2024-11-27 17:28:44.079739] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:07:12.920 [2024-11-27 17:28:44.079790] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev null0 is claimed 00:07:12.920 [2024-11-27 17:28:44.079956] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:07:12.920 [2024-11-27 17:28:44.079978] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 129024, blocklen 512 00:07:12.920 [2024-11-27 17:28:44.080280] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000021f0 00:07:12.920 [2024-11-27 17:28:44.080414] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:07:12.920 [2024-11-27 17:28:44.080439] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Raid, raid_bdev 0x617000001200 00:07:12.920 [2024-11-27 17:28:44.080595] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:12.920 17:28:44 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:12.920 17:28:44 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@929 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:12.920 17:28:44 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:12.920 17:28:44 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@929 -- # jq -r '.[].base_bdevs_list[2].data_offset' 00:07:12.920 17:28:44 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:07:12.920 17:28:44 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:13.180 17:28:44 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@929 -- # (( 2048 == 2048 )) 00:07:13.180 17:28:44 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@931 -- # rpc_cmd bdev_null_delete null0 00:07:13.180 17:28:44 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:13.180 17:28:44 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:07:13.180 [2024-11-27 17:28:44.137456] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: null0 00:07:13.180 17:28:44 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:13.180 17:28:44 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@935 -- # rpc_cmd bdev_malloc_create -b malloc2 512 512 -o 30 00:07:13.180 17:28:44 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:13.180 17:28:44 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:07:13.180 malloc2 00:07:13.180 17:28:44 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:13.180 17:28:44 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@936 -- # rpc_cmd bdev_raid_add_base_bdev Raid malloc2 00:07:13.180 17:28:44 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:13.180 17:28:44 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:07:13.180 [2024-11-27 17:28:44.354527] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:07:13.180 [2024-11-27 17:28:44.362234] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:07:13.180 17:28:44 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:13.180 [2024-11-27 17:28:44.364466] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev Raid 00:07:13.180 17:28:44 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@939 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:13.180 17:28:44 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:13.180 17:28:44 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@939 -- # jq -r '.[].base_bdevs_list[2].data_offset' 00:07:13.180 17:28:44 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:07:13.440 17:28:44 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:13.440 17:28:44 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@939 -- # (( 2070 == 2070 )) 00:07:13.440 17:28:44 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@941 -- # killprocess 71354 00:07:13.440 17:28:44 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@950 -- # '[' -z 71354 ']' 00:07:13.440 17:28:44 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@954 -- # kill -0 71354 00:07:13.440 17:28:44 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@955 -- # uname 00:07:13.440 17:28:44 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:07:13.440 17:28:44 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 71354 00:07:13.440 17:28:44 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:07:13.440 17:28:44 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:07:13.440 killing process with pid 71354 00:07:13.441 17:28:44 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 71354' 00:07:13.441 17:28:44 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@969 -- # kill 71354 00:07:13.441 17:28:44 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@974 -- # wait 71354 00:07:13.441 [2024-11-27 17:28:44.459184] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:13.441 [2024-11-27 17:28:44.460685] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev Raid: Operation canceled 00:07:13.441 [2024-11-27 17:28:44.460773] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:13.441 [2024-11-27 17:28:44.460793] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: malloc2 00:07:13.441 [2024-11-27 17:28:44.469278] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:13.441 [2024-11-27 17:28:44.469621] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:13.441 [2024-11-27 17:28:44.469645] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Raid, state offline 00:07:13.701 [2024-11-27 17:28:44.864879] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:14.272 17:28:45 bdev_raid.raid1_resize_data_offset_test -- bdev/bdev_raid.sh@943 -- # return 0 00:07:14.272 00:07:14.272 real 0m2.151s 00:07:14.272 user 0m1.949s 00:07:14.272 sys 0m0.638s 00:07:14.272 17:28:45 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:07:14.272 17:28:45 bdev_raid.raid1_resize_data_offset_test -- common/autotest_common.sh@10 -- # set +x 00:07:14.272 ************************************ 00:07:14.272 END TEST raid1_resize_data_offset_test 00:07:14.272 ************************************ 00:07:14.272 17:28:45 bdev_raid -- bdev/bdev_raid.sh@953 -- # run_test raid0_resize_superblock_test raid_resize_superblock_test 0 00:07:14.272 17:28:45 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:07:14.272 17:28:45 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:07:14.272 17:28:45 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:14.272 ************************************ 00:07:14.272 START TEST raid0_resize_superblock_test 00:07:14.272 ************************************ 00:07:14.272 17:28:45 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@1125 -- # raid_resize_superblock_test 0 00:07:14.272 17:28:45 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@854 -- # local raid_level=0 00:07:14.272 17:28:45 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@857 -- # raid_pid=71416 00:07:14.272 17:28:45 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@858 -- # echo 'Process raid pid: 71416' 00:07:14.272 17:28:45 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@856 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:07:14.272 Process raid pid: 71416 00:07:14.272 17:28:45 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@859 -- # waitforlisten 71416 00:07:14.272 17:28:45 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@831 -- # '[' -z 71416 ']' 00:07:14.272 17:28:45 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:14.272 17:28:45 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:07:14.272 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:14.272 17:28:45 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:14.272 17:28:45 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:07:14.272 17:28:45 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:14.272 [2024-11-27 17:28:45.395379] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:07:14.272 [2024-11-27 17:28:45.395510] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:07:14.532 [2024-11-27 17:28:45.540102] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:14.532 [2024-11-27 17:28:45.606982] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:07:14.532 [2024-11-27 17:28:45.681778] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:14.532 [2024-11-27 17:28:45.681825] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:15.102 17:28:46 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:07:15.102 17:28:46 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@864 -- # return 0 00:07:15.102 17:28:46 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@861 -- # rpc_cmd bdev_malloc_create -b malloc0 512 512 00:07:15.102 17:28:46 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:15.102 17:28:46 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:15.362 malloc0 00:07:15.362 17:28:46 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:15.362 17:28:46 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@863 -- # rpc_cmd bdev_passthru_create -b malloc0 -p pt0 00:07:15.362 17:28:46 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:15.362 17:28:46 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:15.362 [2024-11-27 17:28:46.447668] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc0 00:07:15.362 [2024-11-27 17:28:46.447743] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:15.362 [2024-11-27 17:28:46.447775] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:07:15.362 [2024-11-27 17:28:46.447796] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:15.362 [2024-11-27 17:28:46.450238] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:15.362 [2024-11-27 17:28:46.450273] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt0 00:07:15.362 pt0 00:07:15.362 17:28:46 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:15.362 17:28:46 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@864 -- # rpc_cmd bdev_lvol_create_lvstore pt0 lvs0 00:07:15.362 17:28:46 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:15.362 17:28:46 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:15.623 ec95ceb1-db17-42c7-aa02-93c1003f91c7 00:07:15.623 17:28:46 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:15.623 17:28:46 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@866 -- # rpc_cmd bdev_lvol_create -l lvs0 lvol0 64 00:07:15.623 17:28:46 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:15.623 17:28:46 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:15.623 145692cf-349a-4c64-90f1-5dc2c418a1fe 00:07:15.623 17:28:46 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:15.623 17:28:46 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@867 -- # rpc_cmd bdev_lvol_create -l lvs0 lvol1 64 00:07:15.623 17:28:46 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:15.623 17:28:46 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:15.623 a4fa57da-5302-4ff5-b9c3-e96bf7128347 00:07:15.623 17:28:46 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:15.623 17:28:46 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@869 -- # case $raid_level in 00:07:15.623 17:28:46 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@870 -- # rpc_cmd bdev_raid_create -n Raid -r 0 -z 64 -b ''\''lvs0/lvol0 lvs0/lvol1'\''' -s 00:07:15.623 17:28:46 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:15.623 17:28:46 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:15.623 [2024-11-27 17:28:46.655844] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev 145692cf-349a-4c64-90f1-5dc2c418a1fe is claimed 00:07:15.623 [2024-11-27 17:28:46.655934] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev a4fa57da-5302-4ff5-b9c3-e96bf7128347 is claimed 00:07:15.623 [2024-11-27 17:28:46.656053] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:07:15.623 [2024-11-27 17:28:46.656067] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 245760, blocklen 512 00:07:15.623 [2024-11-27 17:28:46.656365] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:07:15.623 [2024-11-27 17:28:46.656509] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:07:15.623 [2024-11-27 17:28:46.656530] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Raid, raid_bdev 0x617000001200 00:07:15.623 [2024-11-27 17:28:46.656668] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:15.623 17:28:46 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:15.623 17:28:46 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@875 -- # rpc_cmd bdev_get_bdevs -b lvs0/lvol0 00:07:15.623 17:28:46 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:15.623 17:28:46 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@875 -- # jq '.[].num_blocks' 00:07:15.623 17:28:46 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:15.623 17:28:46 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:15.623 17:28:46 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@875 -- # (( 64 == 64 )) 00:07:15.623 17:28:46 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@876 -- # rpc_cmd bdev_get_bdevs -b lvs0/lvol1 00:07:15.623 17:28:46 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@876 -- # jq '.[].num_blocks' 00:07:15.623 17:28:46 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:15.623 17:28:46 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:15.623 17:28:46 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:15.623 17:28:46 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@876 -- # (( 64 == 64 )) 00:07:15.623 17:28:46 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@879 -- # case $raid_level in 00:07:15.623 17:28:46 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@880 -- # rpc_cmd bdev_get_bdevs -b Raid 00:07:15.623 17:28:46 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:15.623 17:28:46 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:15.623 17:28:46 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@879 -- # case $raid_level in 00:07:15.623 17:28:46 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@880 -- # jq '.[].num_blocks' 00:07:15.623 [2024-11-27 17:28:46.767843] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:15.623 17:28:46 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:15.623 17:28:46 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@879 -- # case $raid_level in 00:07:15.623 17:28:46 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@879 -- # case $raid_level in 00:07:15.623 17:28:46 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@880 -- # (( 245760 == 245760 )) 00:07:15.623 17:28:46 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@885 -- # rpc_cmd bdev_lvol_resize lvs0/lvol0 100 00:07:15.623 17:28:46 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:15.623 17:28:46 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:15.623 [2024-11-27 17:28:46.811690] bdev_raid.c:2313:raid_bdev_resize_base_bdev: *DEBUG*: raid_bdev_resize_base_bdev 00:07:15.623 [2024-11-27 17:28:46.811717] bdev_raid.c:2326:raid_bdev_resize_base_bdev: *NOTICE*: base_bdev '145692cf-349a-4c64-90f1-5dc2c418a1fe' was resized: old size 131072, new size 204800 00:07:15.884 17:28:46 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:15.884 17:28:46 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@886 -- # rpc_cmd bdev_lvol_resize lvs0/lvol1 100 00:07:15.884 17:28:46 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:15.884 17:28:46 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:15.884 [2024-11-27 17:28:46.823611] bdev_raid.c:2313:raid_bdev_resize_base_bdev: *DEBUG*: raid_bdev_resize_base_bdev 00:07:15.884 [2024-11-27 17:28:46.823634] bdev_raid.c:2326:raid_bdev_resize_base_bdev: *NOTICE*: base_bdev 'a4fa57da-5302-4ff5-b9c3-e96bf7128347' was resized: old size 131072, new size 204800 00:07:15.884 [2024-11-27 17:28:46.823658] bdev_raid.c:2340:raid_bdev_resize_base_bdev: *NOTICE*: raid bdev 'Raid': block count was changed from 245760 to 393216 00:07:15.884 17:28:46 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:15.884 17:28:46 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@889 -- # rpc_cmd bdev_get_bdevs -b lvs0/lvol0 00:07:15.884 17:28:46 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:15.884 17:28:46 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:15.884 17:28:46 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@889 -- # jq '.[].num_blocks' 00:07:15.884 17:28:46 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:15.884 17:28:46 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@889 -- # (( 100 == 100 )) 00:07:15.884 17:28:46 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@890 -- # rpc_cmd bdev_get_bdevs -b lvs0/lvol1 00:07:15.885 17:28:46 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:15.885 17:28:46 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:15.885 17:28:46 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@890 -- # jq '.[].num_blocks' 00:07:15.885 17:28:46 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:15.885 17:28:46 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@890 -- # (( 100 == 100 )) 00:07:15.885 17:28:46 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@893 -- # case $raid_level in 00:07:15.885 17:28:46 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@894 -- # rpc_cmd bdev_get_bdevs -b Raid 00:07:15.885 17:28:46 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:15.885 17:28:46 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:15.885 17:28:46 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@893 -- # case $raid_level in 00:07:15.885 17:28:46 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@894 -- # jq '.[].num_blocks' 00:07:15.885 [2024-11-27 17:28:46.935516] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:15.885 17:28:46 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:15.885 17:28:46 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@893 -- # case $raid_level in 00:07:15.885 17:28:46 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@893 -- # case $raid_level in 00:07:15.885 17:28:46 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@894 -- # (( 393216 == 393216 )) 00:07:15.885 17:28:46 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@898 -- # rpc_cmd bdev_passthru_delete pt0 00:07:15.885 17:28:46 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:15.885 17:28:46 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:15.885 [2024-11-27 17:28:46.979283] vbdev_lvol.c: 150:vbdev_lvs_hotremove_cb: *NOTICE*: bdev pt0 being removed: closing lvstore lvs0 00:07:15.885 [2024-11-27 17:28:46.979341] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: lvs0/lvol0 00:07:15.885 [2024-11-27 17:28:46.979363] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:15.885 [2024-11-27 17:28:46.979375] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: lvs0/lvol1 00:07:15.885 [2024-11-27 17:28:46.979482] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:15.885 [2024-11-27 17:28:46.979514] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:15.885 [2024-11-27 17:28:46.979526] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Raid, state offline 00:07:15.885 17:28:46 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:15.885 17:28:46 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@899 -- # rpc_cmd bdev_passthru_create -b malloc0 -p pt0 00:07:15.885 17:28:46 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:15.885 17:28:46 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:15.885 [2024-11-27 17:28:46.991235] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc0 00:07:15.885 [2024-11-27 17:28:46.991280] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:15.885 [2024-11-27 17:28:46.991298] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:07:15.885 [2024-11-27 17:28:46.991309] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:15.885 [2024-11-27 17:28:46.993716] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:15.885 [2024-11-27 17:28:46.993748] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt0 00:07:15.885 [2024-11-27 17:28:46.995147] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev 145692cf-349a-4c64-90f1-5dc2c418a1fe 00:07:15.885 [2024-11-27 17:28:46.995229] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev 145692cf-349a-4c64-90f1-5dc2c418a1fe is claimed 00:07:15.885 [2024-11-27 17:28:46.995327] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev a4fa57da-5302-4ff5-b9c3-e96bf7128347 00:07:15.885 [2024-11-27 17:28:46.995351] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev a4fa57da-5302-4ff5-b9c3-e96bf7128347 is claimed 00:07:15.885 [2024-11-27 17:28:46.995482] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev a4fa57da-5302-4ff5-b9c3-e96bf7128347 (2) smaller than existing raid bdev Raid (3) 00:07:15.885 [2024-11-27 17:28:46.995520] bdev_raid.c:3884:raid_bdev_examine_done: *ERROR*: Failed to examine bdev 145692cf-349a-4c64-90f1-5dc2c418a1fe: File exists 00:07:15.885 [2024-11-27 17:28:46.995557] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001580 00:07:15.885 [2024-11-27 17:28:46.995566] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 393216, blocklen 512 00:07:15.885 [2024-11-27 17:28:46.995810] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000026d0 00:07:15.885 [2024-11-27 17:28:46.995964] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001580 00:07:15.885 [2024-11-27 17:28:46.995980] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Raid, raid_bdev 0x617000001580 00:07:15.885 [2024-11-27 17:28:46.996091] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:15.885 pt0 00:07:15.885 17:28:46 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:15.885 17:28:46 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@900 -- # rpc_cmd bdev_wait_for_examine 00:07:15.885 17:28:46 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:15.885 17:28:46 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:15.885 17:28:47 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:15.885 17:28:47 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@904 -- # case $raid_level in 00:07:15.885 17:28:47 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@905 -- # jq '.[].num_blocks' 00:07:15.885 17:28:47 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@904 -- # case $raid_level in 00:07:15.885 17:28:47 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@905 -- # rpc_cmd bdev_get_bdevs -b Raid 00:07:15.885 17:28:47 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:15.885 17:28:47 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:15.885 [2024-11-27 17:28:47.019405] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:15.885 17:28:47 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:15.885 17:28:47 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@904 -- # case $raid_level in 00:07:15.885 17:28:47 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@904 -- # case $raid_level in 00:07:15.885 17:28:47 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@905 -- # (( 393216 == 393216 )) 00:07:15.885 17:28:47 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@909 -- # killprocess 71416 00:07:15.885 17:28:47 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@950 -- # '[' -z 71416 ']' 00:07:15.885 17:28:47 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@954 -- # kill -0 71416 00:07:15.885 17:28:47 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@955 -- # uname 00:07:15.885 17:28:47 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:07:15.885 17:28:47 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 71416 00:07:16.146 17:28:47 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:07:16.146 17:28:47 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:07:16.146 killing process with pid 71416 00:07:16.146 17:28:47 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 71416' 00:07:16.146 17:28:47 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@969 -- # kill 71416 00:07:16.146 [2024-11-27 17:28:47.097777] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:16.146 [2024-11-27 17:28:47.097847] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:16.146 [2024-11-27 17:28:47.097887] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:16.146 [2024-11-27 17:28:47.097895] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Raid, state offline 00:07:16.146 17:28:47 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@974 -- # wait 71416 00:07:16.408 [2024-11-27 17:28:47.400856] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:16.668 17:28:47 bdev_raid.raid0_resize_superblock_test -- bdev/bdev_raid.sh@911 -- # return 0 00:07:16.668 00:07:16.668 real 0m2.454s 00:07:16.668 user 0m2.549s 00:07:16.668 sys 0m0.657s 00:07:16.668 17:28:47 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:07:16.668 17:28:47 bdev_raid.raid0_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:16.668 ************************************ 00:07:16.668 END TEST raid0_resize_superblock_test 00:07:16.668 ************************************ 00:07:16.669 17:28:47 bdev_raid -- bdev/bdev_raid.sh@954 -- # run_test raid1_resize_superblock_test raid_resize_superblock_test 1 00:07:16.669 17:28:47 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:07:16.669 17:28:47 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:07:16.669 17:28:47 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:16.669 ************************************ 00:07:16.669 START TEST raid1_resize_superblock_test 00:07:16.669 ************************************ 00:07:16.669 17:28:47 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@1125 -- # raid_resize_superblock_test 1 00:07:16.669 17:28:47 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@854 -- # local raid_level=1 00:07:16.669 17:28:47 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@857 -- # raid_pid=71487 00:07:16.669 Process raid pid: 71487 00:07:16.669 17:28:47 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@856 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:07:16.669 17:28:47 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@858 -- # echo 'Process raid pid: 71487' 00:07:16.669 17:28:47 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@859 -- # waitforlisten 71487 00:07:16.669 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:16.669 17:28:47 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@831 -- # '[' -z 71487 ']' 00:07:16.669 17:28:47 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:16.669 17:28:47 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:07:16.669 17:28:47 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:16.669 17:28:47 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:07:16.669 17:28:47 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:16.928 [2024-11-27 17:28:47.925815] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:07:16.928 [2024-11-27 17:28:47.926012] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:07:16.928 [2024-11-27 17:28:48.073927] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:17.187 [2024-11-27 17:28:48.144681] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:07:17.187 [2024-11-27 17:28:48.223400] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:17.187 [2024-11-27 17:28:48.223467] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:17.758 17:28:48 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:07:17.758 17:28:48 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@864 -- # return 0 00:07:17.758 17:28:48 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@861 -- # rpc_cmd bdev_malloc_create -b malloc0 512 512 00:07:17.758 17:28:48 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:17.758 17:28:48 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:18.018 malloc0 00:07:18.018 17:28:48 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:18.018 17:28:48 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@863 -- # rpc_cmd bdev_passthru_create -b malloc0 -p pt0 00:07:18.018 17:28:48 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:18.018 17:28:48 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:18.018 [2024-11-27 17:28:48.960171] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc0 00:07:18.018 [2024-11-27 17:28:48.960237] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:18.018 [2024-11-27 17:28:48.960275] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:07:18.018 [2024-11-27 17:28:48.960296] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:18.018 [2024-11-27 17:28:48.962707] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:18.018 [2024-11-27 17:28:48.962752] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt0 00:07:18.018 pt0 00:07:18.018 17:28:48 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:18.018 17:28:48 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@864 -- # rpc_cmd bdev_lvol_create_lvstore pt0 lvs0 00:07:18.018 17:28:48 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:18.018 17:28:48 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:18.018 085caa2d-6f72-41eb-b616-e75d850fcefb 00:07:18.018 17:28:49 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:18.018 17:28:49 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@866 -- # rpc_cmd bdev_lvol_create -l lvs0 lvol0 64 00:07:18.018 17:28:49 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:18.018 17:28:49 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:18.018 47611c10-482f-44e5-ac82-952c11cce073 00:07:18.018 17:28:49 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:18.019 17:28:49 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@867 -- # rpc_cmd bdev_lvol_create -l lvs0 lvol1 64 00:07:18.019 17:28:49 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:18.019 17:28:49 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:18.019 4843f88d-c8e7-403e-acfc-7231a0fe28a0 00:07:18.019 17:28:49 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:18.019 17:28:49 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@869 -- # case $raid_level in 00:07:18.019 17:28:49 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@871 -- # rpc_cmd bdev_raid_create -n Raid -r 1 -b ''\''lvs0/lvol0 lvs0/lvol1'\''' -s 00:07:18.019 17:28:49 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:18.019 17:28:49 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:18.019 [2024-11-27 17:28:49.167321] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev 47611c10-482f-44e5-ac82-952c11cce073 is claimed 00:07:18.019 [2024-11-27 17:28:49.167419] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev 4843f88d-c8e7-403e-acfc-7231a0fe28a0 is claimed 00:07:18.019 [2024-11-27 17:28:49.167535] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:07:18.019 [2024-11-27 17:28:49.167548] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 122880, blocklen 512 00:07:18.019 [2024-11-27 17:28:49.167827] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:07:18.019 [2024-11-27 17:28:49.167982] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:07:18.019 [2024-11-27 17:28:49.167994] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Raid, raid_bdev 0x617000001200 00:07:18.019 [2024-11-27 17:28:49.168134] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:18.019 17:28:49 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:18.019 17:28:49 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@875 -- # rpc_cmd bdev_get_bdevs -b lvs0/lvol0 00:07:18.019 17:28:49 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:18.019 17:28:49 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@875 -- # jq '.[].num_blocks' 00:07:18.019 17:28:49 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:18.019 17:28:49 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:18.287 17:28:49 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@875 -- # (( 64 == 64 )) 00:07:18.287 17:28:49 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@876 -- # jq '.[].num_blocks' 00:07:18.287 17:28:49 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@876 -- # rpc_cmd bdev_get_bdevs -b lvs0/lvol1 00:07:18.287 17:28:49 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:18.287 17:28:49 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:18.287 17:28:49 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:18.287 17:28:49 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@876 -- # (( 64 == 64 )) 00:07:18.287 17:28:49 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@879 -- # case $raid_level in 00:07:18.287 17:28:49 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@881 -- # jq '.[].num_blocks' 00:07:18.287 17:28:49 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@879 -- # case $raid_level in 00:07:18.287 17:28:49 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@881 -- # rpc_cmd bdev_get_bdevs -b Raid 00:07:18.287 17:28:49 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:18.287 17:28:49 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:18.287 [2024-11-27 17:28:49.283322] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:18.287 17:28:49 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:18.287 17:28:49 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@879 -- # case $raid_level in 00:07:18.287 17:28:49 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@879 -- # case $raid_level in 00:07:18.287 17:28:49 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@881 -- # (( 122880 == 122880 )) 00:07:18.287 17:28:49 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@885 -- # rpc_cmd bdev_lvol_resize lvs0/lvol0 100 00:07:18.287 17:28:49 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:18.287 17:28:49 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:18.287 [2024-11-27 17:28:49.327175] bdev_raid.c:2313:raid_bdev_resize_base_bdev: *DEBUG*: raid_bdev_resize_base_bdev 00:07:18.287 [2024-11-27 17:28:49.327240] bdev_raid.c:2326:raid_bdev_resize_base_bdev: *NOTICE*: base_bdev '47611c10-482f-44e5-ac82-952c11cce073' was resized: old size 131072, new size 204800 00:07:18.287 17:28:49 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:18.287 17:28:49 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@886 -- # rpc_cmd bdev_lvol_resize lvs0/lvol1 100 00:07:18.287 17:28:49 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:18.287 17:28:49 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:18.287 [2024-11-27 17:28:49.339080] bdev_raid.c:2313:raid_bdev_resize_base_bdev: *DEBUG*: raid_bdev_resize_base_bdev 00:07:18.287 [2024-11-27 17:28:49.339162] bdev_raid.c:2326:raid_bdev_resize_base_bdev: *NOTICE*: base_bdev '4843f88d-c8e7-403e-acfc-7231a0fe28a0' was resized: old size 131072, new size 204800 00:07:18.287 [2024-11-27 17:28:49.339193] bdev_raid.c:2340:raid_bdev_resize_base_bdev: *NOTICE*: raid bdev 'Raid': block count was changed from 122880 to 196608 00:07:18.287 17:28:49 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:18.287 17:28:49 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@889 -- # rpc_cmd bdev_get_bdevs -b lvs0/lvol0 00:07:18.287 17:28:49 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@889 -- # jq '.[].num_blocks' 00:07:18.287 17:28:49 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:18.287 17:28:49 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:18.287 17:28:49 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:18.287 17:28:49 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@889 -- # (( 100 == 100 )) 00:07:18.287 17:28:49 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@890 -- # jq '.[].num_blocks' 00:07:18.287 17:28:49 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@890 -- # rpc_cmd bdev_get_bdevs -b lvs0/lvol1 00:07:18.287 17:28:49 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:18.287 17:28:49 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:18.287 17:28:49 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:18.287 17:28:49 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@890 -- # (( 100 == 100 )) 00:07:18.287 17:28:49 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@893 -- # case $raid_level in 00:07:18.287 17:28:49 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@895 -- # jq '.[].num_blocks' 00:07:18.287 17:28:49 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@893 -- # case $raid_level in 00:07:18.287 17:28:49 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@895 -- # rpc_cmd bdev_get_bdevs -b Raid 00:07:18.287 17:28:49 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:18.287 17:28:49 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:18.287 [2024-11-27 17:28:49.431060] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:18.287 17:28:49 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:18.287 17:28:49 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@893 -- # case $raid_level in 00:07:18.287 17:28:49 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@893 -- # case $raid_level in 00:07:18.287 17:28:49 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@895 -- # (( 196608 == 196608 )) 00:07:18.287 17:28:49 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@898 -- # rpc_cmd bdev_passthru_delete pt0 00:07:18.287 17:28:49 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:18.287 17:28:49 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:18.287 [2024-11-27 17:28:49.470863] vbdev_lvol.c: 150:vbdev_lvs_hotremove_cb: *NOTICE*: bdev pt0 being removed: closing lvstore lvs0 00:07:18.287 [2024-11-27 17:28:49.470966] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: lvs0/lvol0 00:07:18.287 [2024-11-27 17:28:49.471008] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: lvs0/lvol1 00:07:18.287 [2024-11-27 17:28:49.471211] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:18.287 [2024-11-27 17:28:49.471390] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:18.287 [2024-11-27 17:28:49.471478] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:18.287 [2024-11-27 17:28:49.471532] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Raid, state offline 00:07:18.287 17:28:49 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:18.287 17:28:49 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@899 -- # rpc_cmd bdev_passthru_create -b malloc0 -p pt0 00:07:18.287 17:28:49 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:18.287 17:28:49 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:18.548 [2024-11-27 17:28:49.482803] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc0 00:07:18.548 [2024-11-27 17:28:49.482887] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:18.548 [2024-11-27 17:28:49.482921] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007580 00:07:18.548 [2024-11-27 17:28:49.482934] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:18.548 [2024-11-27 17:28:49.485300] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:18.548 [2024-11-27 17:28:49.485342] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt0 00:07:18.548 [2024-11-27 17:28:49.486659] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev 47611c10-482f-44e5-ac82-952c11cce073 00:07:18.548 [2024-11-27 17:28:49.486710] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev 47611c10-482f-44e5-ac82-952c11cce073 is claimed 00:07:18.548 [2024-11-27 17:28:49.486806] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev 4843f88d-c8e7-403e-acfc-7231a0fe28a0 00:07:18.548 [2024-11-27 17:28:49.486830] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev 4843f88d-c8e7-403e-acfc-7231a0fe28a0 is claimed 00:07:18.548 [2024-11-27 17:28:49.486908] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev 4843f88d-c8e7-403e-acfc-7231a0fe28a0 (2) smaller than existing raid bdev Raid (3) 00:07:18.548 [2024-11-27 17:28:49.486929] bdev_raid.c:3884:raid_bdev_examine_done: *ERROR*: Failed to examine bdev 47611c10-482f-44e5-ac82-952c11cce073: File exists 00:07:18.548 [2024-11-27 17:28:49.486970] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001580 00:07:18.548 [2024-11-27 17:28:49.486979] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 196608, blocklen 512 00:07:18.548 [2024-11-27 17:28:49.487265] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000026d0 00:07:18.548 [2024-11-27 17:28:49.487421] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001580 00:07:18.548 [2024-11-27 17:28:49.487439] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Raid, raid_bdev 0x617000001580 00:07:18.548 [2024-11-27 17:28:49.487588] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:18.548 pt0 00:07:18.548 17:28:49 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:18.548 17:28:49 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@900 -- # rpc_cmd bdev_wait_for_examine 00:07:18.548 17:28:49 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:18.548 17:28:49 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:18.548 17:28:49 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:18.548 17:28:49 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@904 -- # case $raid_level in 00:07:18.548 17:28:49 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@906 -- # rpc_cmd bdev_get_bdevs -b Raid 00:07:18.548 17:28:49 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:18.548 17:28:49 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:18.548 17:28:49 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@904 -- # case $raid_level in 00:07:18.548 17:28:49 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@906 -- # jq '.[].num_blocks' 00:07:18.548 [2024-11-27 17:28:49.511183] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:18.548 17:28:49 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:18.548 17:28:49 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@904 -- # case $raid_level in 00:07:18.548 17:28:49 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@904 -- # case $raid_level in 00:07:18.548 17:28:49 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@906 -- # (( 196608 == 196608 )) 00:07:18.548 17:28:49 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@909 -- # killprocess 71487 00:07:18.548 17:28:49 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@950 -- # '[' -z 71487 ']' 00:07:18.548 17:28:49 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@954 -- # kill -0 71487 00:07:18.548 17:28:49 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@955 -- # uname 00:07:18.548 17:28:49 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:07:18.548 17:28:49 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 71487 00:07:18.548 killing process with pid 71487 00:07:18.548 17:28:49 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:07:18.548 17:28:49 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:07:18.548 17:28:49 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 71487' 00:07:18.548 17:28:49 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@969 -- # kill 71487 00:07:18.548 [2024-11-27 17:28:49.592212] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:18.548 [2024-11-27 17:28:49.592275] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:18.548 [2024-11-27 17:28:49.592316] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:18.548 [2024-11-27 17:28:49.592325] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Raid, state offline 00:07:18.548 17:28:49 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@974 -- # wait 71487 00:07:18.923 [2024-11-27 17:28:49.895699] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:19.183 17:28:50 bdev_raid.raid1_resize_superblock_test -- bdev/bdev_raid.sh@911 -- # return 0 00:07:19.183 00:07:19.183 real 0m2.430s 00:07:19.183 user 0m2.496s 00:07:19.183 sys 0m0.703s 00:07:19.183 17:28:50 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:07:19.183 ************************************ 00:07:19.183 END TEST raid1_resize_superblock_test 00:07:19.183 ************************************ 00:07:19.183 17:28:50 bdev_raid.raid1_resize_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:19.183 17:28:50 bdev_raid -- bdev/bdev_raid.sh@956 -- # uname -s 00:07:19.183 17:28:50 bdev_raid -- bdev/bdev_raid.sh@956 -- # '[' Linux = Linux ']' 00:07:19.183 17:28:50 bdev_raid -- bdev/bdev_raid.sh@956 -- # modprobe -n nbd 00:07:19.183 17:28:50 bdev_raid -- bdev/bdev_raid.sh@957 -- # has_nbd=true 00:07:19.183 17:28:50 bdev_raid -- bdev/bdev_raid.sh@958 -- # modprobe nbd 00:07:19.183 17:28:50 bdev_raid -- bdev/bdev_raid.sh@959 -- # run_test raid_function_test_raid0 raid_function_test raid0 00:07:19.183 17:28:50 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:07:19.183 17:28:50 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:07:19.183 17:28:50 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:19.183 ************************************ 00:07:19.183 START TEST raid_function_test_raid0 00:07:19.183 ************************************ 00:07:19.183 17:28:50 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@1125 -- # raid_function_test raid0 00:07:19.183 17:28:50 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@64 -- # local raid_level=raid0 00:07:19.183 17:28:50 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@65 -- # local nbd=/dev/nbd0 00:07:19.183 17:28:50 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@66 -- # local raid_bdev 00:07:19.183 17:28:50 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@69 -- # raid_pid=71567 00:07:19.183 17:28:50 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@68 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:07:19.183 17:28:50 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@70 -- # echo 'Process raid pid: 71567' 00:07:19.183 Process raid pid: 71567 00:07:19.183 17:28:50 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@71 -- # waitforlisten 71567 00:07:19.183 17:28:50 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@831 -- # '[' -z 71567 ']' 00:07:19.183 17:28:50 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:19.184 17:28:50 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@836 -- # local max_retries=100 00:07:19.184 17:28:50 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:19.184 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:19.184 17:28:50 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@840 -- # xtrace_disable 00:07:19.184 17:28:50 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@10 -- # set +x 00:07:19.443 [2024-11-27 17:28:50.448289] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:07:19.443 [2024-11-27 17:28:50.448490] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:07:19.443 [2024-11-27 17:28:50.598528] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:19.704 [2024-11-27 17:28:50.666936] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:07:19.704 [2024-11-27 17:28:50.744009] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:19.704 [2024-11-27 17:28:50.744195] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:20.275 17:28:51 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:07:20.275 17:28:51 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@864 -- # return 0 00:07:20.275 17:28:51 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@73 -- # rpc_cmd bdev_malloc_create 32 512 -b Base_1 00:07:20.275 17:28:51 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:20.275 17:28:51 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@10 -- # set +x 00:07:20.275 Base_1 00:07:20.275 17:28:51 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:20.275 17:28:51 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@74 -- # rpc_cmd bdev_malloc_create 32 512 -b Base_2 00:07:20.275 17:28:51 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:20.275 17:28:51 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@10 -- # set +x 00:07:20.275 Base_2 00:07:20.275 17:28:51 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:20.275 17:28:51 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@75 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''Base_1 Base_2'\''' -n raid 00:07:20.275 17:28:51 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:20.275 17:28:51 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@10 -- # set +x 00:07:20.275 [2024-11-27 17:28:51.333076] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev Base_1 is claimed 00:07:20.275 [2024-11-27 17:28:51.336647] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev Base_2 is claimed 00:07:20.275 [2024-11-27 17:28:51.336758] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:07:20.275 [2024-11-27 17:28:51.336780] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 131072, blocklen 512 00:07:20.275 [2024-11-27 17:28:51.337299] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:07:20.275 [2024-11-27 17:28:51.337537] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:07:20.275 [2024-11-27 17:28:51.337556] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid, raid_bdev 0x617000001200 00:07:20.275 [2024-11-27 17:28:51.337894] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:20.275 17:28:51 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:20.275 17:28:51 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@77 -- # rpc_cmd bdev_raid_get_bdevs online 00:07:20.275 17:28:51 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@77 -- # jq -r '.[0]["name"] | select(.)' 00:07:20.275 17:28:51 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:20.275 17:28:51 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@10 -- # set +x 00:07:20.275 17:28:51 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:20.275 17:28:51 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@77 -- # raid_bdev=raid 00:07:20.275 17:28:51 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@78 -- # '[' raid = '' ']' 00:07:20.275 17:28:51 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@83 -- # nbd_start_disks /var/tmp/spdk.sock raid /dev/nbd0 00:07:20.275 17:28:51 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:07:20.275 17:28:51 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@10 -- # bdev_list=('raid') 00:07:20.275 17:28:51 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@10 -- # local bdev_list 00:07:20.275 17:28:51 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:07:20.275 17:28:51 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@11 -- # local nbd_list 00:07:20.275 17:28:51 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@12 -- # local i 00:07:20.275 17:28:51 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:07:20.275 17:28:51 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:07:20.275 17:28:51 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid /dev/nbd0 00:07:20.535 [2024-11-27 17:28:51.581460] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:07:20.535 /dev/nbd0 00:07:20.535 17:28:51 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:07:20.535 17:28:51 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:07:20.535 17:28:51 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:07:20.535 17:28:51 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@869 -- # local i 00:07:20.535 17:28:51 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:07:20.535 17:28:51 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:07:20.535 17:28:51 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:07:20.535 17:28:51 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@873 -- # break 00:07:20.535 17:28:51 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:07:20.535 17:28:51 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:07:20.535 17:28:51 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:07:20.535 1+0 records in 00:07:20.535 1+0 records out 00:07:20.535 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000540818 s, 7.6 MB/s 00:07:20.535 17:28:51 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:07:20.535 17:28:51 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@886 -- # size=4096 00:07:20.535 17:28:51 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:07:20.535 17:28:51 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:07:20.535 17:28:51 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@889 -- # return 0 00:07:20.535 17:28:51 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:07:20.535 17:28:51 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:07:20.535 17:28:51 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@84 -- # nbd_get_count /var/tmp/spdk.sock 00:07:20.535 17:28:51 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk.sock 00:07:20.535 17:28:51 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_get_disks 00:07:20.795 17:28:51 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[ 00:07:20.795 { 00:07:20.795 "nbd_device": "/dev/nbd0", 00:07:20.795 "bdev_name": "raid" 00:07:20.795 } 00:07:20.795 ]' 00:07:20.795 17:28:51 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@64 -- # echo '[ 00:07:20.795 { 00:07:20.795 "nbd_device": "/dev/nbd0", 00:07:20.795 "bdev_name": "raid" 00:07:20.796 } 00:07:20.796 ]' 00:07:20.796 17:28:51 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:07:20.796 17:28:51 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@64 -- # nbd_disks_name=/dev/nbd0 00:07:20.796 17:28:51 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@65 -- # echo /dev/nbd0 00:07:20.796 17:28:51 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:07:20.796 17:28:51 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@65 -- # count=1 00:07:20.796 17:28:51 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@66 -- # echo 1 00:07:20.796 17:28:51 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@84 -- # count=1 00:07:20.796 17:28:51 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@85 -- # '[' 1 -ne 1 ']' 00:07:20.796 17:28:51 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@89 -- # raid_unmap_data_verify /dev/nbd0 00:07:20.796 17:28:51 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@17 -- # hash blkdiscard 00:07:20.796 17:28:51 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@18 -- # local nbd=/dev/nbd0 00:07:20.796 17:28:51 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@19 -- # local blksize 00:07:20.796 17:28:51 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@20 -- # lsblk -o LOG-SEC /dev/nbd0 00:07:20.796 17:28:51 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@20 -- # cut -d ' ' -f 5 00:07:20.796 17:28:51 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@20 -- # grep -v LOG-SEC 00:07:20.796 17:28:51 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@20 -- # blksize=512 00:07:20.796 17:28:51 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@21 -- # local rw_blk_num=4096 00:07:20.796 17:28:51 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@22 -- # local rw_len=2097152 00:07:20.796 17:28:51 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@23 -- # unmap_blk_offs=('0' '1028' '321') 00:07:20.796 17:28:51 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@23 -- # local unmap_blk_offs 00:07:20.796 17:28:51 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@24 -- # unmap_blk_nums=('128' '2035' '456') 00:07:20.796 17:28:51 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@24 -- # local unmap_blk_nums 00:07:20.796 17:28:51 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@25 -- # local unmap_off 00:07:20.796 17:28:51 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@26 -- # local unmap_len 00:07:20.796 17:28:51 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@29 -- # dd if=/dev/urandom of=/raidtest/raidrandtest bs=512 count=4096 00:07:20.796 4096+0 records in 00:07:20.796 4096+0 records out 00:07:20.796 2097152 bytes (2.1 MB, 2.0 MiB) copied, 0.0331786 s, 63.2 MB/s 00:07:20.796 17:28:51 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@30 -- # dd if=/raidtest/raidrandtest of=/dev/nbd0 bs=512 count=4096 oflag=direct 00:07:21.056 4096+0 records in 00:07:21.056 4096+0 records out 00:07:21.056 2097152 bytes (2.1 MB, 2.0 MiB) copied, 0.21805 s, 9.6 MB/s 00:07:21.056 17:28:52 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@31 -- # blockdev --flushbufs /dev/nbd0 00:07:21.056 17:28:52 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@34 -- # cmp -b -n 2097152 /raidtest/raidrandtest /dev/nbd0 00:07:21.056 17:28:52 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@36 -- # (( i = 0 )) 00:07:21.056 17:28:52 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@36 -- # (( i < 3 )) 00:07:21.056 17:28:52 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@37 -- # unmap_off=0 00:07:21.056 17:28:52 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@38 -- # unmap_len=65536 00:07:21.056 17:28:52 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@41 -- # dd if=/dev/zero of=/raidtest/raidrandtest bs=512 seek=0 count=128 conv=notrunc 00:07:21.056 128+0 records in 00:07:21.056 128+0 records out 00:07:21.056 65536 bytes (66 kB, 64 KiB) copied, 0.00118826 s, 55.2 MB/s 00:07:21.056 17:28:52 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@44 -- # blkdiscard -o 0 -l 65536 /dev/nbd0 00:07:21.056 17:28:52 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@45 -- # blockdev --flushbufs /dev/nbd0 00:07:21.056 17:28:52 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@48 -- # cmp -b -n 2097152 /raidtest/raidrandtest /dev/nbd0 00:07:21.056 17:28:52 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@36 -- # (( i++ )) 00:07:21.056 17:28:52 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@36 -- # (( i < 3 )) 00:07:21.056 17:28:52 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@37 -- # unmap_off=526336 00:07:21.056 17:28:52 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@38 -- # unmap_len=1041920 00:07:21.056 17:28:52 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@41 -- # dd if=/dev/zero of=/raidtest/raidrandtest bs=512 seek=1028 count=2035 conv=notrunc 00:07:21.056 2035+0 records in 00:07:21.056 2035+0 records out 00:07:21.056 1041920 bytes (1.0 MB, 1018 KiB) copied, 0.0117902 s, 88.4 MB/s 00:07:21.056 17:28:52 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@44 -- # blkdiscard -o 526336 -l 1041920 /dev/nbd0 00:07:21.316 17:28:52 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@45 -- # blockdev --flushbufs /dev/nbd0 00:07:21.316 17:28:52 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@48 -- # cmp -b -n 2097152 /raidtest/raidrandtest /dev/nbd0 00:07:21.316 17:28:52 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@36 -- # (( i++ )) 00:07:21.316 17:28:52 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@36 -- # (( i < 3 )) 00:07:21.316 17:28:52 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@37 -- # unmap_off=164352 00:07:21.316 17:28:52 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@38 -- # unmap_len=233472 00:07:21.316 17:28:52 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@41 -- # dd if=/dev/zero of=/raidtest/raidrandtest bs=512 seek=321 count=456 conv=notrunc 00:07:21.316 456+0 records in 00:07:21.316 456+0 records out 00:07:21.316 233472 bytes (233 kB, 228 KiB) copied, 0.00332947 s, 70.1 MB/s 00:07:21.316 17:28:52 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@44 -- # blkdiscard -o 164352 -l 233472 /dev/nbd0 00:07:21.316 17:28:52 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@45 -- # blockdev --flushbufs /dev/nbd0 00:07:21.316 17:28:52 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@48 -- # cmp -b -n 2097152 /raidtest/raidrandtest /dev/nbd0 00:07:21.316 17:28:52 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@36 -- # (( i++ )) 00:07:21.316 17:28:52 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@36 -- # (( i < 3 )) 00:07:21.316 17:28:52 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@52 -- # return 0 00:07:21.316 17:28:52 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@91 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:07:21.316 17:28:52 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:07:21.316 17:28:52 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:07:21.316 17:28:52 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@50 -- # local nbd_list 00:07:21.316 17:28:52 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@51 -- # local i 00:07:21.316 17:28:52 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:07:21.316 17:28:52 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:07:21.316 17:28:52 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:07:21.316 [2024-11-27 17:28:52.495803] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:21.316 17:28:52 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:07:21.316 17:28:52 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:07:21.316 17:28:52 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:07:21.316 17:28:52 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:07:21.316 17:28:52 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:07:21.316 17:28:52 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@41 -- # break 00:07:21.577 17:28:52 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@45 -- # return 0 00:07:21.577 17:28:52 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@92 -- # nbd_get_count /var/tmp/spdk.sock 00:07:21.577 17:28:52 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk.sock 00:07:21.577 17:28:52 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_get_disks 00:07:21.577 17:28:52 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:07:21.577 17:28:52 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@64 -- # echo '[]' 00:07:21.577 17:28:52 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:07:21.577 17:28:52 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:07:21.577 17:28:52 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@65 -- # echo '' 00:07:21.577 17:28:52 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:07:21.577 17:28:52 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@65 -- # true 00:07:21.577 17:28:52 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@65 -- # count=0 00:07:21.577 17:28:52 bdev_raid.raid_function_test_raid0 -- bdev/nbd_common.sh@66 -- # echo 0 00:07:21.838 17:28:52 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@92 -- # count=0 00:07:21.838 17:28:52 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@93 -- # '[' 0 -ne 0 ']' 00:07:21.838 17:28:52 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@97 -- # killprocess 71567 00:07:21.838 17:28:52 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@950 -- # '[' -z 71567 ']' 00:07:21.838 17:28:52 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@954 -- # kill -0 71567 00:07:21.838 17:28:52 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@955 -- # uname 00:07:21.838 17:28:52 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:07:21.838 17:28:52 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 71567 00:07:21.838 killing process with pid 71567 00:07:21.838 17:28:52 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:07:21.838 17:28:52 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:07:21.838 17:28:52 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@968 -- # echo 'killing process with pid 71567' 00:07:21.838 17:28:52 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@969 -- # kill 71567 00:07:21.838 17:28:52 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@974 -- # wait 71567 00:07:21.838 [2024-11-27 17:28:52.802796] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:21.838 [2024-11-27 17:28:52.802996] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:21.838 [2024-11-27 17:28:52.803075] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:21.838 [2024-11-27 17:28:52.803092] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid, state offline 00:07:21.838 [2024-11-27 17:28:52.843256] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:22.099 17:28:53 bdev_raid.raid_function_test_raid0 -- bdev/bdev_raid.sh@99 -- # return 0 00:07:22.099 00:07:22.099 real 0m2.850s 00:07:22.099 user 0m3.339s 00:07:22.099 sys 0m1.014s 00:07:22.099 17:28:53 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@1126 -- # xtrace_disable 00:07:22.099 17:28:53 bdev_raid.raid_function_test_raid0 -- common/autotest_common.sh@10 -- # set +x 00:07:22.099 ************************************ 00:07:22.099 END TEST raid_function_test_raid0 00:07:22.099 ************************************ 00:07:22.099 17:28:53 bdev_raid -- bdev/bdev_raid.sh@960 -- # run_test raid_function_test_concat raid_function_test concat 00:07:22.099 17:28:53 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:07:22.099 17:28:53 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:07:22.099 17:28:53 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:22.099 ************************************ 00:07:22.099 START TEST raid_function_test_concat 00:07:22.099 ************************************ 00:07:22.099 17:28:53 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@1125 -- # raid_function_test concat 00:07:22.099 17:28:53 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@64 -- # local raid_level=concat 00:07:22.099 17:28:53 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@65 -- # local nbd=/dev/nbd0 00:07:22.099 17:28:53 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@66 -- # local raid_bdev 00:07:22.099 17:28:53 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@69 -- # raid_pid=71687 00:07:22.099 Process raid pid: 71687 00:07:22.099 17:28:53 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@70 -- # echo 'Process raid pid: 71687' 00:07:22.099 17:28:53 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@68 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:07:22.099 17:28:53 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@71 -- # waitforlisten 71687 00:07:22.099 17:28:53 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@831 -- # '[' -z 71687 ']' 00:07:22.099 17:28:53 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:22.099 17:28:53 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@836 -- # local max_retries=100 00:07:22.099 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:22.099 17:28:53 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:22.099 17:28:53 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@840 -- # xtrace_disable 00:07:22.099 17:28:53 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@10 -- # set +x 00:07:22.360 [2024-11-27 17:28:53.368275] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:07:22.360 [2024-11-27 17:28:53.368430] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:07:22.360 [2024-11-27 17:28:53.496659] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:22.620 [2024-11-27 17:28:53.566844] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:07:22.620 [2024-11-27 17:28:53.642151] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:22.620 [2024-11-27 17:28:53.642196] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:23.190 17:28:54 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:07:23.190 17:28:54 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@864 -- # return 0 00:07:23.190 17:28:54 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@73 -- # rpc_cmd bdev_malloc_create 32 512 -b Base_1 00:07:23.190 17:28:54 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:23.190 17:28:54 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@10 -- # set +x 00:07:23.190 Base_1 00:07:23.190 17:28:54 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:23.190 17:28:54 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@74 -- # rpc_cmd bdev_malloc_create 32 512 -b Base_2 00:07:23.190 17:28:54 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:23.190 17:28:54 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@10 -- # set +x 00:07:23.190 Base_2 00:07:23.190 17:28:54 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:23.190 17:28:54 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@75 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''Base_1 Base_2'\''' -n raid 00:07:23.190 17:28:54 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:23.190 17:28:54 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@10 -- # set +x 00:07:23.190 [2024-11-27 17:28:54.234260] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev Base_1 is claimed 00:07:23.190 [2024-11-27 17:28:54.236373] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev Base_2 is claimed 00:07:23.190 [2024-11-27 17:28:54.236462] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:07:23.190 [2024-11-27 17:28:54.236474] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 131072, blocklen 512 00:07:23.190 [2024-11-27 17:28:54.236752] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:07:23.190 [2024-11-27 17:28:54.236912] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:07:23.190 [2024-11-27 17:28:54.236934] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid, raid_bdev 0x617000001200 00:07:23.190 [2024-11-27 17:28:54.237076] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:23.190 17:28:54 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:23.190 17:28:54 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@77 -- # rpc_cmd bdev_raid_get_bdevs online 00:07:23.190 17:28:54 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:23.190 17:28:54 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@10 -- # set +x 00:07:23.190 17:28:54 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@77 -- # jq -r '.[0]["name"] | select(.)' 00:07:23.190 17:28:54 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:23.190 17:28:54 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@77 -- # raid_bdev=raid 00:07:23.190 17:28:54 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@78 -- # '[' raid = '' ']' 00:07:23.190 17:28:54 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@83 -- # nbd_start_disks /var/tmp/spdk.sock raid /dev/nbd0 00:07:23.190 17:28:54 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:07:23.190 17:28:54 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@10 -- # bdev_list=('raid') 00:07:23.190 17:28:54 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@10 -- # local bdev_list 00:07:23.190 17:28:54 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:07:23.190 17:28:54 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@11 -- # local nbd_list 00:07:23.190 17:28:54 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@12 -- # local i 00:07:23.190 17:28:54 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:07:23.190 17:28:54 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:07:23.190 17:28:54 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid /dev/nbd0 00:07:23.450 [2024-11-27 17:28:54.453927] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:07:23.450 /dev/nbd0 00:07:23.450 17:28:54 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:07:23.450 17:28:54 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:07:23.450 17:28:54 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:07:23.450 17:28:54 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@869 -- # local i 00:07:23.450 17:28:54 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:07:23.450 17:28:54 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:07:23.450 17:28:54 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:07:23.450 17:28:54 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@873 -- # break 00:07:23.450 17:28:54 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:07:23.450 17:28:54 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:07:23.450 17:28:54 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:07:23.450 1+0 records in 00:07:23.450 1+0 records out 00:07:23.450 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000478246 s, 8.6 MB/s 00:07:23.450 17:28:54 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:07:23.450 17:28:54 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@886 -- # size=4096 00:07:23.450 17:28:54 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:07:23.450 17:28:54 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:07:23.450 17:28:54 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@889 -- # return 0 00:07:23.450 17:28:54 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:07:23.450 17:28:54 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:07:23.450 17:28:54 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@84 -- # nbd_get_count /var/tmp/spdk.sock 00:07:23.450 17:28:54 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk.sock 00:07:23.450 17:28:54 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_get_disks 00:07:23.710 17:28:54 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[ 00:07:23.710 { 00:07:23.710 "nbd_device": "/dev/nbd0", 00:07:23.710 "bdev_name": "raid" 00:07:23.710 } 00:07:23.710 ]' 00:07:23.710 17:28:54 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@64 -- # echo '[ 00:07:23.710 { 00:07:23.710 "nbd_device": "/dev/nbd0", 00:07:23.710 "bdev_name": "raid" 00:07:23.710 } 00:07:23.710 ]' 00:07:23.710 17:28:54 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:07:23.710 17:28:54 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@64 -- # nbd_disks_name=/dev/nbd0 00:07:23.710 17:28:54 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@65 -- # echo /dev/nbd0 00:07:23.710 17:28:54 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:07:23.710 17:28:54 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@65 -- # count=1 00:07:23.710 17:28:54 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@66 -- # echo 1 00:07:23.710 17:28:54 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@84 -- # count=1 00:07:23.710 17:28:54 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@85 -- # '[' 1 -ne 1 ']' 00:07:23.710 17:28:54 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@89 -- # raid_unmap_data_verify /dev/nbd0 00:07:23.710 17:28:54 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@17 -- # hash blkdiscard 00:07:23.710 17:28:54 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@18 -- # local nbd=/dev/nbd0 00:07:23.710 17:28:54 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@19 -- # local blksize 00:07:23.710 17:28:54 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@20 -- # grep -v LOG-SEC 00:07:23.710 17:28:54 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@20 -- # lsblk -o LOG-SEC /dev/nbd0 00:07:23.710 17:28:54 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@20 -- # cut -d ' ' -f 5 00:07:23.710 17:28:54 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@20 -- # blksize=512 00:07:23.710 17:28:54 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@21 -- # local rw_blk_num=4096 00:07:23.710 17:28:54 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@22 -- # local rw_len=2097152 00:07:23.710 17:28:54 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@23 -- # unmap_blk_offs=('0' '1028' '321') 00:07:23.710 17:28:54 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@23 -- # local unmap_blk_offs 00:07:23.710 17:28:54 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@24 -- # unmap_blk_nums=('128' '2035' '456') 00:07:23.710 17:28:54 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@24 -- # local unmap_blk_nums 00:07:23.710 17:28:54 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@25 -- # local unmap_off 00:07:23.710 17:28:54 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@26 -- # local unmap_len 00:07:23.710 17:28:54 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@29 -- # dd if=/dev/urandom of=/raidtest/raidrandtest bs=512 count=4096 00:07:23.710 4096+0 records in 00:07:23.710 4096+0 records out 00:07:23.710 2097152 bytes (2.1 MB, 2.0 MiB) copied, 0.0339 s, 61.9 MB/s 00:07:23.710 17:28:54 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@30 -- # dd if=/raidtest/raidrandtest of=/dev/nbd0 bs=512 count=4096 oflag=direct 00:07:23.970 4096+0 records in 00:07:23.970 4096+0 records out 00:07:23.970 2097152 bytes (2.1 MB, 2.0 MiB) copied, 0.208907 s, 10.0 MB/s 00:07:23.970 17:28:55 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@31 -- # blockdev --flushbufs /dev/nbd0 00:07:23.970 17:28:55 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@34 -- # cmp -b -n 2097152 /raidtest/raidrandtest /dev/nbd0 00:07:23.970 17:28:55 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@36 -- # (( i = 0 )) 00:07:23.970 17:28:55 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@36 -- # (( i < 3 )) 00:07:23.970 17:28:55 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@37 -- # unmap_off=0 00:07:23.971 17:28:55 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@38 -- # unmap_len=65536 00:07:23.971 17:28:55 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@41 -- # dd if=/dev/zero of=/raidtest/raidrandtest bs=512 seek=0 count=128 conv=notrunc 00:07:23.971 128+0 records in 00:07:23.971 128+0 records out 00:07:23.971 65536 bytes (66 kB, 64 KiB) copied, 0.0011794 s, 55.6 MB/s 00:07:23.971 17:28:55 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@44 -- # blkdiscard -o 0 -l 65536 /dev/nbd0 00:07:23.971 17:28:55 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@45 -- # blockdev --flushbufs /dev/nbd0 00:07:23.971 17:28:55 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@48 -- # cmp -b -n 2097152 /raidtest/raidrandtest /dev/nbd0 00:07:23.971 17:28:55 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@36 -- # (( i++ )) 00:07:23.971 17:28:55 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@36 -- # (( i < 3 )) 00:07:23.971 17:28:55 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@37 -- # unmap_off=526336 00:07:23.971 17:28:55 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@38 -- # unmap_len=1041920 00:07:23.971 17:28:55 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@41 -- # dd if=/dev/zero of=/raidtest/raidrandtest bs=512 seek=1028 count=2035 conv=notrunc 00:07:23.971 2035+0 records in 00:07:23.971 2035+0 records out 00:07:23.971 1041920 bytes (1.0 MB, 1018 KiB) copied, 0.0146846 s, 71.0 MB/s 00:07:23.971 17:28:55 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@44 -- # blkdiscard -o 526336 -l 1041920 /dev/nbd0 00:07:23.971 17:28:55 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@45 -- # blockdev --flushbufs /dev/nbd0 00:07:23.971 17:28:55 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@48 -- # cmp -b -n 2097152 /raidtest/raidrandtest /dev/nbd0 00:07:23.971 17:28:55 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@36 -- # (( i++ )) 00:07:23.971 17:28:55 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@36 -- # (( i < 3 )) 00:07:23.971 17:28:55 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@37 -- # unmap_off=164352 00:07:23.971 17:28:55 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@38 -- # unmap_len=233472 00:07:23.971 17:28:55 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@41 -- # dd if=/dev/zero of=/raidtest/raidrandtest bs=512 seek=321 count=456 conv=notrunc 00:07:23.971 456+0 records in 00:07:23.971 456+0 records out 00:07:23.971 233472 bytes (233 kB, 228 KiB) copied, 0.00294021 s, 79.4 MB/s 00:07:23.971 17:28:55 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@44 -- # blkdiscard -o 164352 -l 233472 /dev/nbd0 00:07:23.971 17:28:55 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@45 -- # blockdev --flushbufs /dev/nbd0 00:07:23.971 17:28:55 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@48 -- # cmp -b -n 2097152 /raidtest/raidrandtest /dev/nbd0 00:07:23.971 17:28:55 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@36 -- # (( i++ )) 00:07:23.971 17:28:55 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@36 -- # (( i < 3 )) 00:07:23.971 17:28:55 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@52 -- # return 0 00:07:23.971 17:28:55 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@91 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:07:23.971 17:28:55 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:07:23.971 17:28:55 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:07:23.971 17:28:55 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@50 -- # local nbd_list 00:07:23.971 17:28:55 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@51 -- # local i 00:07:23.971 17:28:55 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:07:23.971 17:28:55 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:07:24.231 17:28:55 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:07:24.231 [2024-11-27 17:28:55.376201] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:24.231 17:28:55 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:07:24.231 17:28:55 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:07:24.231 17:28:55 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:07:24.231 17:28:55 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:07:24.231 17:28:55 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:07:24.231 17:28:55 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@41 -- # break 00:07:24.231 17:28:55 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@45 -- # return 0 00:07:24.231 17:28:55 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@92 -- # nbd_get_count /var/tmp/spdk.sock 00:07:24.231 17:28:55 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk.sock 00:07:24.231 17:28:55 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_get_disks 00:07:24.491 17:28:55 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:07:24.491 17:28:55 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@64 -- # echo '[]' 00:07:24.491 17:28:55 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:07:24.491 17:28:55 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:07:24.491 17:28:55 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:07:24.491 17:28:55 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@65 -- # echo '' 00:07:24.491 17:28:55 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@65 -- # true 00:07:24.491 17:28:55 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@65 -- # count=0 00:07:24.491 17:28:55 bdev_raid.raid_function_test_concat -- bdev/nbd_common.sh@66 -- # echo 0 00:07:24.491 17:28:55 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@92 -- # count=0 00:07:24.491 17:28:55 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@93 -- # '[' 0 -ne 0 ']' 00:07:24.491 17:28:55 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@97 -- # killprocess 71687 00:07:24.491 17:28:55 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@950 -- # '[' -z 71687 ']' 00:07:24.491 17:28:55 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@954 -- # kill -0 71687 00:07:24.491 17:28:55 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@955 -- # uname 00:07:24.491 17:28:55 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:07:24.491 17:28:55 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 71687 00:07:24.752 17:28:55 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:07:24.752 17:28:55 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:07:24.752 killing process with pid 71687 00:07:24.752 17:28:55 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@968 -- # echo 'killing process with pid 71687' 00:07:24.752 17:28:55 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@969 -- # kill 71687 00:07:24.752 [2024-11-27 17:28:55.709508] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:24.752 17:28:55 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@974 -- # wait 71687 00:07:24.752 [2024-11-27 17:28:55.709668] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:24.752 [2024-11-27 17:28:55.709733] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:24.752 [2024-11-27 17:28:55.709754] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid, state offline 00:07:24.752 [2024-11-27 17:28:55.750799] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:25.012 17:28:56 bdev_raid.raid_function_test_concat -- bdev/bdev_raid.sh@99 -- # return 0 00:07:25.012 00:07:25.012 real 0m2.837s 00:07:25.012 user 0m3.343s 00:07:25.012 sys 0m1.004s 00:07:25.012 17:28:56 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@1126 -- # xtrace_disable 00:07:25.012 17:28:56 bdev_raid.raid_function_test_concat -- common/autotest_common.sh@10 -- # set +x 00:07:25.012 ************************************ 00:07:25.012 END TEST raid_function_test_concat 00:07:25.012 ************************************ 00:07:25.012 17:28:56 bdev_raid -- bdev/bdev_raid.sh@963 -- # run_test raid0_resize_test raid_resize_test 0 00:07:25.012 17:28:56 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:07:25.012 17:28:56 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:07:25.012 17:28:56 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:25.012 ************************************ 00:07:25.012 START TEST raid0_resize_test 00:07:25.012 ************************************ 00:07:25.012 17:28:56 bdev_raid.raid0_resize_test -- common/autotest_common.sh@1125 -- # raid_resize_test 0 00:07:25.012 17:28:56 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@332 -- # local raid_level=0 00:07:25.012 17:28:56 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@333 -- # local blksize=512 00:07:25.013 17:28:56 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@334 -- # local bdev_size_mb=32 00:07:25.013 17:28:56 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@335 -- # local new_bdev_size_mb=64 00:07:25.013 17:28:56 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@336 -- # local blkcnt 00:07:25.013 17:28:56 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@337 -- # local raid_size_mb 00:07:25.013 17:28:56 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@338 -- # local new_raid_size_mb 00:07:25.013 17:28:56 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@339 -- # local expected_size 00:07:25.013 17:28:56 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@342 -- # raid_pid=71799 00:07:25.013 17:28:56 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@343 -- # echo 'Process raid pid: 71799' 00:07:25.013 17:28:56 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@341 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:07:25.013 Process raid pid: 71799 00:07:25.013 17:28:56 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@344 -- # waitforlisten 71799 00:07:25.013 17:28:56 bdev_raid.raid0_resize_test -- common/autotest_common.sh@831 -- # '[' -z 71799 ']' 00:07:25.013 17:28:56 bdev_raid.raid0_resize_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:25.013 17:28:56 bdev_raid.raid0_resize_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:07:25.013 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:25.013 17:28:56 bdev_raid.raid0_resize_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:25.013 17:28:56 bdev_raid.raid0_resize_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:07:25.013 17:28:56 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:07:25.273 [2024-11-27 17:28:56.294342] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:07:25.273 [2024-11-27 17:28:56.294521] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:07:25.273 [2024-11-27 17:28:56.443011] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:25.533 [2024-11-27 17:28:56.512807] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:07:25.533 [2024-11-27 17:28:56.588816] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:25.533 [2024-11-27 17:28:56.588861] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:26.104 17:28:57 bdev_raid.raid0_resize_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:07:26.104 17:28:57 bdev_raid.raid0_resize_test -- common/autotest_common.sh@864 -- # return 0 00:07:26.104 17:28:57 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@346 -- # rpc_cmd bdev_null_create Base_1 32 512 00:07:26.104 17:28:57 bdev_raid.raid0_resize_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:26.104 17:28:57 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:07:26.104 Base_1 00:07:26.104 17:28:57 bdev_raid.raid0_resize_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:26.104 17:28:57 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@347 -- # rpc_cmd bdev_null_create Base_2 32 512 00:07:26.104 17:28:57 bdev_raid.raid0_resize_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:26.104 17:28:57 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:07:26.104 Base_2 00:07:26.104 17:28:57 bdev_raid.raid0_resize_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:26.104 17:28:57 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@349 -- # '[' 0 -eq 0 ']' 00:07:26.104 17:28:57 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@350 -- # rpc_cmd bdev_raid_create -z 64 -r 0 -b ''\''Base_1 Base_2'\''' -n Raid 00:07:26.104 17:28:57 bdev_raid.raid0_resize_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:26.104 17:28:57 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:07:26.104 [2024-11-27 17:28:57.119929] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev Base_1 is claimed 00:07:26.104 [2024-11-27 17:28:57.122015] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev Base_2 is claimed 00:07:26.104 [2024-11-27 17:28:57.122074] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:07:26.104 [2024-11-27 17:28:57.122085] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 131072, blocklen 512 00:07:26.104 [2024-11-27 17:28:57.122417] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000021f0 00:07:26.104 [2024-11-27 17:28:57.122531] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:07:26.104 [2024-11-27 17:28:57.122549] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Raid, raid_bdev 0x617000001200 00:07:26.104 [2024-11-27 17:28:57.122650] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:26.104 17:28:57 bdev_raid.raid0_resize_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:26.104 17:28:57 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@356 -- # rpc_cmd bdev_null_resize Base_1 64 00:07:26.104 17:28:57 bdev_raid.raid0_resize_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:26.104 17:28:57 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:07:26.104 [2024-11-27 17:28:57.131878] bdev_raid.c:2313:raid_bdev_resize_base_bdev: *DEBUG*: raid_bdev_resize_base_bdev 00:07:26.104 [2024-11-27 17:28:57.131912] bdev_raid.c:2326:raid_bdev_resize_base_bdev: *NOTICE*: base_bdev 'Base_1' was resized: old size 65536, new size 131072 00:07:26.104 true 00:07:26.104 17:28:57 bdev_raid.raid0_resize_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:26.104 17:28:57 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@359 -- # rpc_cmd bdev_get_bdevs -b Raid 00:07:26.104 17:28:57 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@359 -- # jq '.[].num_blocks' 00:07:26.104 17:28:57 bdev_raid.raid0_resize_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:26.104 17:28:57 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:07:26.104 [2024-11-27 17:28:57.148033] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:26.104 17:28:57 bdev_raid.raid0_resize_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:26.104 17:28:57 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@359 -- # blkcnt=131072 00:07:26.104 17:28:57 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@360 -- # raid_size_mb=64 00:07:26.104 17:28:57 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@361 -- # '[' 0 -eq 0 ']' 00:07:26.104 17:28:57 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@362 -- # expected_size=64 00:07:26.104 17:28:57 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@366 -- # '[' 64 '!=' 64 ']' 00:07:26.104 17:28:57 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@372 -- # rpc_cmd bdev_null_resize Base_2 64 00:07:26.104 17:28:57 bdev_raid.raid0_resize_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:26.104 17:28:57 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:07:26.104 [2024-11-27 17:28:57.191755] bdev_raid.c:2313:raid_bdev_resize_base_bdev: *DEBUG*: raid_bdev_resize_base_bdev 00:07:26.104 [2024-11-27 17:28:57.191780] bdev_raid.c:2326:raid_bdev_resize_base_bdev: *NOTICE*: base_bdev 'Base_2' was resized: old size 65536, new size 131072 00:07:26.104 [2024-11-27 17:28:57.191803] bdev_raid.c:2340:raid_bdev_resize_base_bdev: *NOTICE*: raid bdev 'Raid': block count was changed from 131072 to 262144 00:07:26.104 true 00:07:26.104 17:28:57 bdev_raid.raid0_resize_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:26.104 17:28:57 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@375 -- # rpc_cmd bdev_get_bdevs -b Raid 00:07:26.104 17:28:57 bdev_raid.raid0_resize_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:26.104 17:28:57 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:07:26.104 17:28:57 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@375 -- # jq '.[].num_blocks' 00:07:26.104 [2024-11-27 17:28:57.203900] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:26.104 17:28:57 bdev_raid.raid0_resize_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:26.104 17:28:57 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@375 -- # blkcnt=262144 00:07:26.104 17:28:57 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@376 -- # raid_size_mb=128 00:07:26.105 17:28:57 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@377 -- # '[' 0 -eq 0 ']' 00:07:26.105 17:28:57 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@378 -- # expected_size=128 00:07:26.105 17:28:57 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@382 -- # '[' 128 '!=' 128 ']' 00:07:26.105 17:28:57 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@387 -- # killprocess 71799 00:07:26.105 17:28:57 bdev_raid.raid0_resize_test -- common/autotest_common.sh@950 -- # '[' -z 71799 ']' 00:07:26.105 17:28:57 bdev_raid.raid0_resize_test -- common/autotest_common.sh@954 -- # kill -0 71799 00:07:26.105 17:28:57 bdev_raid.raid0_resize_test -- common/autotest_common.sh@955 -- # uname 00:07:26.105 17:28:57 bdev_raid.raid0_resize_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:07:26.105 17:28:57 bdev_raid.raid0_resize_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 71799 00:07:26.105 17:28:57 bdev_raid.raid0_resize_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:07:26.105 17:28:57 bdev_raid.raid0_resize_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:07:26.105 killing process with pid 71799 00:07:26.105 17:28:57 bdev_raid.raid0_resize_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 71799' 00:07:26.105 17:28:57 bdev_raid.raid0_resize_test -- common/autotest_common.sh@969 -- # kill 71799 00:07:26.105 [2024-11-27 17:28:57.282241] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:26.105 [2024-11-27 17:28:57.282318] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:26.105 [2024-11-27 17:28:57.282375] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:26.105 [2024-11-27 17:28:57.282386] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Raid, state offline 00:07:26.105 17:28:57 bdev_raid.raid0_resize_test -- common/autotest_common.sh@974 -- # wait 71799 00:07:26.105 [2024-11-27 17:28:57.284495] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:26.676 17:28:57 bdev_raid.raid0_resize_test -- bdev/bdev_raid.sh@389 -- # return 0 00:07:26.676 00:07:26.676 real 0m1.464s 00:07:26.676 user 0m1.542s 00:07:26.676 sys 0m0.369s 00:07:26.676 17:28:57 bdev_raid.raid0_resize_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:07:26.676 17:28:57 bdev_raid.raid0_resize_test -- common/autotest_common.sh@10 -- # set +x 00:07:26.676 ************************************ 00:07:26.676 END TEST raid0_resize_test 00:07:26.676 ************************************ 00:07:26.676 17:28:57 bdev_raid -- bdev/bdev_raid.sh@964 -- # run_test raid1_resize_test raid_resize_test 1 00:07:26.676 17:28:57 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:07:26.676 17:28:57 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:07:26.676 17:28:57 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:26.676 ************************************ 00:07:26.676 START TEST raid1_resize_test 00:07:26.676 ************************************ 00:07:26.676 17:28:57 bdev_raid.raid1_resize_test -- common/autotest_common.sh@1125 -- # raid_resize_test 1 00:07:26.676 17:28:57 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@332 -- # local raid_level=1 00:07:26.676 17:28:57 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@333 -- # local blksize=512 00:07:26.676 17:28:57 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@334 -- # local bdev_size_mb=32 00:07:26.676 17:28:57 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@335 -- # local new_bdev_size_mb=64 00:07:26.676 17:28:57 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@336 -- # local blkcnt 00:07:26.676 17:28:57 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@337 -- # local raid_size_mb 00:07:26.676 17:28:57 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@338 -- # local new_raid_size_mb 00:07:26.676 17:28:57 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@339 -- # local expected_size 00:07:26.676 17:28:57 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@342 -- # raid_pid=71855 00:07:26.676 17:28:57 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@341 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:07:26.676 Process raid pid: 71855 00:07:26.676 17:28:57 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@343 -- # echo 'Process raid pid: 71855' 00:07:26.676 17:28:57 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@344 -- # waitforlisten 71855 00:07:26.676 17:28:57 bdev_raid.raid1_resize_test -- common/autotest_common.sh@831 -- # '[' -z 71855 ']' 00:07:26.676 17:28:57 bdev_raid.raid1_resize_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:26.676 17:28:57 bdev_raid.raid1_resize_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:07:26.676 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:26.676 17:28:57 bdev_raid.raid1_resize_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:26.676 17:28:57 bdev_raid.raid1_resize_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:07:26.676 17:28:57 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:07:26.676 [2024-11-27 17:28:57.802065] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:07:26.676 [2024-11-27 17:28:57.802193] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:07:26.937 [2024-11-27 17:28:57.949228] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:26.937 [2024-11-27 17:28:58.019498] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:07:26.937 [2024-11-27 17:28:58.094953] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:26.937 [2024-11-27 17:28:58.094998] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:27.507 17:28:58 bdev_raid.raid1_resize_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:07:27.507 17:28:58 bdev_raid.raid1_resize_test -- common/autotest_common.sh@864 -- # return 0 00:07:27.507 17:28:58 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@346 -- # rpc_cmd bdev_null_create Base_1 32 512 00:07:27.507 17:28:58 bdev_raid.raid1_resize_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:27.507 17:28:58 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:07:27.507 Base_1 00:07:27.507 17:28:58 bdev_raid.raid1_resize_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:27.507 17:28:58 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@347 -- # rpc_cmd bdev_null_create Base_2 32 512 00:07:27.507 17:28:58 bdev_raid.raid1_resize_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:27.507 17:28:58 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:07:27.507 Base_2 00:07:27.507 17:28:58 bdev_raid.raid1_resize_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:27.507 17:28:58 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@349 -- # '[' 1 -eq 0 ']' 00:07:27.507 17:28:58 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@352 -- # rpc_cmd bdev_raid_create -r 1 -b ''\''Base_1 Base_2'\''' -n Raid 00:07:27.507 17:28:58 bdev_raid.raid1_resize_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:27.507 17:28:58 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:07:27.507 [2024-11-27 17:28:58.675106] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev Base_1 is claimed 00:07:27.507 [2024-11-27 17:28:58.677186] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev Base_2 is claimed 00:07:27.507 [2024-11-27 17:28:58.677252] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:07:27.507 [2024-11-27 17:28:58.677265] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:07:27.507 [2024-11-27 17:28:58.677543] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000021f0 00:07:27.507 [2024-11-27 17:28:58.677667] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:07:27.507 [2024-11-27 17:28:58.677686] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Raid, raid_bdev 0x617000001200 00:07:27.507 [2024-11-27 17:28:58.677820] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:27.507 17:28:58 bdev_raid.raid1_resize_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:27.507 17:28:58 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@356 -- # rpc_cmd bdev_null_resize Base_1 64 00:07:27.507 17:28:58 bdev_raid.raid1_resize_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:27.507 17:28:58 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:07:27.507 [2024-11-27 17:28:58.687041] bdev_raid.c:2313:raid_bdev_resize_base_bdev: *DEBUG*: raid_bdev_resize_base_bdev 00:07:27.507 [2024-11-27 17:28:58.687069] bdev_raid.c:2326:raid_bdev_resize_base_bdev: *NOTICE*: base_bdev 'Base_1' was resized: old size 65536, new size 131072 00:07:27.507 true 00:07:27.507 17:28:58 bdev_raid.raid1_resize_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:27.507 17:28:58 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@359 -- # jq '.[].num_blocks' 00:07:27.507 17:28:58 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@359 -- # rpc_cmd bdev_get_bdevs -b Raid 00:07:27.507 17:28:58 bdev_raid.raid1_resize_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:27.507 17:28:58 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:07:27.768 [2024-11-27 17:28:58.703196] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:27.768 17:28:58 bdev_raid.raid1_resize_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:27.768 17:28:58 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@359 -- # blkcnt=65536 00:07:27.768 17:28:58 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@360 -- # raid_size_mb=32 00:07:27.768 17:28:58 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@361 -- # '[' 1 -eq 0 ']' 00:07:27.768 17:28:58 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@364 -- # expected_size=32 00:07:27.768 17:28:58 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@366 -- # '[' 32 '!=' 32 ']' 00:07:27.768 17:28:58 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@372 -- # rpc_cmd bdev_null_resize Base_2 64 00:07:27.768 17:28:58 bdev_raid.raid1_resize_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:27.768 17:28:58 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:07:27.768 [2024-11-27 17:28:58.738936] bdev_raid.c:2313:raid_bdev_resize_base_bdev: *DEBUG*: raid_bdev_resize_base_bdev 00:07:27.768 [2024-11-27 17:28:58.738959] bdev_raid.c:2326:raid_bdev_resize_base_bdev: *NOTICE*: base_bdev 'Base_2' was resized: old size 65536, new size 131072 00:07:27.768 [2024-11-27 17:28:58.738983] bdev_raid.c:2340:raid_bdev_resize_base_bdev: *NOTICE*: raid bdev 'Raid': block count was changed from 65536 to 131072 00:07:27.768 true 00:07:27.768 17:28:58 bdev_raid.raid1_resize_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:27.768 17:28:58 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@375 -- # rpc_cmd bdev_get_bdevs -b Raid 00:07:27.768 17:28:58 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@375 -- # jq '.[].num_blocks' 00:07:27.768 17:28:58 bdev_raid.raid1_resize_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:27.768 17:28:58 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:07:27.768 [2024-11-27 17:28:58.755081] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:27.768 17:28:58 bdev_raid.raid1_resize_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:27.768 17:28:58 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@375 -- # blkcnt=131072 00:07:27.768 17:28:58 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@376 -- # raid_size_mb=64 00:07:27.768 17:28:58 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@377 -- # '[' 1 -eq 0 ']' 00:07:27.768 17:28:58 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@380 -- # expected_size=64 00:07:27.768 17:28:58 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@382 -- # '[' 64 '!=' 64 ']' 00:07:27.768 17:28:58 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@387 -- # killprocess 71855 00:07:27.768 17:28:58 bdev_raid.raid1_resize_test -- common/autotest_common.sh@950 -- # '[' -z 71855 ']' 00:07:27.768 17:28:58 bdev_raid.raid1_resize_test -- common/autotest_common.sh@954 -- # kill -0 71855 00:07:27.768 17:28:58 bdev_raid.raid1_resize_test -- common/autotest_common.sh@955 -- # uname 00:07:27.768 17:28:58 bdev_raid.raid1_resize_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:07:27.768 17:28:58 bdev_raid.raid1_resize_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 71855 00:07:27.768 17:28:58 bdev_raid.raid1_resize_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:07:27.768 17:28:58 bdev_raid.raid1_resize_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:07:27.768 killing process with pid 71855 00:07:27.768 17:28:58 bdev_raid.raid1_resize_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 71855' 00:07:27.768 17:28:58 bdev_raid.raid1_resize_test -- common/autotest_common.sh@969 -- # kill 71855 00:07:27.768 [2024-11-27 17:28:58.834634] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:27.768 [2024-11-27 17:28:58.834730] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:27.768 17:28:58 bdev_raid.raid1_resize_test -- common/autotest_common.sh@974 -- # wait 71855 00:07:27.768 [2024-11-27 17:28:58.835208] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:27.768 [2024-11-27 17:28:58.835234] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Raid, state offline 00:07:27.768 [2024-11-27 17:28:58.836935] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:28.028 17:28:59 bdev_raid.raid1_resize_test -- bdev/bdev_raid.sh@389 -- # return 0 00:07:28.028 00:07:28.028 real 0m1.486s 00:07:28.028 user 0m1.571s 00:07:28.028 sys 0m0.395s 00:07:28.028 17:28:59 bdev_raid.raid1_resize_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:07:28.028 17:28:59 bdev_raid.raid1_resize_test -- common/autotest_common.sh@10 -- # set +x 00:07:28.028 ************************************ 00:07:28.028 END TEST raid1_resize_test 00:07:28.028 ************************************ 00:07:28.288 17:28:59 bdev_raid -- bdev/bdev_raid.sh@966 -- # for n in {2..4} 00:07:28.288 17:28:59 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:07:28.288 17:28:59 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test raid0 2 false 00:07:28.288 17:28:59 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:07:28.288 17:28:59 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:07:28.288 17:28:59 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:28.288 ************************************ 00:07:28.288 START TEST raid_state_function_test 00:07:28.288 ************************************ 00:07:28.288 17:28:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1125 -- # raid_state_function_test raid0 2 false 00:07:28.288 17:28:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=raid0 00:07:28.288 17:28:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:07:28.288 17:28:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:07:28.288 17:28:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:07:28.288 17:28:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:07:28.288 17:28:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:28.288 17:28:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:07:28.288 17:28:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:07:28.288 17:28:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:28.288 17:28:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:07:28.288 17:28:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:07:28.288 17:28:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:28.288 17:28:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:07:28.288 17:28:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:07:28.288 17:28:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:07:28.288 17:28:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:07:28.288 17:28:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:07:28.288 17:28:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:07:28.288 17:28:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' raid0 '!=' raid1 ']' 00:07:28.288 17:28:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:07:28.288 17:28:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:07:28.288 17:28:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:07:28.288 17:28:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:07:28.288 17:28:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=71901 00:07:28.288 Process raid pid: 71901 00:07:28.288 17:28:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 71901' 00:07:28.288 17:28:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:07:28.288 17:28:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 71901 00:07:28.288 17:28:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@831 -- # '[' -z 71901 ']' 00:07:28.288 17:28:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:28.288 17:28:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:07:28.288 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:28.288 17:28:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:28.288 17:28:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:07:28.288 17:28:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:28.288 [2024-11-27 17:28:59.373346] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:07:28.288 [2024-11-27 17:28:59.373515] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:07:28.548 [2024-11-27 17:28:59.514525] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:28.548 [2024-11-27 17:28:59.590595] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:07:28.548 [2024-11-27 17:28:59.670916] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:28.548 [2024-11-27 17:28:59.670963] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:29.117 17:29:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:07:29.117 17:29:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@864 -- # return 0 00:07:29.117 17:29:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:29.117 17:29:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:29.117 17:29:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:29.117 [2024-11-27 17:29:00.200423] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:07:29.117 [2024-11-27 17:29:00.200483] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:07:29.117 [2024-11-27 17:29:00.200504] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:29.117 [2024-11-27 17:29:00.200515] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:29.117 17:29:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:29.117 17:29:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 2 00:07:29.117 17:29:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:29.117 17:29:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:29.117 17:29:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:29.117 17:29:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:29.117 17:29:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:29.117 17:29:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:29.117 17:29:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:29.117 17:29:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:29.117 17:29:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:29.117 17:29:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:29.117 17:29:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:29.117 17:29:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:29.117 17:29:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:29.117 17:29:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:29.117 17:29:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:29.117 "name": "Existed_Raid", 00:07:29.117 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:29.117 "strip_size_kb": 64, 00:07:29.117 "state": "configuring", 00:07:29.117 "raid_level": "raid0", 00:07:29.117 "superblock": false, 00:07:29.117 "num_base_bdevs": 2, 00:07:29.117 "num_base_bdevs_discovered": 0, 00:07:29.117 "num_base_bdevs_operational": 2, 00:07:29.117 "base_bdevs_list": [ 00:07:29.117 { 00:07:29.117 "name": "BaseBdev1", 00:07:29.117 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:29.117 "is_configured": false, 00:07:29.117 "data_offset": 0, 00:07:29.117 "data_size": 0 00:07:29.117 }, 00:07:29.117 { 00:07:29.117 "name": "BaseBdev2", 00:07:29.117 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:29.117 "is_configured": false, 00:07:29.117 "data_offset": 0, 00:07:29.117 "data_size": 0 00:07:29.117 } 00:07:29.117 ] 00:07:29.117 }' 00:07:29.117 17:29:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:29.117 17:29:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:29.689 17:29:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:07:29.689 17:29:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:29.689 17:29:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:29.689 [2024-11-27 17:29:00.659496] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:07:29.689 [2024-11-27 17:29:00.659546] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:07:29.689 17:29:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:29.689 17:29:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:29.689 17:29:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:29.690 17:29:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:29.690 [2024-11-27 17:29:00.671488] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:07:29.690 [2024-11-27 17:29:00.671530] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:07:29.690 [2024-11-27 17:29:00.671549] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:29.690 [2024-11-27 17:29:00.671560] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:29.690 17:29:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:29.690 17:29:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:07:29.690 17:29:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:29.690 17:29:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:29.690 [2024-11-27 17:29:00.699164] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:29.690 BaseBdev1 00:07:29.690 17:29:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:29.690 17:29:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:07:29.690 17:29:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:07:29.690 17:29:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:07:29.690 17:29:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:07:29.690 17:29:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:07:29.690 17:29:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:07:29.690 17:29:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:07:29.690 17:29:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:29.690 17:29:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:29.690 17:29:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:29.690 17:29:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:07:29.690 17:29:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:29.690 17:29:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:29.690 [ 00:07:29.690 { 00:07:29.690 "name": "BaseBdev1", 00:07:29.690 "aliases": [ 00:07:29.690 "cd444473-9e9d-4e7b-aa09-a5a2c52cf698" 00:07:29.690 ], 00:07:29.690 "product_name": "Malloc disk", 00:07:29.690 "block_size": 512, 00:07:29.690 "num_blocks": 65536, 00:07:29.690 "uuid": "cd444473-9e9d-4e7b-aa09-a5a2c52cf698", 00:07:29.690 "assigned_rate_limits": { 00:07:29.690 "rw_ios_per_sec": 0, 00:07:29.690 "rw_mbytes_per_sec": 0, 00:07:29.690 "r_mbytes_per_sec": 0, 00:07:29.690 "w_mbytes_per_sec": 0 00:07:29.690 }, 00:07:29.690 "claimed": true, 00:07:29.690 "claim_type": "exclusive_write", 00:07:29.690 "zoned": false, 00:07:29.690 "supported_io_types": { 00:07:29.690 "read": true, 00:07:29.690 "write": true, 00:07:29.690 "unmap": true, 00:07:29.690 "flush": true, 00:07:29.690 "reset": true, 00:07:29.690 "nvme_admin": false, 00:07:29.690 "nvme_io": false, 00:07:29.690 "nvme_io_md": false, 00:07:29.690 "write_zeroes": true, 00:07:29.690 "zcopy": true, 00:07:29.690 "get_zone_info": false, 00:07:29.690 "zone_management": false, 00:07:29.690 "zone_append": false, 00:07:29.690 "compare": false, 00:07:29.690 "compare_and_write": false, 00:07:29.690 "abort": true, 00:07:29.690 "seek_hole": false, 00:07:29.690 "seek_data": false, 00:07:29.690 "copy": true, 00:07:29.690 "nvme_iov_md": false 00:07:29.690 }, 00:07:29.690 "memory_domains": [ 00:07:29.690 { 00:07:29.690 "dma_device_id": "system", 00:07:29.690 "dma_device_type": 1 00:07:29.690 }, 00:07:29.690 { 00:07:29.690 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:29.690 "dma_device_type": 2 00:07:29.690 } 00:07:29.690 ], 00:07:29.690 "driver_specific": {} 00:07:29.690 } 00:07:29.690 ] 00:07:29.690 17:29:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:29.690 17:29:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:07:29.690 17:29:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 2 00:07:29.690 17:29:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:29.690 17:29:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:29.690 17:29:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:29.690 17:29:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:29.690 17:29:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:29.690 17:29:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:29.690 17:29:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:29.690 17:29:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:29.690 17:29:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:29.690 17:29:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:29.690 17:29:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:29.690 17:29:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:29.690 17:29:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:29.690 17:29:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:29.690 17:29:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:29.690 "name": "Existed_Raid", 00:07:29.690 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:29.690 "strip_size_kb": 64, 00:07:29.690 "state": "configuring", 00:07:29.690 "raid_level": "raid0", 00:07:29.690 "superblock": false, 00:07:29.690 "num_base_bdevs": 2, 00:07:29.690 "num_base_bdevs_discovered": 1, 00:07:29.690 "num_base_bdevs_operational": 2, 00:07:29.690 "base_bdevs_list": [ 00:07:29.690 { 00:07:29.690 "name": "BaseBdev1", 00:07:29.690 "uuid": "cd444473-9e9d-4e7b-aa09-a5a2c52cf698", 00:07:29.690 "is_configured": true, 00:07:29.690 "data_offset": 0, 00:07:29.690 "data_size": 65536 00:07:29.690 }, 00:07:29.690 { 00:07:29.690 "name": "BaseBdev2", 00:07:29.690 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:29.690 "is_configured": false, 00:07:29.690 "data_offset": 0, 00:07:29.690 "data_size": 0 00:07:29.690 } 00:07:29.690 ] 00:07:29.690 }' 00:07:29.690 17:29:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:29.690 17:29:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:29.958 17:29:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:07:29.958 17:29:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:29.958 17:29:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:29.958 [2024-11-27 17:29:01.138451] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:07:29.958 [2024-11-27 17:29:01.138510] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:07:29.958 17:29:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:29.958 17:29:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:29.958 17:29:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:29.958 17:29:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:30.218 [2024-11-27 17:29:01.150476] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:30.218 [2024-11-27 17:29:01.152734] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:30.218 [2024-11-27 17:29:01.152778] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:30.218 17:29:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:30.218 17:29:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:07:30.218 17:29:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:07:30.218 17:29:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 2 00:07:30.218 17:29:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:30.218 17:29:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:30.218 17:29:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:30.218 17:29:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:30.218 17:29:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:30.218 17:29:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:30.218 17:29:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:30.218 17:29:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:30.218 17:29:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:30.218 17:29:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:30.218 17:29:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:30.218 17:29:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:30.218 17:29:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:30.218 17:29:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:30.218 17:29:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:30.218 "name": "Existed_Raid", 00:07:30.218 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:30.218 "strip_size_kb": 64, 00:07:30.218 "state": "configuring", 00:07:30.218 "raid_level": "raid0", 00:07:30.218 "superblock": false, 00:07:30.218 "num_base_bdevs": 2, 00:07:30.218 "num_base_bdevs_discovered": 1, 00:07:30.218 "num_base_bdevs_operational": 2, 00:07:30.218 "base_bdevs_list": [ 00:07:30.218 { 00:07:30.218 "name": "BaseBdev1", 00:07:30.218 "uuid": "cd444473-9e9d-4e7b-aa09-a5a2c52cf698", 00:07:30.218 "is_configured": true, 00:07:30.218 "data_offset": 0, 00:07:30.218 "data_size": 65536 00:07:30.218 }, 00:07:30.218 { 00:07:30.219 "name": "BaseBdev2", 00:07:30.219 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:30.219 "is_configured": false, 00:07:30.219 "data_offset": 0, 00:07:30.219 "data_size": 0 00:07:30.219 } 00:07:30.219 ] 00:07:30.219 }' 00:07:30.219 17:29:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:30.219 17:29:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:30.479 17:29:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:07:30.479 17:29:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:30.479 17:29:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:30.479 [2024-11-27 17:29:01.554395] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:07:30.479 [2024-11-27 17:29:01.554458] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:07:30.479 [2024-11-27 17:29:01.554470] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 131072, blocklen 512 00:07:30.479 [2024-11-27 17:29:01.554825] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:07:30.479 [2024-11-27 17:29:01.555004] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:07:30.479 [2024-11-27 17:29:01.555028] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:07:30.479 [2024-11-27 17:29:01.555297] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:30.479 BaseBdev2 00:07:30.479 17:29:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:30.479 17:29:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:07:30.479 17:29:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:07:30.479 17:29:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:07:30.479 17:29:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:07:30.479 17:29:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:07:30.479 17:29:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:07:30.479 17:29:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:07:30.479 17:29:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:30.479 17:29:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:30.479 17:29:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:30.479 17:29:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:07:30.479 17:29:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:30.479 17:29:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:30.479 [ 00:07:30.479 { 00:07:30.479 "name": "BaseBdev2", 00:07:30.479 "aliases": [ 00:07:30.479 "2e672b32-c631-4d02-bace-ab13a18ab211" 00:07:30.479 ], 00:07:30.479 "product_name": "Malloc disk", 00:07:30.479 "block_size": 512, 00:07:30.479 "num_blocks": 65536, 00:07:30.479 "uuid": "2e672b32-c631-4d02-bace-ab13a18ab211", 00:07:30.479 "assigned_rate_limits": { 00:07:30.479 "rw_ios_per_sec": 0, 00:07:30.479 "rw_mbytes_per_sec": 0, 00:07:30.479 "r_mbytes_per_sec": 0, 00:07:30.479 "w_mbytes_per_sec": 0 00:07:30.479 }, 00:07:30.479 "claimed": true, 00:07:30.479 "claim_type": "exclusive_write", 00:07:30.479 "zoned": false, 00:07:30.479 "supported_io_types": { 00:07:30.479 "read": true, 00:07:30.479 "write": true, 00:07:30.479 "unmap": true, 00:07:30.479 "flush": true, 00:07:30.479 "reset": true, 00:07:30.479 "nvme_admin": false, 00:07:30.479 "nvme_io": false, 00:07:30.479 "nvme_io_md": false, 00:07:30.479 "write_zeroes": true, 00:07:30.479 "zcopy": true, 00:07:30.479 "get_zone_info": false, 00:07:30.479 "zone_management": false, 00:07:30.479 "zone_append": false, 00:07:30.479 "compare": false, 00:07:30.479 "compare_and_write": false, 00:07:30.479 "abort": true, 00:07:30.479 "seek_hole": false, 00:07:30.479 "seek_data": false, 00:07:30.479 "copy": true, 00:07:30.479 "nvme_iov_md": false 00:07:30.479 }, 00:07:30.479 "memory_domains": [ 00:07:30.479 { 00:07:30.479 "dma_device_id": "system", 00:07:30.479 "dma_device_type": 1 00:07:30.479 }, 00:07:30.479 { 00:07:30.479 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:30.479 "dma_device_type": 2 00:07:30.479 } 00:07:30.479 ], 00:07:30.479 "driver_specific": {} 00:07:30.479 } 00:07:30.479 ] 00:07:30.479 17:29:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:30.479 17:29:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:07:30.479 17:29:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:07:30.479 17:29:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:07:30.479 17:29:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid0 64 2 00:07:30.479 17:29:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:30.479 17:29:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:30.479 17:29:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:30.479 17:29:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:30.479 17:29:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:30.479 17:29:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:30.479 17:29:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:30.479 17:29:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:30.479 17:29:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:30.479 17:29:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:30.479 17:29:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:30.479 17:29:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:30.479 17:29:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:30.479 17:29:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:30.479 17:29:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:30.479 "name": "Existed_Raid", 00:07:30.479 "uuid": "6a19fd6e-fb12-49ca-95cf-4ec8bd280e22", 00:07:30.479 "strip_size_kb": 64, 00:07:30.479 "state": "online", 00:07:30.479 "raid_level": "raid0", 00:07:30.479 "superblock": false, 00:07:30.479 "num_base_bdevs": 2, 00:07:30.479 "num_base_bdevs_discovered": 2, 00:07:30.479 "num_base_bdevs_operational": 2, 00:07:30.479 "base_bdevs_list": [ 00:07:30.479 { 00:07:30.479 "name": "BaseBdev1", 00:07:30.479 "uuid": "cd444473-9e9d-4e7b-aa09-a5a2c52cf698", 00:07:30.479 "is_configured": true, 00:07:30.479 "data_offset": 0, 00:07:30.479 "data_size": 65536 00:07:30.479 }, 00:07:30.479 { 00:07:30.479 "name": "BaseBdev2", 00:07:30.479 "uuid": "2e672b32-c631-4d02-bace-ab13a18ab211", 00:07:30.479 "is_configured": true, 00:07:30.479 "data_offset": 0, 00:07:30.479 "data_size": 65536 00:07:30.479 } 00:07:30.480 ] 00:07:30.480 }' 00:07:30.480 17:29:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:30.480 17:29:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:31.050 17:29:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:07:31.050 17:29:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:07:31.050 17:29:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:07:31.050 17:29:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:07:31.050 17:29:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:07:31.050 17:29:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:07:31.050 17:29:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:07:31.050 17:29:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:07:31.050 17:29:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:31.050 17:29:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:31.050 [2024-11-27 17:29:02.069815] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:31.050 17:29:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:31.050 17:29:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:07:31.050 "name": "Existed_Raid", 00:07:31.050 "aliases": [ 00:07:31.050 "6a19fd6e-fb12-49ca-95cf-4ec8bd280e22" 00:07:31.050 ], 00:07:31.050 "product_name": "Raid Volume", 00:07:31.050 "block_size": 512, 00:07:31.050 "num_blocks": 131072, 00:07:31.050 "uuid": "6a19fd6e-fb12-49ca-95cf-4ec8bd280e22", 00:07:31.050 "assigned_rate_limits": { 00:07:31.050 "rw_ios_per_sec": 0, 00:07:31.050 "rw_mbytes_per_sec": 0, 00:07:31.050 "r_mbytes_per_sec": 0, 00:07:31.050 "w_mbytes_per_sec": 0 00:07:31.050 }, 00:07:31.050 "claimed": false, 00:07:31.050 "zoned": false, 00:07:31.050 "supported_io_types": { 00:07:31.050 "read": true, 00:07:31.050 "write": true, 00:07:31.050 "unmap": true, 00:07:31.050 "flush": true, 00:07:31.050 "reset": true, 00:07:31.050 "nvme_admin": false, 00:07:31.050 "nvme_io": false, 00:07:31.050 "nvme_io_md": false, 00:07:31.050 "write_zeroes": true, 00:07:31.050 "zcopy": false, 00:07:31.050 "get_zone_info": false, 00:07:31.050 "zone_management": false, 00:07:31.050 "zone_append": false, 00:07:31.050 "compare": false, 00:07:31.050 "compare_and_write": false, 00:07:31.050 "abort": false, 00:07:31.051 "seek_hole": false, 00:07:31.051 "seek_data": false, 00:07:31.051 "copy": false, 00:07:31.051 "nvme_iov_md": false 00:07:31.051 }, 00:07:31.051 "memory_domains": [ 00:07:31.051 { 00:07:31.051 "dma_device_id": "system", 00:07:31.051 "dma_device_type": 1 00:07:31.051 }, 00:07:31.051 { 00:07:31.051 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:31.051 "dma_device_type": 2 00:07:31.051 }, 00:07:31.051 { 00:07:31.051 "dma_device_id": "system", 00:07:31.051 "dma_device_type": 1 00:07:31.051 }, 00:07:31.051 { 00:07:31.051 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:31.051 "dma_device_type": 2 00:07:31.051 } 00:07:31.051 ], 00:07:31.051 "driver_specific": { 00:07:31.051 "raid": { 00:07:31.051 "uuid": "6a19fd6e-fb12-49ca-95cf-4ec8bd280e22", 00:07:31.051 "strip_size_kb": 64, 00:07:31.051 "state": "online", 00:07:31.051 "raid_level": "raid0", 00:07:31.051 "superblock": false, 00:07:31.051 "num_base_bdevs": 2, 00:07:31.051 "num_base_bdevs_discovered": 2, 00:07:31.051 "num_base_bdevs_operational": 2, 00:07:31.051 "base_bdevs_list": [ 00:07:31.051 { 00:07:31.051 "name": "BaseBdev1", 00:07:31.051 "uuid": "cd444473-9e9d-4e7b-aa09-a5a2c52cf698", 00:07:31.051 "is_configured": true, 00:07:31.051 "data_offset": 0, 00:07:31.051 "data_size": 65536 00:07:31.051 }, 00:07:31.051 { 00:07:31.051 "name": "BaseBdev2", 00:07:31.051 "uuid": "2e672b32-c631-4d02-bace-ab13a18ab211", 00:07:31.051 "is_configured": true, 00:07:31.051 "data_offset": 0, 00:07:31.051 "data_size": 65536 00:07:31.051 } 00:07:31.051 ] 00:07:31.051 } 00:07:31.051 } 00:07:31.051 }' 00:07:31.051 17:29:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:07:31.051 17:29:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:07:31.051 BaseBdev2' 00:07:31.051 17:29:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:31.051 17:29:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:07:31.051 17:29:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:31.051 17:29:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:31.051 17:29:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:07:31.051 17:29:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:31.051 17:29:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:31.051 17:29:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:31.051 17:29:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:31.051 17:29:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:31.051 17:29:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:31.051 17:29:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:07:31.051 17:29:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:31.051 17:29:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:31.051 17:29:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:31.051 17:29:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:31.310 17:29:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:31.310 17:29:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:31.310 17:29:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:07:31.310 17:29:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:31.310 17:29:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:31.310 [2024-11-27 17:29:02.273264] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:07:31.311 [2024-11-27 17:29:02.273294] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:31.311 [2024-11-27 17:29:02.273362] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:31.311 17:29:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:31.311 17:29:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:07:31.311 17:29:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy raid0 00:07:31.311 17:29:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:07:31.311 17:29:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # return 1 00:07:31.311 17:29:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:07:31.311 17:29:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline raid0 64 1 00:07:31.311 17:29:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:31.311 17:29:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:07:31.311 17:29:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:31.311 17:29:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:31.311 17:29:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:07:31.311 17:29:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:31.311 17:29:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:31.311 17:29:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:31.311 17:29:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:31.311 17:29:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:31.311 17:29:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:31.311 17:29:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:31.311 17:29:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:31.311 17:29:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:31.311 17:29:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:31.311 "name": "Existed_Raid", 00:07:31.311 "uuid": "6a19fd6e-fb12-49ca-95cf-4ec8bd280e22", 00:07:31.311 "strip_size_kb": 64, 00:07:31.311 "state": "offline", 00:07:31.311 "raid_level": "raid0", 00:07:31.311 "superblock": false, 00:07:31.311 "num_base_bdevs": 2, 00:07:31.311 "num_base_bdevs_discovered": 1, 00:07:31.311 "num_base_bdevs_operational": 1, 00:07:31.311 "base_bdevs_list": [ 00:07:31.311 { 00:07:31.311 "name": null, 00:07:31.311 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:31.311 "is_configured": false, 00:07:31.311 "data_offset": 0, 00:07:31.311 "data_size": 65536 00:07:31.311 }, 00:07:31.311 { 00:07:31.311 "name": "BaseBdev2", 00:07:31.311 "uuid": "2e672b32-c631-4d02-bace-ab13a18ab211", 00:07:31.311 "is_configured": true, 00:07:31.311 "data_offset": 0, 00:07:31.311 "data_size": 65536 00:07:31.311 } 00:07:31.311 ] 00:07:31.311 }' 00:07:31.311 17:29:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:31.311 17:29:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:31.573 17:29:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:07:31.573 17:29:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:07:31.573 17:29:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:07:31.573 17:29:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:31.573 17:29:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:31.573 17:29:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:31.573 17:29:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:31.573 17:29:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:07:31.573 17:29:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:07:31.573 17:29:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:07:31.573 17:29:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:31.573 17:29:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:31.573 [2024-11-27 17:29:02.741216] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:07:31.573 [2024-11-27 17:29:02.741272] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:07:31.833 17:29:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:31.833 17:29:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:07:31.833 17:29:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:07:31.833 17:29:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:31.833 17:29:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:07:31.833 17:29:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:31.833 17:29:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:31.833 17:29:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:31.833 17:29:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:07:31.833 17:29:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:07:31.834 17:29:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:07:31.834 17:29:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 71901 00:07:31.834 17:29:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@950 -- # '[' -z 71901 ']' 00:07:31.834 17:29:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # kill -0 71901 00:07:31.834 17:29:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@955 -- # uname 00:07:31.834 17:29:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:07:31.834 17:29:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 71901 00:07:31.834 17:29:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:07:31.834 17:29:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:07:31.834 killing process with pid 71901 00:07:31.834 17:29:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 71901' 00:07:31.834 17:29:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@969 -- # kill 71901 00:07:31.834 [2024-11-27 17:29:02.861351] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:31.834 17:29:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@974 -- # wait 71901 00:07:31.834 [2024-11-27 17:29:02.862930] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:32.093 17:29:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:07:32.094 00:07:32.094 real 0m3.957s 00:07:32.094 user 0m5.999s 00:07:32.094 sys 0m0.892s 00:07:32.094 17:29:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:07:32.094 17:29:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:32.094 ************************************ 00:07:32.094 END TEST raid_state_function_test 00:07:32.094 ************************************ 00:07:32.354 17:29:03 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test raid0 2 true 00:07:32.354 17:29:03 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:07:32.354 17:29:03 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:07:32.354 17:29:03 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:32.354 ************************************ 00:07:32.354 START TEST raid_state_function_test_sb 00:07:32.354 ************************************ 00:07:32.354 17:29:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1125 -- # raid_state_function_test raid0 2 true 00:07:32.354 17:29:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=raid0 00:07:32.354 17:29:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:07:32.354 17:29:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:07:32.354 17:29:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:07:32.354 17:29:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:07:32.354 17:29:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:32.354 17:29:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:07:32.354 17:29:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:07:32.354 17:29:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:32.354 17:29:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:07:32.354 17:29:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:07:32.354 17:29:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:32.354 17:29:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:07:32.354 17:29:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:07:32.354 17:29:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:07:32.354 17:29:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:07:32.354 17:29:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:07:32.354 17:29:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:07:32.354 17:29:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' raid0 '!=' raid1 ']' 00:07:32.354 17:29:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:07:32.354 17:29:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:07:32.354 17:29:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:07:32.354 17:29:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:07:32.354 17:29:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=72143 00:07:32.354 17:29:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:07:32.354 17:29:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 72143' 00:07:32.354 Process raid pid: 72143 00:07:32.354 17:29:03 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 72143 00:07:32.354 17:29:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@831 -- # '[' -z 72143 ']' 00:07:32.354 17:29:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:32.354 17:29:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@836 -- # local max_retries=100 00:07:32.354 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:32.354 17:29:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:32.354 17:29:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # xtrace_disable 00:07:32.354 17:29:03 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:32.354 [2024-11-27 17:29:03.418831] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:07:32.354 [2024-11-27 17:29:03.419019] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:07:32.614 [2024-11-27 17:29:03.566736] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:32.615 [2024-11-27 17:29:03.637630] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:07:32.615 [2024-11-27 17:29:03.713254] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:32.615 [2024-11-27 17:29:03.713303] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:33.183 17:29:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:07:33.183 17:29:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@864 -- # return 0 00:07:33.183 17:29:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:33.183 17:29:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:33.183 17:29:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:33.183 [2024-11-27 17:29:04.248918] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:07:33.183 [2024-11-27 17:29:04.248974] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:07:33.183 [2024-11-27 17:29:04.248988] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:33.183 [2024-11-27 17:29:04.248998] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:33.183 17:29:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:33.183 17:29:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 2 00:07:33.183 17:29:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:33.183 17:29:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:33.183 17:29:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:33.183 17:29:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:33.183 17:29:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:33.184 17:29:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:33.184 17:29:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:33.184 17:29:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:33.184 17:29:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:33.184 17:29:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:33.184 17:29:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:33.184 17:29:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:33.184 17:29:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:33.184 17:29:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:33.184 17:29:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:33.184 "name": "Existed_Raid", 00:07:33.184 "uuid": "cf0cfb00-5433-48a7-82c2-cae25812bd7c", 00:07:33.184 "strip_size_kb": 64, 00:07:33.184 "state": "configuring", 00:07:33.184 "raid_level": "raid0", 00:07:33.184 "superblock": true, 00:07:33.184 "num_base_bdevs": 2, 00:07:33.184 "num_base_bdevs_discovered": 0, 00:07:33.184 "num_base_bdevs_operational": 2, 00:07:33.184 "base_bdevs_list": [ 00:07:33.184 { 00:07:33.184 "name": "BaseBdev1", 00:07:33.184 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:33.184 "is_configured": false, 00:07:33.184 "data_offset": 0, 00:07:33.184 "data_size": 0 00:07:33.184 }, 00:07:33.184 { 00:07:33.184 "name": "BaseBdev2", 00:07:33.184 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:33.184 "is_configured": false, 00:07:33.184 "data_offset": 0, 00:07:33.184 "data_size": 0 00:07:33.184 } 00:07:33.184 ] 00:07:33.184 }' 00:07:33.184 17:29:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:33.184 17:29:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:33.754 17:29:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:07:33.754 17:29:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:33.754 17:29:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:33.754 [2024-11-27 17:29:04.688029] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:07:33.754 [2024-11-27 17:29:04.688082] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:07:33.754 17:29:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:33.754 17:29:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:33.754 17:29:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:33.754 17:29:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:33.754 [2024-11-27 17:29:04.696059] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:07:33.754 [2024-11-27 17:29:04.696100] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:07:33.754 [2024-11-27 17:29:04.696133] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:33.754 [2024-11-27 17:29:04.696144] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:33.754 17:29:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:33.754 17:29:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:07:33.754 17:29:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:33.754 17:29:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:33.754 [2024-11-27 17:29:04.719074] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:33.754 BaseBdev1 00:07:33.754 17:29:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:33.754 17:29:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:07:33.754 17:29:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:07:33.754 17:29:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:07:33.754 17:29:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:07:33.754 17:29:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:07:33.754 17:29:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:07:33.754 17:29:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:07:33.754 17:29:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:33.754 17:29:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:33.754 17:29:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:33.754 17:29:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:07:33.754 17:29:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:33.754 17:29:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:33.754 [ 00:07:33.754 { 00:07:33.754 "name": "BaseBdev1", 00:07:33.754 "aliases": [ 00:07:33.754 "062f4015-c869-4a5e-9fe6-8310e1d66d9b" 00:07:33.755 ], 00:07:33.755 "product_name": "Malloc disk", 00:07:33.755 "block_size": 512, 00:07:33.755 "num_blocks": 65536, 00:07:33.755 "uuid": "062f4015-c869-4a5e-9fe6-8310e1d66d9b", 00:07:33.755 "assigned_rate_limits": { 00:07:33.755 "rw_ios_per_sec": 0, 00:07:33.755 "rw_mbytes_per_sec": 0, 00:07:33.755 "r_mbytes_per_sec": 0, 00:07:33.755 "w_mbytes_per_sec": 0 00:07:33.755 }, 00:07:33.755 "claimed": true, 00:07:33.755 "claim_type": "exclusive_write", 00:07:33.755 "zoned": false, 00:07:33.755 "supported_io_types": { 00:07:33.755 "read": true, 00:07:33.755 "write": true, 00:07:33.755 "unmap": true, 00:07:33.755 "flush": true, 00:07:33.755 "reset": true, 00:07:33.755 "nvme_admin": false, 00:07:33.755 "nvme_io": false, 00:07:33.755 "nvme_io_md": false, 00:07:33.755 "write_zeroes": true, 00:07:33.755 "zcopy": true, 00:07:33.755 "get_zone_info": false, 00:07:33.755 "zone_management": false, 00:07:33.755 "zone_append": false, 00:07:33.755 "compare": false, 00:07:33.755 "compare_and_write": false, 00:07:33.755 "abort": true, 00:07:33.755 "seek_hole": false, 00:07:33.755 "seek_data": false, 00:07:33.755 "copy": true, 00:07:33.755 "nvme_iov_md": false 00:07:33.755 }, 00:07:33.755 "memory_domains": [ 00:07:33.755 { 00:07:33.755 "dma_device_id": "system", 00:07:33.755 "dma_device_type": 1 00:07:33.755 }, 00:07:33.755 { 00:07:33.755 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:33.755 "dma_device_type": 2 00:07:33.755 } 00:07:33.755 ], 00:07:33.755 "driver_specific": {} 00:07:33.755 } 00:07:33.755 ] 00:07:33.755 17:29:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:33.755 17:29:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:07:33.755 17:29:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 2 00:07:33.755 17:29:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:33.755 17:29:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:33.755 17:29:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:33.755 17:29:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:33.755 17:29:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:33.755 17:29:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:33.755 17:29:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:33.755 17:29:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:33.755 17:29:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:33.755 17:29:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:33.755 17:29:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:33.755 17:29:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:33.755 17:29:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:33.755 17:29:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:33.755 17:29:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:33.755 "name": "Existed_Raid", 00:07:33.755 "uuid": "62105695-7fa1-4f2b-979d-86e1da422e0b", 00:07:33.755 "strip_size_kb": 64, 00:07:33.755 "state": "configuring", 00:07:33.755 "raid_level": "raid0", 00:07:33.755 "superblock": true, 00:07:33.755 "num_base_bdevs": 2, 00:07:33.755 "num_base_bdevs_discovered": 1, 00:07:33.755 "num_base_bdevs_operational": 2, 00:07:33.755 "base_bdevs_list": [ 00:07:33.755 { 00:07:33.755 "name": "BaseBdev1", 00:07:33.755 "uuid": "062f4015-c869-4a5e-9fe6-8310e1d66d9b", 00:07:33.755 "is_configured": true, 00:07:33.755 "data_offset": 2048, 00:07:33.755 "data_size": 63488 00:07:33.755 }, 00:07:33.755 { 00:07:33.755 "name": "BaseBdev2", 00:07:33.755 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:33.755 "is_configured": false, 00:07:33.755 "data_offset": 0, 00:07:33.755 "data_size": 0 00:07:33.755 } 00:07:33.755 ] 00:07:33.755 }' 00:07:33.755 17:29:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:33.755 17:29:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:34.016 17:29:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:07:34.016 17:29:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:34.016 17:29:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:34.016 [2024-11-27 17:29:05.186336] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:07:34.016 [2024-11-27 17:29:05.186389] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:07:34.016 17:29:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:34.016 17:29:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:34.016 17:29:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:34.016 17:29:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:34.016 [2024-11-27 17:29:05.194382] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:34.016 [2024-11-27 17:29:05.196543] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:34.016 [2024-11-27 17:29:05.196582] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:34.016 17:29:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:34.016 17:29:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:07:34.016 17:29:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:07:34.016 17:29:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 2 00:07:34.016 17:29:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:34.016 17:29:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:34.016 17:29:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:34.016 17:29:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:34.016 17:29:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:34.016 17:29:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:34.016 17:29:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:34.016 17:29:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:34.016 17:29:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:34.016 17:29:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:34.016 17:29:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:34.016 17:29:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:34.275 17:29:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:34.275 17:29:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:34.275 17:29:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:34.275 "name": "Existed_Raid", 00:07:34.275 "uuid": "a8168e96-813d-40b9-8b3b-e48d421ed65e", 00:07:34.275 "strip_size_kb": 64, 00:07:34.275 "state": "configuring", 00:07:34.275 "raid_level": "raid0", 00:07:34.275 "superblock": true, 00:07:34.275 "num_base_bdevs": 2, 00:07:34.275 "num_base_bdevs_discovered": 1, 00:07:34.275 "num_base_bdevs_operational": 2, 00:07:34.275 "base_bdevs_list": [ 00:07:34.275 { 00:07:34.275 "name": "BaseBdev1", 00:07:34.275 "uuid": "062f4015-c869-4a5e-9fe6-8310e1d66d9b", 00:07:34.275 "is_configured": true, 00:07:34.275 "data_offset": 2048, 00:07:34.275 "data_size": 63488 00:07:34.275 }, 00:07:34.275 { 00:07:34.275 "name": "BaseBdev2", 00:07:34.275 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:34.275 "is_configured": false, 00:07:34.275 "data_offset": 0, 00:07:34.275 "data_size": 0 00:07:34.275 } 00:07:34.275 ] 00:07:34.275 }' 00:07:34.275 17:29:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:34.275 17:29:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:34.535 17:29:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:07:34.535 17:29:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:34.535 17:29:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:34.535 [2024-11-27 17:29:05.652089] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:07:34.535 [2024-11-27 17:29:05.652628] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:07:34.535 [2024-11-27 17:29:05.652688] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:07:34.535 BaseBdev2 00:07:34.535 [2024-11-27 17:29:05.653543] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:07:34.535 17:29:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:34.535 [2024-11-27 17:29:05.654023] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:07:34.535 [2024-11-27 17:29:05.654075] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:07:34.535 17:29:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:07:34.535 [2024-11-27 17:29:05.654479] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:34.535 17:29:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:07:34.535 17:29:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:07:34.535 17:29:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:07:34.536 17:29:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:07:34.536 17:29:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:07:34.536 17:29:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:07:34.536 17:29:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:34.536 17:29:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:34.536 17:29:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:34.536 17:29:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:07:34.536 17:29:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:34.536 17:29:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:34.536 [ 00:07:34.536 { 00:07:34.536 "name": "BaseBdev2", 00:07:34.536 "aliases": [ 00:07:34.536 "17caa770-dffc-4cb0-a4c7-19f855b3a190" 00:07:34.536 ], 00:07:34.536 "product_name": "Malloc disk", 00:07:34.536 "block_size": 512, 00:07:34.536 "num_blocks": 65536, 00:07:34.536 "uuid": "17caa770-dffc-4cb0-a4c7-19f855b3a190", 00:07:34.536 "assigned_rate_limits": { 00:07:34.536 "rw_ios_per_sec": 0, 00:07:34.536 "rw_mbytes_per_sec": 0, 00:07:34.536 "r_mbytes_per_sec": 0, 00:07:34.536 "w_mbytes_per_sec": 0 00:07:34.536 }, 00:07:34.536 "claimed": true, 00:07:34.536 "claim_type": "exclusive_write", 00:07:34.536 "zoned": false, 00:07:34.536 "supported_io_types": { 00:07:34.536 "read": true, 00:07:34.536 "write": true, 00:07:34.536 "unmap": true, 00:07:34.536 "flush": true, 00:07:34.536 "reset": true, 00:07:34.536 "nvme_admin": false, 00:07:34.536 "nvme_io": false, 00:07:34.536 "nvme_io_md": false, 00:07:34.536 "write_zeroes": true, 00:07:34.536 "zcopy": true, 00:07:34.536 "get_zone_info": false, 00:07:34.536 "zone_management": false, 00:07:34.536 "zone_append": false, 00:07:34.536 "compare": false, 00:07:34.536 "compare_and_write": false, 00:07:34.536 "abort": true, 00:07:34.536 "seek_hole": false, 00:07:34.536 "seek_data": false, 00:07:34.536 "copy": true, 00:07:34.536 "nvme_iov_md": false 00:07:34.536 }, 00:07:34.536 "memory_domains": [ 00:07:34.536 { 00:07:34.536 "dma_device_id": "system", 00:07:34.536 "dma_device_type": 1 00:07:34.536 }, 00:07:34.536 { 00:07:34.536 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:34.536 "dma_device_type": 2 00:07:34.536 } 00:07:34.536 ], 00:07:34.536 "driver_specific": {} 00:07:34.536 } 00:07:34.536 ] 00:07:34.536 17:29:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:34.536 17:29:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:07:34.536 17:29:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:07:34.536 17:29:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:07:34.536 17:29:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid0 64 2 00:07:34.536 17:29:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:34.536 17:29:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:34.536 17:29:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:34.536 17:29:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:34.536 17:29:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:34.536 17:29:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:34.536 17:29:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:34.536 17:29:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:34.536 17:29:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:34.536 17:29:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:34.536 17:29:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:34.536 17:29:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:34.536 17:29:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:34.536 17:29:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:34.796 17:29:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:34.796 "name": "Existed_Raid", 00:07:34.796 "uuid": "a8168e96-813d-40b9-8b3b-e48d421ed65e", 00:07:34.796 "strip_size_kb": 64, 00:07:34.796 "state": "online", 00:07:34.796 "raid_level": "raid0", 00:07:34.796 "superblock": true, 00:07:34.796 "num_base_bdevs": 2, 00:07:34.796 "num_base_bdevs_discovered": 2, 00:07:34.796 "num_base_bdevs_operational": 2, 00:07:34.796 "base_bdevs_list": [ 00:07:34.796 { 00:07:34.796 "name": "BaseBdev1", 00:07:34.796 "uuid": "062f4015-c869-4a5e-9fe6-8310e1d66d9b", 00:07:34.796 "is_configured": true, 00:07:34.796 "data_offset": 2048, 00:07:34.796 "data_size": 63488 00:07:34.796 }, 00:07:34.796 { 00:07:34.796 "name": "BaseBdev2", 00:07:34.796 "uuid": "17caa770-dffc-4cb0-a4c7-19f855b3a190", 00:07:34.796 "is_configured": true, 00:07:34.796 "data_offset": 2048, 00:07:34.796 "data_size": 63488 00:07:34.796 } 00:07:34.796 ] 00:07:34.796 }' 00:07:34.796 17:29:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:34.796 17:29:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:35.055 17:29:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:07:35.055 17:29:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:07:35.055 17:29:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:07:35.055 17:29:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:07:35.056 17:29:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:07:35.056 17:29:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:07:35.056 17:29:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:07:35.056 17:29:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:35.056 17:29:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:35.056 17:29:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:07:35.056 [2024-11-27 17:29:06.115538] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:35.056 17:29:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:35.056 17:29:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:07:35.056 "name": "Existed_Raid", 00:07:35.056 "aliases": [ 00:07:35.056 "a8168e96-813d-40b9-8b3b-e48d421ed65e" 00:07:35.056 ], 00:07:35.056 "product_name": "Raid Volume", 00:07:35.056 "block_size": 512, 00:07:35.056 "num_blocks": 126976, 00:07:35.056 "uuid": "a8168e96-813d-40b9-8b3b-e48d421ed65e", 00:07:35.056 "assigned_rate_limits": { 00:07:35.056 "rw_ios_per_sec": 0, 00:07:35.056 "rw_mbytes_per_sec": 0, 00:07:35.056 "r_mbytes_per_sec": 0, 00:07:35.056 "w_mbytes_per_sec": 0 00:07:35.056 }, 00:07:35.056 "claimed": false, 00:07:35.056 "zoned": false, 00:07:35.056 "supported_io_types": { 00:07:35.056 "read": true, 00:07:35.056 "write": true, 00:07:35.056 "unmap": true, 00:07:35.056 "flush": true, 00:07:35.056 "reset": true, 00:07:35.056 "nvme_admin": false, 00:07:35.056 "nvme_io": false, 00:07:35.056 "nvme_io_md": false, 00:07:35.056 "write_zeroes": true, 00:07:35.056 "zcopy": false, 00:07:35.056 "get_zone_info": false, 00:07:35.056 "zone_management": false, 00:07:35.056 "zone_append": false, 00:07:35.056 "compare": false, 00:07:35.056 "compare_and_write": false, 00:07:35.056 "abort": false, 00:07:35.056 "seek_hole": false, 00:07:35.056 "seek_data": false, 00:07:35.056 "copy": false, 00:07:35.056 "nvme_iov_md": false 00:07:35.056 }, 00:07:35.056 "memory_domains": [ 00:07:35.056 { 00:07:35.056 "dma_device_id": "system", 00:07:35.056 "dma_device_type": 1 00:07:35.056 }, 00:07:35.056 { 00:07:35.056 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:35.056 "dma_device_type": 2 00:07:35.056 }, 00:07:35.056 { 00:07:35.056 "dma_device_id": "system", 00:07:35.056 "dma_device_type": 1 00:07:35.056 }, 00:07:35.056 { 00:07:35.056 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:35.056 "dma_device_type": 2 00:07:35.056 } 00:07:35.056 ], 00:07:35.056 "driver_specific": { 00:07:35.056 "raid": { 00:07:35.056 "uuid": "a8168e96-813d-40b9-8b3b-e48d421ed65e", 00:07:35.056 "strip_size_kb": 64, 00:07:35.056 "state": "online", 00:07:35.056 "raid_level": "raid0", 00:07:35.056 "superblock": true, 00:07:35.056 "num_base_bdevs": 2, 00:07:35.056 "num_base_bdevs_discovered": 2, 00:07:35.056 "num_base_bdevs_operational": 2, 00:07:35.056 "base_bdevs_list": [ 00:07:35.056 { 00:07:35.056 "name": "BaseBdev1", 00:07:35.056 "uuid": "062f4015-c869-4a5e-9fe6-8310e1d66d9b", 00:07:35.056 "is_configured": true, 00:07:35.056 "data_offset": 2048, 00:07:35.056 "data_size": 63488 00:07:35.056 }, 00:07:35.056 { 00:07:35.056 "name": "BaseBdev2", 00:07:35.056 "uuid": "17caa770-dffc-4cb0-a4c7-19f855b3a190", 00:07:35.056 "is_configured": true, 00:07:35.056 "data_offset": 2048, 00:07:35.056 "data_size": 63488 00:07:35.056 } 00:07:35.056 ] 00:07:35.056 } 00:07:35.056 } 00:07:35.056 }' 00:07:35.056 17:29:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:07:35.056 17:29:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:07:35.056 BaseBdev2' 00:07:35.056 17:29:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:35.056 17:29:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:07:35.056 17:29:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:35.056 17:29:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:07:35.056 17:29:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:35.056 17:29:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:35.056 17:29:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:35.316 17:29:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:35.316 17:29:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:35.316 17:29:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:35.316 17:29:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:35.316 17:29:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:07:35.316 17:29:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:35.316 17:29:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:35.316 17:29:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:35.316 17:29:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:35.316 17:29:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:35.316 17:29:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:35.316 17:29:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:07:35.316 17:29:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:35.316 17:29:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:35.316 [2024-11-27 17:29:06.338926] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:07:35.316 [2024-11-27 17:29:06.338966] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:35.316 [2024-11-27 17:29:06.339024] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:35.316 17:29:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:35.316 17:29:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:07:35.316 17:29:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy raid0 00:07:35.316 17:29:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:07:35.316 17:29:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # return 1 00:07:35.316 17:29:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:07:35.316 17:29:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline raid0 64 1 00:07:35.316 17:29:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:35.316 17:29:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:07:35.316 17:29:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:35.316 17:29:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:35.316 17:29:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:07:35.316 17:29:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:35.316 17:29:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:35.316 17:29:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:35.316 17:29:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:35.316 17:29:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:35.316 17:29:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:35.316 17:29:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:35.316 17:29:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:35.316 17:29:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:35.316 17:29:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:35.316 "name": "Existed_Raid", 00:07:35.316 "uuid": "a8168e96-813d-40b9-8b3b-e48d421ed65e", 00:07:35.316 "strip_size_kb": 64, 00:07:35.316 "state": "offline", 00:07:35.316 "raid_level": "raid0", 00:07:35.316 "superblock": true, 00:07:35.316 "num_base_bdevs": 2, 00:07:35.316 "num_base_bdevs_discovered": 1, 00:07:35.316 "num_base_bdevs_operational": 1, 00:07:35.316 "base_bdevs_list": [ 00:07:35.316 { 00:07:35.316 "name": null, 00:07:35.316 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:35.316 "is_configured": false, 00:07:35.317 "data_offset": 0, 00:07:35.317 "data_size": 63488 00:07:35.317 }, 00:07:35.317 { 00:07:35.317 "name": "BaseBdev2", 00:07:35.317 "uuid": "17caa770-dffc-4cb0-a4c7-19f855b3a190", 00:07:35.317 "is_configured": true, 00:07:35.317 "data_offset": 2048, 00:07:35.317 "data_size": 63488 00:07:35.317 } 00:07:35.317 ] 00:07:35.317 }' 00:07:35.317 17:29:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:35.317 17:29:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:35.576 17:29:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:07:35.576 17:29:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:07:35.576 17:29:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:35.576 17:29:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:07:35.576 17:29:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:35.576 17:29:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:35.576 17:29:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:35.835 17:29:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:07:35.835 17:29:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:07:35.835 17:29:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:07:35.835 17:29:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:35.835 17:29:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:35.835 [2024-11-27 17:29:06.795186] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:07:35.835 [2024-11-27 17:29:06.795246] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:07:35.835 17:29:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:35.835 17:29:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:07:35.835 17:29:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:07:35.835 17:29:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:35.835 17:29:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:07:35.835 17:29:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:35.836 17:29:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:35.836 17:29:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:35.836 17:29:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:07:35.836 17:29:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:07:35.836 17:29:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:07:35.836 17:29:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 72143 00:07:35.836 17:29:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@950 -- # '[' -z 72143 ']' 00:07:35.836 17:29:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # kill -0 72143 00:07:35.836 17:29:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@955 -- # uname 00:07:35.836 17:29:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:07:35.836 17:29:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 72143 00:07:35.836 17:29:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:07:35.836 17:29:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:07:35.836 killing process with pid 72143 00:07:35.836 17:29:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@968 -- # echo 'killing process with pid 72143' 00:07:35.836 17:29:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@969 -- # kill 72143 00:07:35.836 [2024-11-27 17:29:06.916993] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:35.836 17:29:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@974 -- # wait 72143 00:07:35.836 [2024-11-27 17:29:06.918561] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:36.096 17:29:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:07:36.096 00:07:36.096 real 0m3.966s 00:07:36.096 user 0m6.034s 00:07:36.096 sys 0m0.877s 00:07:36.096 17:29:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1126 -- # xtrace_disable 00:07:36.096 17:29:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:36.096 ************************************ 00:07:36.096 END TEST raid_state_function_test_sb 00:07:36.096 ************************************ 00:07:36.357 17:29:07 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test raid0 2 00:07:36.357 17:29:07 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:07:36.357 17:29:07 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:07:36.357 17:29:07 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:36.357 ************************************ 00:07:36.357 START TEST raid_superblock_test 00:07:36.357 ************************************ 00:07:36.357 17:29:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1125 -- # raid_superblock_test raid0 2 00:07:36.357 17:29:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=raid0 00:07:36.357 17:29:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=2 00:07:36.357 17:29:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:07:36.357 17:29:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:07:36.357 17:29:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:07:36.357 17:29:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:07:36.357 17:29:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:07:36.357 17:29:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:07:36.357 17:29:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:07:36.357 17:29:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:07:36.357 17:29:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:07:36.357 17:29:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:07:36.357 17:29:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:07:36.357 17:29:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' raid0 '!=' raid1 ']' 00:07:36.357 17:29:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@405 -- # strip_size=64 00:07:36.357 17:29:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@406 -- # strip_size_create_arg='-z 64' 00:07:36.357 17:29:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=72384 00:07:36.357 17:29:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 72384 00:07:36.357 17:29:07 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:07:36.357 17:29:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@831 -- # '[' -z 72384 ']' 00:07:36.357 17:29:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:36.357 17:29:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:07:36.357 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:36.357 17:29:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:36.357 17:29:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:07:36.357 17:29:07 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:36.357 [2024-11-27 17:29:07.446340] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:07:36.357 [2024-11-27 17:29:07.446486] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid72384 ] 00:07:36.617 [2024-11-27 17:29:07.592456] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:36.617 [2024-11-27 17:29:07.660612] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:07:36.617 [2024-11-27 17:29:07.739420] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:36.617 [2024-11-27 17:29:07.739478] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:37.186 17:29:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:07:37.186 17:29:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@864 -- # return 0 00:07:37.186 17:29:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:07:37.186 17:29:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:07:37.186 17:29:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:07:37.186 17:29:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:07:37.186 17:29:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:07:37.186 17:29:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:07:37.186 17:29:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:07:37.186 17:29:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:07:37.186 17:29:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:07:37.186 17:29:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:37.186 17:29:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:37.186 malloc1 00:07:37.186 17:29:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:37.186 17:29:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:07:37.186 17:29:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:37.186 17:29:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:37.186 [2024-11-27 17:29:08.298873] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:07:37.186 [2024-11-27 17:29:08.298942] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:37.186 [2024-11-27 17:29:08.298961] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:07:37.186 [2024-11-27 17:29:08.298976] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:37.186 [2024-11-27 17:29:08.301409] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:37.186 [2024-11-27 17:29:08.301449] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:07:37.186 pt1 00:07:37.186 17:29:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:37.186 17:29:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:07:37.186 17:29:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:07:37.186 17:29:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:07:37.186 17:29:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:07:37.186 17:29:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:07:37.186 17:29:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:07:37.186 17:29:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:07:37.186 17:29:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:07:37.186 17:29:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:07:37.186 17:29:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:37.186 17:29:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:37.186 malloc2 00:07:37.186 17:29:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:37.186 17:29:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:07:37.186 17:29:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:37.186 17:29:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:37.186 [2024-11-27 17:29:08.348216] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:07:37.186 [2024-11-27 17:29:08.348301] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:37.186 [2024-11-27 17:29:08.348335] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:07:37.186 [2024-11-27 17:29:08.348359] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:37.187 [2024-11-27 17:29:08.353091] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:37.187 [2024-11-27 17:29:08.353181] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:07:37.187 pt2 00:07:37.187 17:29:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:37.187 17:29:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:07:37.187 17:29:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:07:37.187 17:29:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''pt1 pt2'\''' -n raid_bdev1 -s 00:07:37.187 17:29:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:37.187 17:29:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:37.187 [2024-11-27 17:29:08.361488] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:07:37.187 [2024-11-27 17:29:08.364622] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:07:37.187 [2024-11-27 17:29:08.364839] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:07:37.187 [2024-11-27 17:29:08.364872] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:07:37.187 [2024-11-27 17:29:08.365316] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:07:37.187 [2024-11-27 17:29:08.365535] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:07:37.187 [2024-11-27 17:29:08.365559] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:07:37.187 [2024-11-27 17:29:08.365797] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:37.187 17:29:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:37.187 17:29:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 2 00:07:37.187 17:29:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:37.187 17:29:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:37.187 17:29:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:37.187 17:29:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:37.187 17:29:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:37.187 17:29:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:37.187 17:29:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:37.187 17:29:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:37.187 17:29:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:37.187 17:29:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:37.187 17:29:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:37.187 17:29:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:37.187 17:29:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:37.446 17:29:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:37.446 17:29:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:37.446 "name": "raid_bdev1", 00:07:37.446 "uuid": "aab0478f-d039-488a-96ee-c1d562883896", 00:07:37.446 "strip_size_kb": 64, 00:07:37.446 "state": "online", 00:07:37.446 "raid_level": "raid0", 00:07:37.446 "superblock": true, 00:07:37.446 "num_base_bdevs": 2, 00:07:37.446 "num_base_bdevs_discovered": 2, 00:07:37.446 "num_base_bdevs_operational": 2, 00:07:37.446 "base_bdevs_list": [ 00:07:37.446 { 00:07:37.446 "name": "pt1", 00:07:37.446 "uuid": "00000000-0000-0000-0000-000000000001", 00:07:37.446 "is_configured": true, 00:07:37.446 "data_offset": 2048, 00:07:37.446 "data_size": 63488 00:07:37.446 }, 00:07:37.446 { 00:07:37.446 "name": "pt2", 00:07:37.446 "uuid": "00000000-0000-0000-0000-000000000002", 00:07:37.446 "is_configured": true, 00:07:37.446 "data_offset": 2048, 00:07:37.446 "data_size": 63488 00:07:37.446 } 00:07:37.446 ] 00:07:37.446 }' 00:07:37.446 17:29:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:37.446 17:29:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:37.706 17:29:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:07:37.706 17:29:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:07:37.706 17:29:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:07:37.706 17:29:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:07:37.706 17:29:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:07:37.706 17:29:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:07:37.706 17:29:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:07:37.706 17:29:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:37.706 17:29:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:07:37.706 17:29:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:37.706 [2024-11-27 17:29:08.825276] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:37.706 17:29:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:37.706 17:29:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:07:37.706 "name": "raid_bdev1", 00:07:37.706 "aliases": [ 00:07:37.706 "aab0478f-d039-488a-96ee-c1d562883896" 00:07:37.706 ], 00:07:37.706 "product_name": "Raid Volume", 00:07:37.706 "block_size": 512, 00:07:37.706 "num_blocks": 126976, 00:07:37.706 "uuid": "aab0478f-d039-488a-96ee-c1d562883896", 00:07:37.706 "assigned_rate_limits": { 00:07:37.706 "rw_ios_per_sec": 0, 00:07:37.706 "rw_mbytes_per_sec": 0, 00:07:37.706 "r_mbytes_per_sec": 0, 00:07:37.706 "w_mbytes_per_sec": 0 00:07:37.706 }, 00:07:37.706 "claimed": false, 00:07:37.706 "zoned": false, 00:07:37.706 "supported_io_types": { 00:07:37.706 "read": true, 00:07:37.706 "write": true, 00:07:37.706 "unmap": true, 00:07:37.706 "flush": true, 00:07:37.706 "reset": true, 00:07:37.706 "nvme_admin": false, 00:07:37.706 "nvme_io": false, 00:07:37.706 "nvme_io_md": false, 00:07:37.706 "write_zeroes": true, 00:07:37.706 "zcopy": false, 00:07:37.706 "get_zone_info": false, 00:07:37.706 "zone_management": false, 00:07:37.706 "zone_append": false, 00:07:37.706 "compare": false, 00:07:37.706 "compare_and_write": false, 00:07:37.706 "abort": false, 00:07:37.706 "seek_hole": false, 00:07:37.706 "seek_data": false, 00:07:37.706 "copy": false, 00:07:37.706 "nvme_iov_md": false 00:07:37.706 }, 00:07:37.706 "memory_domains": [ 00:07:37.706 { 00:07:37.706 "dma_device_id": "system", 00:07:37.706 "dma_device_type": 1 00:07:37.706 }, 00:07:37.706 { 00:07:37.706 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:37.706 "dma_device_type": 2 00:07:37.706 }, 00:07:37.706 { 00:07:37.706 "dma_device_id": "system", 00:07:37.706 "dma_device_type": 1 00:07:37.706 }, 00:07:37.706 { 00:07:37.706 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:37.706 "dma_device_type": 2 00:07:37.706 } 00:07:37.706 ], 00:07:37.706 "driver_specific": { 00:07:37.706 "raid": { 00:07:37.706 "uuid": "aab0478f-d039-488a-96ee-c1d562883896", 00:07:37.706 "strip_size_kb": 64, 00:07:37.706 "state": "online", 00:07:37.706 "raid_level": "raid0", 00:07:37.706 "superblock": true, 00:07:37.706 "num_base_bdevs": 2, 00:07:37.706 "num_base_bdevs_discovered": 2, 00:07:37.706 "num_base_bdevs_operational": 2, 00:07:37.706 "base_bdevs_list": [ 00:07:37.706 { 00:07:37.706 "name": "pt1", 00:07:37.706 "uuid": "00000000-0000-0000-0000-000000000001", 00:07:37.706 "is_configured": true, 00:07:37.706 "data_offset": 2048, 00:07:37.706 "data_size": 63488 00:07:37.706 }, 00:07:37.706 { 00:07:37.706 "name": "pt2", 00:07:37.706 "uuid": "00000000-0000-0000-0000-000000000002", 00:07:37.706 "is_configured": true, 00:07:37.706 "data_offset": 2048, 00:07:37.706 "data_size": 63488 00:07:37.706 } 00:07:37.706 ] 00:07:37.706 } 00:07:37.706 } 00:07:37.706 }' 00:07:37.706 17:29:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:07:37.966 17:29:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:07:37.966 pt2' 00:07:37.966 17:29:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:37.966 17:29:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:07:37.966 17:29:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:37.966 17:29:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:07:37.966 17:29:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:37.966 17:29:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:37.966 17:29:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:37.966 17:29:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:37.966 17:29:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:37.967 17:29:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:37.967 17:29:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:37.967 17:29:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:07:37.967 17:29:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:37.967 17:29:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:37.967 17:29:08 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:37.967 17:29:08 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:37.967 17:29:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:37.967 17:29:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:37.967 17:29:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:07:37.967 17:29:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:07:37.967 17:29:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:37.967 17:29:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:37.967 [2024-11-27 17:29:09.032779] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:37.967 17:29:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:37.967 17:29:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=aab0478f-d039-488a-96ee-c1d562883896 00:07:37.967 17:29:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z aab0478f-d039-488a-96ee-c1d562883896 ']' 00:07:37.967 17:29:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:07:37.967 17:29:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:37.967 17:29:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:37.967 [2024-11-27 17:29:09.060492] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:07:37.967 [2024-11-27 17:29:09.060528] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:37.967 [2024-11-27 17:29:09.060605] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:37.967 [2024-11-27 17:29:09.060680] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:37.967 [2024-11-27 17:29:09.060693] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:07:37.967 17:29:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:37.967 17:29:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:37.967 17:29:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:37.967 17:29:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:37.967 17:29:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:07:37.967 17:29:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:37.967 17:29:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:07:37.967 17:29:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:07:37.967 17:29:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:07:37.967 17:29:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:07:37.967 17:29:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:37.967 17:29:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:37.967 17:29:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:37.967 17:29:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:07:37.967 17:29:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:07:37.967 17:29:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:37.967 17:29:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:37.967 17:29:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:37.967 17:29:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:07:37.967 17:29:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:37.967 17:29:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:07:37.967 17:29:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:38.226 17:29:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:38.226 17:29:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:07:38.226 17:29:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:07:38.226 17:29:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@650 -- # local es=0 00:07:38.226 17:29:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:07:38.226 17:29:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:07:38.226 17:29:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:07:38.226 17:29:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:07:38.226 17:29:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:07:38.226 17:29:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:07:38.226 17:29:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:38.226 17:29:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:38.226 [2024-11-27 17:29:09.204282] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:07:38.226 [2024-11-27 17:29:09.206348] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:07:38.226 [2024-11-27 17:29:09.206426] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:07:38.226 [2024-11-27 17:29:09.206465] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:07:38.226 [2024-11-27 17:29:09.206498] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:07:38.226 [2024-11-27 17:29:09.206514] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state configuring 00:07:38.226 request: 00:07:38.226 { 00:07:38.226 "name": "raid_bdev1", 00:07:38.226 "raid_level": "raid0", 00:07:38.226 "base_bdevs": [ 00:07:38.226 "malloc1", 00:07:38.226 "malloc2" 00:07:38.226 ], 00:07:38.226 "strip_size_kb": 64, 00:07:38.226 "superblock": false, 00:07:38.226 "method": "bdev_raid_create", 00:07:38.226 "req_id": 1 00:07:38.226 } 00:07:38.226 Got JSON-RPC error response 00:07:38.226 response: 00:07:38.226 { 00:07:38.226 "code": -17, 00:07:38.226 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:07:38.226 } 00:07:38.226 17:29:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:07:38.226 17:29:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@653 -- # es=1 00:07:38.226 17:29:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:07:38.226 17:29:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:07:38.226 17:29:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:07:38.226 17:29:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:38.226 17:29:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:38.226 17:29:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:38.226 17:29:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:07:38.226 17:29:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:38.226 17:29:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:07:38.226 17:29:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:07:38.226 17:29:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:07:38.226 17:29:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:38.226 17:29:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:38.226 [2024-11-27 17:29:09.268124] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:07:38.226 [2024-11-27 17:29:09.268180] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:38.226 [2024-11-27 17:29:09.268203] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:07:38.226 [2024-11-27 17:29:09.268212] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:38.226 [2024-11-27 17:29:09.270608] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:38.226 [2024-11-27 17:29:09.270645] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:07:38.226 [2024-11-27 17:29:09.270709] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:07:38.226 [2024-11-27 17:29:09.270755] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:07:38.226 pt1 00:07:38.226 17:29:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:38.226 17:29:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid0 64 2 00:07:38.226 17:29:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:38.226 17:29:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:38.226 17:29:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:38.226 17:29:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:38.226 17:29:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:38.226 17:29:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:38.226 17:29:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:38.226 17:29:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:38.226 17:29:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:38.226 17:29:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:38.226 17:29:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:38.226 17:29:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:38.226 17:29:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:38.226 17:29:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:38.226 17:29:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:38.226 "name": "raid_bdev1", 00:07:38.226 "uuid": "aab0478f-d039-488a-96ee-c1d562883896", 00:07:38.226 "strip_size_kb": 64, 00:07:38.226 "state": "configuring", 00:07:38.226 "raid_level": "raid0", 00:07:38.226 "superblock": true, 00:07:38.226 "num_base_bdevs": 2, 00:07:38.226 "num_base_bdevs_discovered": 1, 00:07:38.226 "num_base_bdevs_operational": 2, 00:07:38.226 "base_bdevs_list": [ 00:07:38.226 { 00:07:38.226 "name": "pt1", 00:07:38.226 "uuid": "00000000-0000-0000-0000-000000000001", 00:07:38.226 "is_configured": true, 00:07:38.226 "data_offset": 2048, 00:07:38.226 "data_size": 63488 00:07:38.226 }, 00:07:38.226 { 00:07:38.226 "name": null, 00:07:38.226 "uuid": "00000000-0000-0000-0000-000000000002", 00:07:38.226 "is_configured": false, 00:07:38.226 "data_offset": 2048, 00:07:38.226 "data_size": 63488 00:07:38.226 } 00:07:38.226 ] 00:07:38.226 }' 00:07:38.226 17:29:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:38.226 17:29:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:38.795 17:29:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 2 -gt 2 ']' 00:07:38.795 17:29:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:07:38.795 17:29:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:07:38.795 17:29:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:07:38.795 17:29:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:38.795 17:29:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:38.795 [2024-11-27 17:29:09.707389] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:07:38.795 [2024-11-27 17:29:09.707440] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:38.795 [2024-11-27 17:29:09.707460] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:07:38.795 [2024-11-27 17:29:09.707469] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:38.795 [2024-11-27 17:29:09.707872] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:38.795 [2024-11-27 17:29:09.707896] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:07:38.795 [2024-11-27 17:29:09.707963] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:07:38.795 [2024-11-27 17:29:09.707986] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:07:38.795 [2024-11-27 17:29:09.708076] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:07:38.795 [2024-11-27 17:29:09.708085] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:07:38.795 [2024-11-27 17:29:09.708357] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:07:38.796 [2024-11-27 17:29:09.708473] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:07:38.796 [2024-11-27 17:29:09.708492] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:07:38.796 [2024-11-27 17:29:09.708591] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:38.796 pt2 00:07:38.796 17:29:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:38.796 17:29:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:07:38.796 17:29:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:07:38.796 17:29:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 2 00:07:38.796 17:29:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:38.796 17:29:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:38.796 17:29:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:38.796 17:29:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:38.796 17:29:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:38.796 17:29:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:38.796 17:29:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:38.796 17:29:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:38.796 17:29:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:38.796 17:29:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:38.796 17:29:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:38.796 17:29:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:38.796 17:29:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:38.796 17:29:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:38.796 17:29:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:38.796 "name": "raid_bdev1", 00:07:38.796 "uuid": "aab0478f-d039-488a-96ee-c1d562883896", 00:07:38.796 "strip_size_kb": 64, 00:07:38.796 "state": "online", 00:07:38.796 "raid_level": "raid0", 00:07:38.796 "superblock": true, 00:07:38.796 "num_base_bdevs": 2, 00:07:38.796 "num_base_bdevs_discovered": 2, 00:07:38.796 "num_base_bdevs_operational": 2, 00:07:38.796 "base_bdevs_list": [ 00:07:38.796 { 00:07:38.796 "name": "pt1", 00:07:38.796 "uuid": "00000000-0000-0000-0000-000000000001", 00:07:38.796 "is_configured": true, 00:07:38.796 "data_offset": 2048, 00:07:38.796 "data_size": 63488 00:07:38.796 }, 00:07:38.796 { 00:07:38.796 "name": "pt2", 00:07:38.796 "uuid": "00000000-0000-0000-0000-000000000002", 00:07:38.796 "is_configured": true, 00:07:38.796 "data_offset": 2048, 00:07:38.796 "data_size": 63488 00:07:38.796 } 00:07:38.796 ] 00:07:38.796 }' 00:07:38.796 17:29:09 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:38.796 17:29:09 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:39.055 17:29:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:07:39.055 17:29:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:07:39.055 17:29:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:07:39.055 17:29:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:07:39.055 17:29:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:07:39.055 17:29:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:07:39.055 17:29:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:07:39.055 17:29:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:07:39.055 17:29:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:39.055 17:29:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:39.055 [2024-11-27 17:29:10.166913] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:39.055 17:29:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:39.055 17:29:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:07:39.055 "name": "raid_bdev1", 00:07:39.055 "aliases": [ 00:07:39.055 "aab0478f-d039-488a-96ee-c1d562883896" 00:07:39.055 ], 00:07:39.055 "product_name": "Raid Volume", 00:07:39.055 "block_size": 512, 00:07:39.055 "num_blocks": 126976, 00:07:39.055 "uuid": "aab0478f-d039-488a-96ee-c1d562883896", 00:07:39.055 "assigned_rate_limits": { 00:07:39.055 "rw_ios_per_sec": 0, 00:07:39.055 "rw_mbytes_per_sec": 0, 00:07:39.055 "r_mbytes_per_sec": 0, 00:07:39.055 "w_mbytes_per_sec": 0 00:07:39.055 }, 00:07:39.055 "claimed": false, 00:07:39.055 "zoned": false, 00:07:39.055 "supported_io_types": { 00:07:39.055 "read": true, 00:07:39.055 "write": true, 00:07:39.055 "unmap": true, 00:07:39.055 "flush": true, 00:07:39.055 "reset": true, 00:07:39.055 "nvme_admin": false, 00:07:39.055 "nvme_io": false, 00:07:39.055 "nvme_io_md": false, 00:07:39.055 "write_zeroes": true, 00:07:39.056 "zcopy": false, 00:07:39.056 "get_zone_info": false, 00:07:39.056 "zone_management": false, 00:07:39.056 "zone_append": false, 00:07:39.056 "compare": false, 00:07:39.056 "compare_and_write": false, 00:07:39.056 "abort": false, 00:07:39.056 "seek_hole": false, 00:07:39.056 "seek_data": false, 00:07:39.056 "copy": false, 00:07:39.056 "nvme_iov_md": false 00:07:39.056 }, 00:07:39.056 "memory_domains": [ 00:07:39.056 { 00:07:39.056 "dma_device_id": "system", 00:07:39.056 "dma_device_type": 1 00:07:39.056 }, 00:07:39.056 { 00:07:39.056 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:39.056 "dma_device_type": 2 00:07:39.056 }, 00:07:39.056 { 00:07:39.056 "dma_device_id": "system", 00:07:39.056 "dma_device_type": 1 00:07:39.056 }, 00:07:39.056 { 00:07:39.056 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:39.056 "dma_device_type": 2 00:07:39.056 } 00:07:39.056 ], 00:07:39.056 "driver_specific": { 00:07:39.056 "raid": { 00:07:39.056 "uuid": "aab0478f-d039-488a-96ee-c1d562883896", 00:07:39.056 "strip_size_kb": 64, 00:07:39.056 "state": "online", 00:07:39.056 "raid_level": "raid0", 00:07:39.056 "superblock": true, 00:07:39.056 "num_base_bdevs": 2, 00:07:39.056 "num_base_bdevs_discovered": 2, 00:07:39.056 "num_base_bdevs_operational": 2, 00:07:39.056 "base_bdevs_list": [ 00:07:39.056 { 00:07:39.056 "name": "pt1", 00:07:39.056 "uuid": "00000000-0000-0000-0000-000000000001", 00:07:39.056 "is_configured": true, 00:07:39.056 "data_offset": 2048, 00:07:39.056 "data_size": 63488 00:07:39.056 }, 00:07:39.056 { 00:07:39.056 "name": "pt2", 00:07:39.056 "uuid": "00000000-0000-0000-0000-000000000002", 00:07:39.056 "is_configured": true, 00:07:39.056 "data_offset": 2048, 00:07:39.056 "data_size": 63488 00:07:39.056 } 00:07:39.056 ] 00:07:39.056 } 00:07:39.056 } 00:07:39.056 }' 00:07:39.056 17:29:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:07:39.316 17:29:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:07:39.316 pt2' 00:07:39.316 17:29:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:39.316 17:29:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:07:39.316 17:29:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:39.316 17:29:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:07:39.316 17:29:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:39.316 17:29:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:39.316 17:29:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:39.316 17:29:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:39.316 17:29:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:39.316 17:29:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:39.316 17:29:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:39.316 17:29:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:39.316 17:29:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:07:39.316 17:29:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:39.316 17:29:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:39.316 17:29:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:39.316 17:29:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:39.316 17:29:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:39.316 17:29:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:07:39.316 17:29:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:39.316 17:29:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:39.316 17:29:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:07:39.316 [2024-11-27 17:29:10.398477] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:39.316 17:29:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:39.316 17:29:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' aab0478f-d039-488a-96ee-c1d562883896 '!=' aab0478f-d039-488a-96ee-c1d562883896 ']' 00:07:39.316 17:29:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy raid0 00:07:39.316 17:29:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:07:39.316 17:29:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # return 1 00:07:39.316 17:29:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 72384 00:07:39.316 17:29:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@950 -- # '[' -z 72384 ']' 00:07:39.316 17:29:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # kill -0 72384 00:07:39.316 17:29:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@955 -- # uname 00:07:39.316 17:29:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:07:39.317 17:29:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 72384 00:07:39.317 17:29:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:07:39.317 17:29:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:07:39.317 killing process with pid 72384 00:07:39.317 17:29:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 72384' 00:07:39.317 17:29:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@969 -- # kill 72384 00:07:39.317 [2024-11-27 17:29:10.479796] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:39.317 [2024-11-27 17:29:10.479881] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:39.317 [2024-11-27 17:29:10.479930] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:39.317 [2024-11-27 17:29:10.479939] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:07:39.317 17:29:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@974 -- # wait 72384 00:07:39.577 [2024-11-27 17:29:10.520399] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:39.837 17:29:10 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:07:39.837 00:07:39.837 real 0m3.530s 00:07:39.837 user 0m5.246s 00:07:39.837 sys 0m0.818s 00:07:39.837 17:29:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:07:39.837 17:29:10 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:39.837 ************************************ 00:07:39.837 END TEST raid_superblock_test 00:07:39.837 ************************************ 00:07:39.837 17:29:10 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test raid0 2 read 00:07:39.837 17:29:10 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:07:39.837 17:29:10 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:07:39.837 17:29:10 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:39.837 ************************************ 00:07:39.837 START TEST raid_read_error_test 00:07:39.837 ************************************ 00:07:39.837 17:29:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1125 -- # raid_io_error_test raid0 2 read 00:07:39.837 17:29:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid0 00:07:39.837 17:29:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=2 00:07:39.837 17:29:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:07:39.837 17:29:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:07:39.837 17:29:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:39.837 17:29:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:07:39.837 17:29:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:07:39.837 17:29:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:39.837 17:29:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:07:39.837 17:29:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:07:39.837 17:29:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:39.837 17:29:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:07:39.837 17:29:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:07:39.837 17:29:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:07:39.837 17:29:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:07:39.837 17:29:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:07:39.837 17:29:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:07:39.837 17:29:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:07:39.837 17:29:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid0 '!=' raid1 ']' 00:07:39.837 17:29:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:07:39.837 17:29:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:07:39.837 17:29:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:07:39.837 17:29:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.L40S5SRKbx 00:07:39.837 17:29:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=72579 00:07:39.837 17:29:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:07:39.837 17:29:10 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 72579 00:07:39.837 17:29:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@831 -- # '[' -z 72579 ']' 00:07:39.837 17:29:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:39.837 17:29:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:07:39.837 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:39.837 17:29:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:39.837 17:29:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:07:39.837 17:29:10 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:40.097 [2024-11-27 17:29:11.075201] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:07:40.097 [2024-11-27 17:29:11.075365] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid72579 ] 00:07:40.097 [2024-11-27 17:29:11.220517] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:40.357 [2024-11-27 17:29:11.289849] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:07:40.357 [2024-11-27 17:29:11.366055] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:40.357 [2024-11-27 17:29:11.366098] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:40.935 17:29:11 bdev_raid.raid_read_error_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:07:40.935 17:29:11 bdev_raid.raid_read_error_test -- common/autotest_common.sh@864 -- # return 0 00:07:40.935 17:29:11 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:07:40.935 17:29:11 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:07:40.935 17:29:11 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:40.935 17:29:11 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:40.935 BaseBdev1_malloc 00:07:40.935 17:29:11 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:40.935 17:29:11 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:07:40.935 17:29:11 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:40.935 17:29:11 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:40.935 true 00:07:40.935 17:29:11 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:40.935 17:29:11 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:07:40.935 17:29:11 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:40.935 17:29:11 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:40.935 [2024-11-27 17:29:11.935603] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:07:40.935 [2024-11-27 17:29:11.935694] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:40.935 [2024-11-27 17:29:11.935722] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:07:40.935 [2024-11-27 17:29:11.935732] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:40.935 [2024-11-27 17:29:11.938164] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:40.935 [2024-11-27 17:29:11.938197] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:07:40.935 BaseBdev1 00:07:40.935 17:29:11 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:40.935 17:29:11 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:07:40.935 17:29:11 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:07:40.935 17:29:11 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:40.935 17:29:11 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:40.935 BaseBdev2_malloc 00:07:40.935 17:29:11 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:40.935 17:29:11 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:07:40.935 17:29:11 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:40.935 17:29:11 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:40.935 true 00:07:40.935 17:29:11 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:40.935 17:29:11 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:07:40.935 17:29:11 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:40.935 17:29:11 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:40.935 [2024-11-27 17:29:11.999415] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:07:40.935 [2024-11-27 17:29:11.999480] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:40.935 [2024-11-27 17:29:11.999505] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:07:40.935 [2024-11-27 17:29:11.999516] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:40.935 [2024-11-27 17:29:12.002215] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:40.935 [2024-11-27 17:29:12.002253] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:07:40.935 BaseBdev2 00:07:40.935 17:29:12 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:40.935 17:29:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 -s 00:07:40.935 17:29:12 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:40.935 17:29:12 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:40.935 [2024-11-27 17:29:12.011475] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:40.935 [2024-11-27 17:29:12.013709] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:07:40.935 [2024-11-27 17:29:12.013909] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:07:40.935 [2024-11-27 17:29:12.013949] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:07:40.936 [2024-11-27 17:29:12.014260] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:07:40.936 [2024-11-27 17:29:12.014399] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:07:40.936 [2024-11-27 17:29:12.014424] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:07:40.936 [2024-11-27 17:29:12.014572] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:40.936 17:29:12 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:40.936 17:29:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 2 00:07:40.936 17:29:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:40.936 17:29:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:40.936 17:29:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:40.936 17:29:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:40.936 17:29:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:40.936 17:29:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:40.936 17:29:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:40.936 17:29:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:40.936 17:29:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:40.936 17:29:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:40.936 17:29:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:40.936 17:29:12 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:40.936 17:29:12 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:40.936 17:29:12 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:40.936 17:29:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:40.936 "name": "raid_bdev1", 00:07:40.936 "uuid": "0cfa88a1-23a6-431b-b2a2-9398ce1c732e", 00:07:40.936 "strip_size_kb": 64, 00:07:40.936 "state": "online", 00:07:40.936 "raid_level": "raid0", 00:07:40.936 "superblock": true, 00:07:40.936 "num_base_bdevs": 2, 00:07:40.936 "num_base_bdevs_discovered": 2, 00:07:40.936 "num_base_bdevs_operational": 2, 00:07:40.936 "base_bdevs_list": [ 00:07:40.936 { 00:07:40.936 "name": "BaseBdev1", 00:07:40.936 "uuid": "234b2c7a-a593-5d3d-b51b-936ff99e97bb", 00:07:40.936 "is_configured": true, 00:07:40.936 "data_offset": 2048, 00:07:40.936 "data_size": 63488 00:07:40.936 }, 00:07:40.936 { 00:07:40.936 "name": "BaseBdev2", 00:07:40.936 "uuid": "e5eebdda-4761-5040-9051-dda0a7d72060", 00:07:40.936 "is_configured": true, 00:07:40.936 "data_offset": 2048, 00:07:40.936 "data_size": 63488 00:07:40.936 } 00:07:40.936 ] 00:07:40.936 }' 00:07:40.936 17:29:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:40.936 17:29:12 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:41.507 17:29:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:07:41.507 17:29:12 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:07:41.507 [2024-11-27 17:29:12.511173] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:07:42.446 17:29:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:07:42.446 17:29:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:42.446 17:29:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:42.446 17:29:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:42.446 17:29:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:07:42.446 17:29:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid0 = \r\a\i\d\1 ]] 00:07:42.446 17:29:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=2 00:07:42.446 17:29:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 2 00:07:42.446 17:29:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:42.446 17:29:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:42.446 17:29:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:42.446 17:29:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:42.446 17:29:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:42.446 17:29:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:42.446 17:29:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:42.446 17:29:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:42.446 17:29:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:42.446 17:29:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:42.446 17:29:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:42.446 17:29:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:42.446 17:29:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:42.446 17:29:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:42.447 17:29:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:42.447 "name": "raid_bdev1", 00:07:42.447 "uuid": "0cfa88a1-23a6-431b-b2a2-9398ce1c732e", 00:07:42.447 "strip_size_kb": 64, 00:07:42.447 "state": "online", 00:07:42.447 "raid_level": "raid0", 00:07:42.447 "superblock": true, 00:07:42.447 "num_base_bdevs": 2, 00:07:42.447 "num_base_bdevs_discovered": 2, 00:07:42.447 "num_base_bdevs_operational": 2, 00:07:42.447 "base_bdevs_list": [ 00:07:42.447 { 00:07:42.447 "name": "BaseBdev1", 00:07:42.447 "uuid": "234b2c7a-a593-5d3d-b51b-936ff99e97bb", 00:07:42.447 "is_configured": true, 00:07:42.447 "data_offset": 2048, 00:07:42.447 "data_size": 63488 00:07:42.447 }, 00:07:42.447 { 00:07:42.447 "name": "BaseBdev2", 00:07:42.447 "uuid": "e5eebdda-4761-5040-9051-dda0a7d72060", 00:07:42.447 "is_configured": true, 00:07:42.447 "data_offset": 2048, 00:07:42.447 "data_size": 63488 00:07:42.447 } 00:07:42.447 ] 00:07:42.447 }' 00:07:42.447 17:29:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:42.447 17:29:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:42.706 17:29:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:07:42.706 17:29:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:42.706 17:29:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:42.706 [2024-11-27 17:29:13.851482] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:07:42.706 [2024-11-27 17:29:13.851525] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:42.706 [2024-11-27 17:29:13.853981] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:42.706 [2024-11-27 17:29:13.854027] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:42.706 [2024-11-27 17:29:13.854067] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:42.706 [2024-11-27 17:29:13.854082] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:07:42.706 { 00:07:42.706 "results": [ 00:07:42.706 { 00:07:42.706 "job": "raid_bdev1", 00:07:42.706 "core_mask": "0x1", 00:07:42.706 "workload": "randrw", 00:07:42.706 "percentage": 50, 00:07:42.706 "status": "finished", 00:07:42.706 "queue_depth": 1, 00:07:42.706 "io_size": 131072, 00:07:42.706 "runtime": 1.340769, 00:07:42.706 "iops": 15894.609735159449, 00:07:42.706 "mibps": 1986.8262168949311, 00:07:42.706 "io_failed": 1, 00:07:42.706 "io_timeout": 0, 00:07:42.706 "avg_latency_us": 87.92227074235808, 00:07:42.706 "min_latency_us": 24.593886462882097, 00:07:42.706 "max_latency_us": 1416.6078602620087 00:07:42.706 } 00:07:42.706 ], 00:07:42.706 "core_count": 1 00:07:42.706 } 00:07:42.706 17:29:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:42.706 17:29:13 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 72579 00:07:42.706 17:29:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@950 -- # '[' -z 72579 ']' 00:07:42.706 17:29:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # kill -0 72579 00:07:42.706 17:29:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@955 -- # uname 00:07:42.706 17:29:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:07:42.706 17:29:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 72579 00:07:42.967 17:29:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:07:42.967 17:29:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:07:42.967 killing process with pid 72579 00:07:42.967 17:29:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 72579' 00:07:42.967 17:29:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@969 -- # kill 72579 00:07:42.967 [2024-11-27 17:29:13.901230] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:42.967 17:29:13 bdev_raid.raid_read_error_test -- common/autotest_common.sh@974 -- # wait 72579 00:07:42.967 [2024-11-27 17:29:13.928387] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:43.228 17:29:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.L40S5SRKbx 00:07:43.228 17:29:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:07:43.228 17:29:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:07:43.228 17:29:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.75 00:07:43.228 17:29:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid0 00:07:43.228 17:29:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:07:43.228 17:29:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:07:43.228 17:29:14 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.75 != \0\.\0\0 ]] 00:07:43.228 00:07:43.228 real 0m3.342s 00:07:43.228 user 0m4.055s 00:07:43.228 sys 0m0.614s 00:07:43.228 17:29:14 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:07:43.228 17:29:14 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:43.228 ************************************ 00:07:43.228 END TEST raid_read_error_test 00:07:43.228 ************************************ 00:07:43.228 17:29:14 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test raid0 2 write 00:07:43.228 17:29:14 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:07:43.228 17:29:14 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:07:43.228 17:29:14 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:43.228 ************************************ 00:07:43.228 START TEST raid_write_error_test 00:07:43.228 ************************************ 00:07:43.228 17:29:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1125 -- # raid_io_error_test raid0 2 write 00:07:43.228 17:29:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid0 00:07:43.228 17:29:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=2 00:07:43.228 17:29:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:07:43.228 17:29:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:07:43.228 17:29:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:43.228 17:29:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:07:43.228 17:29:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:07:43.228 17:29:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:43.228 17:29:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:07:43.228 17:29:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:07:43.229 17:29:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:43.229 17:29:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:07:43.229 17:29:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:07:43.229 17:29:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:07:43.229 17:29:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:07:43.229 17:29:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:07:43.229 17:29:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:07:43.229 17:29:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:07:43.229 17:29:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid0 '!=' raid1 ']' 00:07:43.229 17:29:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:07:43.229 17:29:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:07:43.229 17:29:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:07:43.229 17:29:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.O54qctQcDO 00:07:43.229 17:29:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=72714 00:07:43.229 17:29:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 72714 00:07:43.229 17:29:14 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:07:43.229 17:29:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@831 -- # '[' -z 72714 ']' 00:07:43.229 17:29:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:43.229 17:29:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:07:43.229 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:43.229 17:29:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:43.229 17:29:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:07:43.229 17:29:14 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:43.489 [2024-11-27 17:29:14.487544] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:07:43.489 [2024-11-27 17:29:14.487676] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid72714 ] 00:07:43.489 [2024-11-27 17:29:14.633548] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:43.749 [2024-11-27 17:29:14.700982] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:07:43.749 [2024-11-27 17:29:14.776062] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:43.749 [2024-11-27 17:29:14.776103] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:44.320 17:29:15 bdev_raid.raid_write_error_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:07:44.320 17:29:15 bdev_raid.raid_write_error_test -- common/autotest_common.sh@864 -- # return 0 00:07:44.320 17:29:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:07:44.320 17:29:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:07:44.320 17:29:15 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:44.320 17:29:15 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:44.320 BaseBdev1_malloc 00:07:44.320 17:29:15 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:44.320 17:29:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:07:44.320 17:29:15 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:44.320 17:29:15 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:44.320 true 00:07:44.320 17:29:15 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:44.320 17:29:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:07:44.320 17:29:15 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:44.320 17:29:15 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:44.320 [2024-11-27 17:29:15.345816] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:07:44.320 [2024-11-27 17:29:15.345881] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:44.320 [2024-11-27 17:29:15.345904] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:07:44.320 [2024-11-27 17:29:15.345913] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:44.320 [2024-11-27 17:29:15.348401] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:44.320 [2024-11-27 17:29:15.348457] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:07:44.320 BaseBdev1 00:07:44.320 17:29:15 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:44.320 17:29:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:07:44.320 17:29:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:07:44.320 17:29:15 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:44.320 17:29:15 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:44.320 BaseBdev2_malloc 00:07:44.320 17:29:15 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:44.320 17:29:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:07:44.320 17:29:15 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:44.320 17:29:15 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:44.320 true 00:07:44.320 17:29:15 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:44.320 17:29:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:07:44.320 17:29:15 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:44.320 17:29:15 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:44.320 [2024-11-27 17:29:15.410926] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:07:44.320 [2024-11-27 17:29:15.411013] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:44.320 [2024-11-27 17:29:15.411048] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:07:44.320 [2024-11-27 17:29:15.411063] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:44.320 [2024-11-27 17:29:15.414695] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:44.320 [2024-11-27 17:29:15.414756] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:07:44.320 BaseBdev2 00:07:44.320 17:29:15 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:44.320 17:29:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 -s 00:07:44.320 17:29:15 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:44.320 17:29:15 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:44.320 [2024-11-27 17:29:15.422998] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:44.320 [2024-11-27 17:29:15.425444] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:07:44.320 [2024-11-27 17:29:15.425657] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:07:44.320 [2024-11-27 17:29:15.425672] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:07:44.320 [2024-11-27 17:29:15.425979] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:07:44.320 [2024-11-27 17:29:15.426122] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:07:44.320 [2024-11-27 17:29:15.426154] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:07:44.320 [2024-11-27 17:29:15.426290] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:44.320 17:29:15 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:44.320 17:29:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 2 00:07:44.320 17:29:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:44.320 17:29:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:44.320 17:29:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:44.320 17:29:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:44.320 17:29:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:44.320 17:29:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:44.320 17:29:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:44.320 17:29:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:44.320 17:29:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:44.320 17:29:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:44.320 17:29:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:44.320 17:29:15 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:44.320 17:29:15 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:44.320 17:29:15 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:44.320 17:29:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:44.321 "name": "raid_bdev1", 00:07:44.321 "uuid": "e6ab30c0-7567-4d4f-afe9-f535296dc283", 00:07:44.321 "strip_size_kb": 64, 00:07:44.321 "state": "online", 00:07:44.321 "raid_level": "raid0", 00:07:44.321 "superblock": true, 00:07:44.321 "num_base_bdevs": 2, 00:07:44.321 "num_base_bdevs_discovered": 2, 00:07:44.321 "num_base_bdevs_operational": 2, 00:07:44.321 "base_bdevs_list": [ 00:07:44.321 { 00:07:44.321 "name": "BaseBdev1", 00:07:44.321 "uuid": "fd320b6d-9161-541f-bd39-2efaee85dc0f", 00:07:44.321 "is_configured": true, 00:07:44.321 "data_offset": 2048, 00:07:44.321 "data_size": 63488 00:07:44.321 }, 00:07:44.321 { 00:07:44.321 "name": "BaseBdev2", 00:07:44.321 "uuid": "b508a0a2-88ed-5bec-ab73-88d3a27eb128", 00:07:44.321 "is_configured": true, 00:07:44.321 "data_offset": 2048, 00:07:44.321 "data_size": 63488 00:07:44.321 } 00:07:44.321 ] 00:07:44.321 }' 00:07:44.321 17:29:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:44.321 17:29:15 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:44.891 17:29:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:07:44.891 17:29:15 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:07:44.891 [2024-11-27 17:29:15.958717] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:07:45.833 17:29:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:07:45.833 17:29:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:45.833 17:29:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:45.833 17:29:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:45.833 17:29:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:07:45.833 17:29:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid0 = \r\a\i\d\1 ]] 00:07:45.833 17:29:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=2 00:07:45.833 17:29:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 2 00:07:45.833 17:29:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:45.833 17:29:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:45.833 17:29:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:07:45.833 17:29:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:45.834 17:29:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:45.834 17:29:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:45.834 17:29:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:45.834 17:29:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:45.834 17:29:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:45.834 17:29:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:45.834 17:29:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:45.834 17:29:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:45.834 17:29:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:45.834 17:29:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:45.834 17:29:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:45.834 "name": "raid_bdev1", 00:07:45.834 "uuid": "e6ab30c0-7567-4d4f-afe9-f535296dc283", 00:07:45.834 "strip_size_kb": 64, 00:07:45.834 "state": "online", 00:07:45.834 "raid_level": "raid0", 00:07:45.834 "superblock": true, 00:07:45.834 "num_base_bdevs": 2, 00:07:45.834 "num_base_bdevs_discovered": 2, 00:07:45.834 "num_base_bdevs_operational": 2, 00:07:45.834 "base_bdevs_list": [ 00:07:45.834 { 00:07:45.834 "name": "BaseBdev1", 00:07:45.834 "uuid": "fd320b6d-9161-541f-bd39-2efaee85dc0f", 00:07:45.834 "is_configured": true, 00:07:45.834 "data_offset": 2048, 00:07:45.834 "data_size": 63488 00:07:45.834 }, 00:07:45.834 { 00:07:45.834 "name": "BaseBdev2", 00:07:45.834 "uuid": "b508a0a2-88ed-5bec-ab73-88d3a27eb128", 00:07:45.834 "is_configured": true, 00:07:45.834 "data_offset": 2048, 00:07:45.834 "data_size": 63488 00:07:45.834 } 00:07:45.834 ] 00:07:45.834 }' 00:07:45.834 17:29:16 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:45.834 17:29:16 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:46.405 17:29:17 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:07:46.405 17:29:17 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:46.405 17:29:17 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:46.405 [2024-11-27 17:29:17.310939] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:07:46.405 [2024-11-27 17:29:17.310985] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:46.405 [2024-11-27 17:29:17.313461] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:46.405 [2024-11-27 17:29:17.313536] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:46.405 [2024-11-27 17:29:17.313574] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:46.405 [2024-11-27 17:29:17.313584] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:07:46.405 { 00:07:46.405 "results": [ 00:07:46.405 { 00:07:46.405 "job": "raid_bdev1", 00:07:46.405 "core_mask": "0x1", 00:07:46.405 "workload": "randrw", 00:07:46.405 "percentage": 50, 00:07:46.405 "status": "finished", 00:07:46.405 "queue_depth": 1, 00:07:46.405 "io_size": 131072, 00:07:46.405 "runtime": 1.352729, 00:07:46.405 "iops": 15887.882938859151, 00:07:46.405 "mibps": 1985.985367357394, 00:07:46.405 "io_failed": 1, 00:07:46.405 "io_timeout": 0, 00:07:46.405 "avg_latency_us": 88.14203076577995, 00:07:46.405 "min_latency_us": 24.482096069868994, 00:07:46.405 "max_latency_us": 1731.4096069868995 00:07:46.405 } 00:07:46.405 ], 00:07:46.405 "core_count": 1 00:07:46.405 } 00:07:46.405 17:29:17 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:46.405 17:29:17 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 72714 00:07:46.405 17:29:17 bdev_raid.raid_write_error_test -- common/autotest_common.sh@950 -- # '[' -z 72714 ']' 00:07:46.405 17:29:17 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # kill -0 72714 00:07:46.405 17:29:17 bdev_raid.raid_write_error_test -- common/autotest_common.sh@955 -- # uname 00:07:46.405 17:29:17 bdev_raid.raid_write_error_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:07:46.405 17:29:17 bdev_raid.raid_write_error_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 72714 00:07:46.405 17:29:17 bdev_raid.raid_write_error_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:07:46.405 17:29:17 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:07:46.405 killing process with pid 72714 00:07:46.405 17:29:17 bdev_raid.raid_write_error_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 72714' 00:07:46.405 17:29:17 bdev_raid.raid_write_error_test -- common/autotest_common.sh@969 -- # kill 72714 00:07:46.405 [2024-11-27 17:29:17.362323] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:46.405 17:29:17 bdev_raid.raid_write_error_test -- common/autotest_common.sh@974 -- # wait 72714 00:07:46.405 [2024-11-27 17:29:17.391589] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:46.665 17:29:17 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:07:46.665 17:29:17 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.O54qctQcDO 00:07:46.665 17:29:17 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:07:46.665 17:29:17 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.74 00:07:46.665 17:29:17 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid0 00:07:46.665 17:29:17 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:07:46.665 17:29:17 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:07:46.665 17:29:17 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.74 != \0\.\0\0 ]] 00:07:46.665 00:07:46.665 real 0m3.384s 00:07:46.665 user 0m4.141s 00:07:46.665 sys 0m0.589s 00:07:46.665 17:29:17 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:07:46.665 17:29:17 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:46.665 ************************************ 00:07:46.665 END TEST raid_write_error_test 00:07:46.665 ************************************ 00:07:46.665 17:29:17 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:07:46.665 17:29:17 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test concat 2 false 00:07:46.665 17:29:17 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:07:46.666 17:29:17 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:07:46.666 17:29:17 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:46.666 ************************************ 00:07:46.666 START TEST raid_state_function_test 00:07:46.666 ************************************ 00:07:46.666 17:29:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1125 -- # raid_state_function_test concat 2 false 00:07:46.666 17:29:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=concat 00:07:46.666 17:29:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:07:46.666 17:29:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:07:46.666 17:29:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:07:46.666 17:29:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:07:46.666 17:29:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:46.666 17:29:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:07:46.666 17:29:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:07:46.666 17:29:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:46.666 17:29:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:07:46.666 17:29:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:07:46.666 17:29:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:46.666 17:29:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:07:46.666 17:29:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:07:46.666 17:29:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:07:46.666 17:29:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:07:46.666 17:29:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:07:46.666 17:29:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:07:46.666 17:29:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' concat '!=' raid1 ']' 00:07:46.666 17:29:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:07:46.666 17:29:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:07:46.666 17:29:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:07:46.666 17:29:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:07:46.666 17:29:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=72846 00:07:46.666 Process raid pid: 72846 00:07:46.666 17:29:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 72846' 00:07:46.666 17:29:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:07:46.666 17:29:17 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 72846 00:07:46.666 17:29:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@831 -- # '[' -z 72846 ']' 00:07:46.666 17:29:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:46.666 17:29:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:07:46.666 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:46.666 17:29:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:46.666 17:29:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:07:46.666 17:29:17 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:46.926 [2024-11-27 17:29:17.932694] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:07:46.926 [2024-11-27 17:29:17.932853] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:07:46.926 [2024-11-27 17:29:18.081130] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:47.186 [2024-11-27 17:29:18.149635] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:07:47.186 [2024-11-27 17:29:18.224570] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:47.186 [2024-11-27 17:29:18.224620] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:47.755 17:29:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:07:47.755 17:29:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@864 -- # return 0 00:07:47.756 17:29:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:47.756 17:29:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:47.756 17:29:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:47.756 [2024-11-27 17:29:18.745769] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:07:47.756 [2024-11-27 17:29:18.745827] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:07:47.756 [2024-11-27 17:29:18.745864] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:47.756 [2024-11-27 17:29:18.745875] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:47.756 17:29:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:47.756 17:29:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 2 00:07:47.756 17:29:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:47.756 17:29:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:47.756 17:29:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:47.756 17:29:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:47.756 17:29:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:47.756 17:29:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:47.756 17:29:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:47.756 17:29:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:47.756 17:29:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:47.756 17:29:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:47.756 17:29:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:47.756 17:29:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:47.756 17:29:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:47.756 17:29:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:47.756 17:29:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:47.756 "name": "Existed_Raid", 00:07:47.756 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:47.756 "strip_size_kb": 64, 00:07:47.756 "state": "configuring", 00:07:47.756 "raid_level": "concat", 00:07:47.756 "superblock": false, 00:07:47.756 "num_base_bdevs": 2, 00:07:47.756 "num_base_bdevs_discovered": 0, 00:07:47.756 "num_base_bdevs_operational": 2, 00:07:47.756 "base_bdevs_list": [ 00:07:47.756 { 00:07:47.756 "name": "BaseBdev1", 00:07:47.756 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:47.756 "is_configured": false, 00:07:47.756 "data_offset": 0, 00:07:47.756 "data_size": 0 00:07:47.756 }, 00:07:47.756 { 00:07:47.756 "name": "BaseBdev2", 00:07:47.756 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:47.756 "is_configured": false, 00:07:47.756 "data_offset": 0, 00:07:47.756 "data_size": 0 00:07:47.756 } 00:07:47.756 ] 00:07:47.756 }' 00:07:47.756 17:29:18 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:47.756 17:29:18 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:48.016 17:29:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:07:48.016 17:29:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:48.016 17:29:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:48.016 [2024-11-27 17:29:19.176897] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:07:48.016 [2024-11-27 17:29:19.176987] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:07:48.016 17:29:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:48.016 17:29:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:48.016 17:29:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:48.016 17:29:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:48.016 [2024-11-27 17:29:19.184898] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:07:48.016 [2024-11-27 17:29:19.184935] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:07:48.016 [2024-11-27 17:29:19.184968] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:48.016 [2024-11-27 17:29:19.184979] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:48.016 17:29:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:48.016 17:29:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:07:48.016 17:29:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:48.016 17:29:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:48.276 [2024-11-27 17:29:19.207950] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:48.276 BaseBdev1 00:07:48.276 17:29:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:48.276 17:29:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:07:48.276 17:29:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:07:48.276 17:29:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:07:48.276 17:29:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:07:48.276 17:29:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:07:48.276 17:29:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:07:48.276 17:29:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:07:48.276 17:29:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:48.276 17:29:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:48.276 17:29:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:48.276 17:29:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:07:48.276 17:29:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:48.276 17:29:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:48.276 [ 00:07:48.276 { 00:07:48.276 "name": "BaseBdev1", 00:07:48.276 "aliases": [ 00:07:48.276 "7aa31275-ccfa-4e9a-84d3-36d74241ef8f" 00:07:48.276 ], 00:07:48.276 "product_name": "Malloc disk", 00:07:48.276 "block_size": 512, 00:07:48.276 "num_blocks": 65536, 00:07:48.276 "uuid": "7aa31275-ccfa-4e9a-84d3-36d74241ef8f", 00:07:48.276 "assigned_rate_limits": { 00:07:48.276 "rw_ios_per_sec": 0, 00:07:48.276 "rw_mbytes_per_sec": 0, 00:07:48.276 "r_mbytes_per_sec": 0, 00:07:48.276 "w_mbytes_per_sec": 0 00:07:48.276 }, 00:07:48.276 "claimed": true, 00:07:48.276 "claim_type": "exclusive_write", 00:07:48.276 "zoned": false, 00:07:48.276 "supported_io_types": { 00:07:48.276 "read": true, 00:07:48.276 "write": true, 00:07:48.276 "unmap": true, 00:07:48.276 "flush": true, 00:07:48.276 "reset": true, 00:07:48.276 "nvme_admin": false, 00:07:48.276 "nvme_io": false, 00:07:48.276 "nvme_io_md": false, 00:07:48.276 "write_zeroes": true, 00:07:48.276 "zcopy": true, 00:07:48.276 "get_zone_info": false, 00:07:48.276 "zone_management": false, 00:07:48.276 "zone_append": false, 00:07:48.276 "compare": false, 00:07:48.276 "compare_and_write": false, 00:07:48.276 "abort": true, 00:07:48.276 "seek_hole": false, 00:07:48.276 "seek_data": false, 00:07:48.276 "copy": true, 00:07:48.276 "nvme_iov_md": false 00:07:48.276 }, 00:07:48.276 "memory_domains": [ 00:07:48.276 { 00:07:48.276 "dma_device_id": "system", 00:07:48.276 "dma_device_type": 1 00:07:48.276 }, 00:07:48.276 { 00:07:48.276 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:48.276 "dma_device_type": 2 00:07:48.276 } 00:07:48.276 ], 00:07:48.276 "driver_specific": {} 00:07:48.276 } 00:07:48.276 ] 00:07:48.276 17:29:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:48.276 17:29:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:07:48.276 17:29:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 2 00:07:48.276 17:29:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:48.276 17:29:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:48.276 17:29:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:48.276 17:29:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:48.276 17:29:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:48.276 17:29:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:48.276 17:29:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:48.276 17:29:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:48.276 17:29:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:48.276 17:29:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:48.276 17:29:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:48.276 17:29:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:48.276 17:29:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:48.276 17:29:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:48.276 17:29:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:48.276 "name": "Existed_Raid", 00:07:48.276 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:48.276 "strip_size_kb": 64, 00:07:48.276 "state": "configuring", 00:07:48.276 "raid_level": "concat", 00:07:48.276 "superblock": false, 00:07:48.276 "num_base_bdevs": 2, 00:07:48.276 "num_base_bdevs_discovered": 1, 00:07:48.276 "num_base_bdevs_operational": 2, 00:07:48.276 "base_bdevs_list": [ 00:07:48.276 { 00:07:48.276 "name": "BaseBdev1", 00:07:48.276 "uuid": "7aa31275-ccfa-4e9a-84d3-36d74241ef8f", 00:07:48.276 "is_configured": true, 00:07:48.276 "data_offset": 0, 00:07:48.276 "data_size": 65536 00:07:48.276 }, 00:07:48.276 { 00:07:48.276 "name": "BaseBdev2", 00:07:48.276 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:48.276 "is_configured": false, 00:07:48.276 "data_offset": 0, 00:07:48.276 "data_size": 0 00:07:48.276 } 00:07:48.276 ] 00:07:48.276 }' 00:07:48.276 17:29:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:48.276 17:29:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:48.536 17:29:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:07:48.536 17:29:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:48.536 17:29:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:48.536 [2024-11-27 17:29:19.659195] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:07:48.536 [2024-11-27 17:29:19.659239] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:07:48.536 17:29:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:48.536 17:29:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:48.536 17:29:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:48.536 17:29:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:48.536 [2024-11-27 17:29:19.671223] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:48.536 [2024-11-27 17:29:19.673311] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:48.536 [2024-11-27 17:29:19.673350] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:48.536 17:29:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:48.536 17:29:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:07:48.536 17:29:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:07:48.536 17:29:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 2 00:07:48.536 17:29:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:48.536 17:29:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:48.536 17:29:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:48.536 17:29:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:48.536 17:29:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:48.536 17:29:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:48.536 17:29:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:48.536 17:29:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:48.536 17:29:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:48.536 17:29:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:48.536 17:29:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:48.536 17:29:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:48.536 17:29:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:48.536 17:29:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:48.795 17:29:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:48.795 "name": "Existed_Raid", 00:07:48.795 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:48.795 "strip_size_kb": 64, 00:07:48.795 "state": "configuring", 00:07:48.795 "raid_level": "concat", 00:07:48.795 "superblock": false, 00:07:48.795 "num_base_bdevs": 2, 00:07:48.795 "num_base_bdevs_discovered": 1, 00:07:48.795 "num_base_bdevs_operational": 2, 00:07:48.795 "base_bdevs_list": [ 00:07:48.795 { 00:07:48.795 "name": "BaseBdev1", 00:07:48.795 "uuid": "7aa31275-ccfa-4e9a-84d3-36d74241ef8f", 00:07:48.795 "is_configured": true, 00:07:48.795 "data_offset": 0, 00:07:48.795 "data_size": 65536 00:07:48.795 }, 00:07:48.795 { 00:07:48.795 "name": "BaseBdev2", 00:07:48.795 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:48.795 "is_configured": false, 00:07:48.795 "data_offset": 0, 00:07:48.795 "data_size": 0 00:07:48.795 } 00:07:48.795 ] 00:07:48.795 }' 00:07:48.795 17:29:19 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:48.795 17:29:19 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:49.056 17:29:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:07:49.056 17:29:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:49.056 17:29:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:49.056 [2024-11-27 17:29:20.083774] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:07:49.056 [2024-11-27 17:29:20.084068] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:07:49.056 [2024-11-27 17:29:20.084224] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 131072, blocklen 512 00:07:49.056 [2024-11-27 17:29:20.085371] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:07:49.056 [2024-11-27 17:29:20.085963] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:07:49.056 [2024-11-27 17:29:20.086179] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:07:49.056 [2024-11-27 17:29:20.087015] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:49.056 BaseBdev2 00:07:49.056 17:29:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:49.056 17:29:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:07:49.056 17:29:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:07:49.056 17:29:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:07:49.056 17:29:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:07:49.056 17:29:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:07:49.056 17:29:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:07:49.056 17:29:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:07:49.056 17:29:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:49.056 17:29:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:49.056 17:29:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:49.056 17:29:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:07:49.056 17:29:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:49.056 17:29:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:49.056 [ 00:07:49.056 { 00:07:49.056 "name": "BaseBdev2", 00:07:49.056 "aliases": [ 00:07:49.056 "bd49f4f2-c4c9-4d66-b6d3-c4e0e2313ce6" 00:07:49.056 ], 00:07:49.056 "product_name": "Malloc disk", 00:07:49.056 "block_size": 512, 00:07:49.056 "num_blocks": 65536, 00:07:49.056 "uuid": "bd49f4f2-c4c9-4d66-b6d3-c4e0e2313ce6", 00:07:49.056 "assigned_rate_limits": { 00:07:49.056 "rw_ios_per_sec": 0, 00:07:49.056 "rw_mbytes_per_sec": 0, 00:07:49.056 "r_mbytes_per_sec": 0, 00:07:49.056 "w_mbytes_per_sec": 0 00:07:49.056 }, 00:07:49.056 "claimed": true, 00:07:49.056 "claim_type": "exclusive_write", 00:07:49.056 "zoned": false, 00:07:49.056 "supported_io_types": { 00:07:49.056 "read": true, 00:07:49.056 "write": true, 00:07:49.056 "unmap": true, 00:07:49.056 "flush": true, 00:07:49.056 "reset": true, 00:07:49.056 "nvme_admin": false, 00:07:49.056 "nvme_io": false, 00:07:49.056 "nvme_io_md": false, 00:07:49.056 "write_zeroes": true, 00:07:49.056 "zcopy": true, 00:07:49.056 "get_zone_info": false, 00:07:49.056 "zone_management": false, 00:07:49.056 "zone_append": false, 00:07:49.056 "compare": false, 00:07:49.056 "compare_and_write": false, 00:07:49.056 "abort": true, 00:07:49.056 "seek_hole": false, 00:07:49.056 "seek_data": false, 00:07:49.056 "copy": true, 00:07:49.056 "nvme_iov_md": false 00:07:49.056 }, 00:07:49.056 "memory_domains": [ 00:07:49.056 { 00:07:49.056 "dma_device_id": "system", 00:07:49.056 "dma_device_type": 1 00:07:49.056 }, 00:07:49.056 { 00:07:49.056 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:49.056 "dma_device_type": 2 00:07:49.056 } 00:07:49.056 ], 00:07:49.056 "driver_specific": {} 00:07:49.056 } 00:07:49.056 ] 00:07:49.056 17:29:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:49.056 17:29:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:07:49.056 17:29:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:07:49.056 17:29:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:07:49.056 17:29:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online concat 64 2 00:07:49.056 17:29:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:49.056 17:29:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:49.056 17:29:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:49.056 17:29:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:49.056 17:29:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:49.056 17:29:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:49.056 17:29:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:49.056 17:29:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:49.056 17:29:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:49.056 17:29:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:49.056 17:29:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:49.056 17:29:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:49.056 17:29:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:49.056 17:29:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:49.056 17:29:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:49.056 "name": "Existed_Raid", 00:07:49.056 "uuid": "5bc8a332-1334-4375-9933-3aeeb10fa8be", 00:07:49.056 "strip_size_kb": 64, 00:07:49.056 "state": "online", 00:07:49.056 "raid_level": "concat", 00:07:49.056 "superblock": false, 00:07:49.056 "num_base_bdevs": 2, 00:07:49.056 "num_base_bdevs_discovered": 2, 00:07:49.056 "num_base_bdevs_operational": 2, 00:07:49.056 "base_bdevs_list": [ 00:07:49.056 { 00:07:49.056 "name": "BaseBdev1", 00:07:49.056 "uuid": "7aa31275-ccfa-4e9a-84d3-36d74241ef8f", 00:07:49.056 "is_configured": true, 00:07:49.056 "data_offset": 0, 00:07:49.056 "data_size": 65536 00:07:49.056 }, 00:07:49.056 { 00:07:49.056 "name": "BaseBdev2", 00:07:49.056 "uuid": "bd49f4f2-c4c9-4d66-b6d3-c4e0e2313ce6", 00:07:49.056 "is_configured": true, 00:07:49.056 "data_offset": 0, 00:07:49.056 "data_size": 65536 00:07:49.056 } 00:07:49.056 ] 00:07:49.056 }' 00:07:49.056 17:29:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:49.056 17:29:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:49.627 17:29:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:07:49.627 17:29:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:07:49.627 17:29:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:07:49.627 17:29:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:07:49.627 17:29:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:07:49.628 17:29:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:07:49.628 17:29:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:07:49.628 17:29:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:07:49.628 17:29:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:49.628 17:29:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:49.628 [2024-11-27 17:29:20.587135] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:49.628 17:29:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:49.628 17:29:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:07:49.628 "name": "Existed_Raid", 00:07:49.628 "aliases": [ 00:07:49.628 "5bc8a332-1334-4375-9933-3aeeb10fa8be" 00:07:49.628 ], 00:07:49.628 "product_name": "Raid Volume", 00:07:49.628 "block_size": 512, 00:07:49.628 "num_blocks": 131072, 00:07:49.628 "uuid": "5bc8a332-1334-4375-9933-3aeeb10fa8be", 00:07:49.628 "assigned_rate_limits": { 00:07:49.628 "rw_ios_per_sec": 0, 00:07:49.628 "rw_mbytes_per_sec": 0, 00:07:49.628 "r_mbytes_per_sec": 0, 00:07:49.628 "w_mbytes_per_sec": 0 00:07:49.628 }, 00:07:49.628 "claimed": false, 00:07:49.628 "zoned": false, 00:07:49.628 "supported_io_types": { 00:07:49.628 "read": true, 00:07:49.628 "write": true, 00:07:49.628 "unmap": true, 00:07:49.628 "flush": true, 00:07:49.628 "reset": true, 00:07:49.628 "nvme_admin": false, 00:07:49.628 "nvme_io": false, 00:07:49.628 "nvme_io_md": false, 00:07:49.628 "write_zeroes": true, 00:07:49.628 "zcopy": false, 00:07:49.628 "get_zone_info": false, 00:07:49.628 "zone_management": false, 00:07:49.628 "zone_append": false, 00:07:49.628 "compare": false, 00:07:49.628 "compare_and_write": false, 00:07:49.628 "abort": false, 00:07:49.628 "seek_hole": false, 00:07:49.628 "seek_data": false, 00:07:49.628 "copy": false, 00:07:49.628 "nvme_iov_md": false 00:07:49.628 }, 00:07:49.628 "memory_domains": [ 00:07:49.628 { 00:07:49.628 "dma_device_id": "system", 00:07:49.628 "dma_device_type": 1 00:07:49.628 }, 00:07:49.628 { 00:07:49.628 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:49.628 "dma_device_type": 2 00:07:49.628 }, 00:07:49.628 { 00:07:49.628 "dma_device_id": "system", 00:07:49.628 "dma_device_type": 1 00:07:49.628 }, 00:07:49.628 { 00:07:49.628 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:49.628 "dma_device_type": 2 00:07:49.628 } 00:07:49.628 ], 00:07:49.628 "driver_specific": { 00:07:49.628 "raid": { 00:07:49.628 "uuid": "5bc8a332-1334-4375-9933-3aeeb10fa8be", 00:07:49.628 "strip_size_kb": 64, 00:07:49.628 "state": "online", 00:07:49.628 "raid_level": "concat", 00:07:49.628 "superblock": false, 00:07:49.628 "num_base_bdevs": 2, 00:07:49.628 "num_base_bdevs_discovered": 2, 00:07:49.628 "num_base_bdevs_operational": 2, 00:07:49.628 "base_bdevs_list": [ 00:07:49.628 { 00:07:49.628 "name": "BaseBdev1", 00:07:49.628 "uuid": "7aa31275-ccfa-4e9a-84d3-36d74241ef8f", 00:07:49.628 "is_configured": true, 00:07:49.628 "data_offset": 0, 00:07:49.628 "data_size": 65536 00:07:49.628 }, 00:07:49.628 { 00:07:49.628 "name": "BaseBdev2", 00:07:49.628 "uuid": "bd49f4f2-c4c9-4d66-b6d3-c4e0e2313ce6", 00:07:49.628 "is_configured": true, 00:07:49.628 "data_offset": 0, 00:07:49.628 "data_size": 65536 00:07:49.628 } 00:07:49.628 ] 00:07:49.628 } 00:07:49.628 } 00:07:49.628 }' 00:07:49.628 17:29:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:07:49.628 17:29:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:07:49.628 BaseBdev2' 00:07:49.628 17:29:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:49.628 17:29:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:07:49.628 17:29:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:49.628 17:29:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:49.628 17:29:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:07:49.628 17:29:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:49.628 17:29:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:49.628 17:29:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:49.628 17:29:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:49.628 17:29:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:49.628 17:29:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:49.628 17:29:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:07:49.628 17:29:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:49.628 17:29:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:49.628 17:29:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:49.628 17:29:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:49.628 17:29:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:49.628 17:29:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:49.628 17:29:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:07:49.628 17:29:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:49.628 17:29:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:49.628 [2024-11-27 17:29:20.806556] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:07:49.628 [2024-11-27 17:29:20.806585] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:49.628 [2024-11-27 17:29:20.806643] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:49.888 17:29:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:49.888 17:29:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:07:49.888 17:29:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy concat 00:07:49.888 17:29:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:07:49.888 17:29:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # return 1 00:07:49.888 17:29:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:07:49.888 17:29:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline concat 64 1 00:07:49.888 17:29:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:49.888 17:29:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:07:49.888 17:29:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:49.888 17:29:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:49.888 17:29:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:07:49.888 17:29:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:49.888 17:29:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:49.888 17:29:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:49.888 17:29:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:49.888 17:29:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:49.888 17:29:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:49.888 17:29:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:49.888 17:29:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:49.888 17:29:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:49.888 17:29:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:49.888 "name": "Existed_Raid", 00:07:49.888 "uuid": "5bc8a332-1334-4375-9933-3aeeb10fa8be", 00:07:49.888 "strip_size_kb": 64, 00:07:49.888 "state": "offline", 00:07:49.888 "raid_level": "concat", 00:07:49.888 "superblock": false, 00:07:49.888 "num_base_bdevs": 2, 00:07:49.888 "num_base_bdevs_discovered": 1, 00:07:49.888 "num_base_bdevs_operational": 1, 00:07:49.888 "base_bdevs_list": [ 00:07:49.888 { 00:07:49.888 "name": null, 00:07:49.888 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:49.888 "is_configured": false, 00:07:49.888 "data_offset": 0, 00:07:49.889 "data_size": 65536 00:07:49.889 }, 00:07:49.889 { 00:07:49.889 "name": "BaseBdev2", 00:07:49.889 "uuid": "bd49f4f2-c4c9-4d66-b6d3-c4e0e2313ce6", 00:07:49.889 "is_configured": true, 00:07:49.889 "data_offset": 0, 00:07:49.889 "data_size": 65536 00:07:49.889 } 00:07:49.889 ] 00:07:49.889 }' 00:07:49.889 17:29:20 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:49.889 17:29:20 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:50.149 17:29:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:07:50.149 17:29:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:07:50.149 17:29:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:50.149 17:29:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:50.149 17:29:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:50.149 17:29:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:07:50.149 17:29:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:50.149 17:29:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:07:50.149 17:29:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:07:50.149 17:29:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:07:50.149 17:29:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:50.149 17:29:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:50.149 [2024-11-27 17:29:21.314195] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:07:50.149 [2024-11-27 17:29:21.314248] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:07:50.149 17:29:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:50.149 17:29:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:07:50.149 17:29:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:07:50.409 17:29:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:50.409 17:29:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:07:50.409 17:29:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:50.409 17:29:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:50.409 17:29:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:50.409 17:29:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:07:50.409 17:29:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:07:50.409 17:29:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:07:50.409 17:29:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 72846 00:07:50.409 17:29:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@950 -- # '[' -z 72846 ']' 00:07:50.409 17:29:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # kill -0 72846 00:07:50.409 17:29:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@955 -- # uname 00:07:50.409 17:29:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:07:50.409 17:29:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 72846 00:07:50.409 killing process with pid 72846 00:07:50.409 17:29:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:07:50.409 17:29:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:07:50.409 17:29:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 72846' 00:07:50.409 17:29:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@969 -- # kill 72846 00:07:50.409 [2024-11-27 17:29:21.419622] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:50.409 17:29:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@974 -- # wait 72846 00:07:50.409 [2024-11-27 17:29:21.421204] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:50.669 17:29:21 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:07:50.669 00:07:50.669 real 0m3.950s 00:07:50.669 user 0m6.037s 00:07:50.669 sys 0m0.827s 00:07:50.669 17:29:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:07:50.669 17:29:21 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:07:50.669 ************************************ 00:07:50.669 END TEST raid_state_function_test 00:07:50.669 ************************************ 00:07:50.669 17:29:21 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test concat 2 true 00:07:50.669 17:29:21 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:07:50.669 17:29:21 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:07:50.669 17:29:21 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:50.930 ************************************ 00:07:50.930 START TEST raid_state_function_test_sb 00:07:50.930 ************************************ 00:07:50.930 17:29:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1125 -- # raid_state_function_test concat 2 true 00:07:50.930 17:29:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=concat 00:07:50.930 17:29:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:07:50.930 17:29:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:07:50.930 17:29:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:07:50.930 17:29:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:07:50.930 17:29:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:50.930 17:29:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:07:50.930 17:29:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:07:50.930 17:29:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:50.930 17:29:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:07:50.930 17:29:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:07:50.930 17:29:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:07:50.930 17:29:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:07:50.930 17:29:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:07:50.930 17:29:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:07:50.930 17:29:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:07:50.930 17:29:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:07:50.930 17:29:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:07:50.930 17:29:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' concat '!=' raid1 ']' 00:07:50.930 17:29:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:07:50.930 17:29:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:07:50.930 17:29:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:07:50.930 17:29:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:07:50.930 17:29:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=73083 00:07:50.930 17:29:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:07:50.930 17:29:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 73083' 00:07:50.930 Process raid pid: 73083 00:07:50.930 17:29:21 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 73083 00:07:50.930 17:29:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@831 -- # '[' -z 73083 ']' 00:07:50.930 17:29:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:50.930 17:29:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@836 -- # local max_retries=100 00:07:50.930 17:29:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:50.930 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:50.930 17:29:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # xtrace_disable 00:07:50.930 17:29:21 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:50.930 [2024-11-27 17:29:21.959730] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:07:50.930 [2024-11-27 17:29:21.959922] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:07:50.930 [2024-11-27 17:29:22.106447] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:51.190 [2024-11-27 17:29:22.174090] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:07:51.190 [2024-11-27 17:29:22.249399] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:51.190 [2024-11-27 17:29:22.249437] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:51.759 17:29:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:07:51.760 17:29:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@864 -- # return 0 00:07:51.760 17:29:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:51.760 17:29:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:51.760 17:29:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:51.760 [2024-11-27 17:29:22.776186] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:07:51.760 [2024-11-27 17:29:22.776243] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:07:51.760 [2024-11-27 17:29:22.776255] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:51.760 [2024-11-27 17:29:22.776265] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:51.760 17:29:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:51.760 17:29:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 2 00:07:51.760 17:29:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:51.760 17:29:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:51.760 17:29:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:51.760 17:29:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:51.760 17:29:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:51.760 17:29:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:51.760 17:29:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:51.760 17:29:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:51.760 17:29:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:51.760 17:29:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:51.760 17:29:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:51.760 17:29:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:51.760 17:29:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:51.760 17:29:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:51.760 17:29:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:51.760 "name": "Existed_Raid", 00:07:51.760 "uuid": "1b507c34-cbe6-44cd-b1cf-98f8bce9c3ce", 00:07:51.760 "strip_size_kb": 64, 00:07:51.760 "state": "configuring", 00:07:51.760 "raid_level": "concat", 00:07:51.760 "superblock": true, 00:07:51.760 "num_base_bdevs": 2, 00:07:51.760 "num_base_bdevs_discovered": 0, 00:07:51.760 "num_base_bdevs_operational": 2, 00:07:51.760 "base_bdevs_list": [ 00:07:51.760 { 00:07:51.760 "name": "BaseBdev1", 00:07:51.760 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:51.760 "is_configured": false, 00:07:51.760 "data_offset": 0, 00:07:51.760 "data_size": 0 00:07:51.760 }, 00:07:51.760 { 00:07:51.760 "name": "BaseBdev2", 00:07:51.760 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:51.760 "is_configured": false, 00:07:51.760 "data_offset": 0, 00:07:51.760 "data_size": 0 00:07:51.760 } 00:07:51.760 ] 00:07:51.760 }' 00:07:51.760 17:29:22 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:51.760 17:29:22 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:52.331 17:29:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:07:52.331 17:29:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:52.332 17:29:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:52.332 [2024-11-27 17:29:23.239234] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:07:52.332 [2024-11-27 17:29:23.239284] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:07:52.332 17:29:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:52.332 17:29:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:52.332 17:29:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:52.332 17:29:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:52.332 [2024-11-27 17:29:23.251248] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:07:52.332 [2024-11-27 17:29:23.251346] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:07:52.332 [2024-11-27 17:29:23.251388] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:52.332 [2024-11-27 17:29:23.251413] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:52.332 17:29:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:52.332 17:29:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:07:52.332 17:29:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:52.332 17:29:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:52.332 [2024-11-27 17:29:23.277968] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:52.332 BaseBdev1 00:07:52.332 17:29:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:52.332 17:29:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:07:52.332 17:29:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:07:52.332 17:29:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:07:52.332 17:29:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:07:52.332 17:29:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:07:52.332 17:29:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:07:52.332 17:29:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:07:52.332 17:29:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:52.332 17:29:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:52.332 17:29:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:52.332 17:29:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:07:52.332 17:29:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:52.332 17:29:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:52.332 [ 00:07:52.332 { 00:07:52.332 "name": "BaseBdev1", 00:07:52.332 "aliases": [ 00:07:52.332 "19907062-2875-4c5e-a5db-ae4ae1e75b7a" 00:07:52.332 ], 00:07:52.332 "product_name": "Malloc disk", 00:07:52.332 "block_size": 512, 00:07:52.332 "num_blocks": 65536, 00:07:52.332 "uuid": "19907062-2875-4c5e-a5db-ae4ae1e75b7a", 00:07:52.332 "assigned_rate_limits": { 00:07:52.332 "rw_ios_per_sec": 0, 00:07:52.332 "rw_mbytes_per_sec": 0, 00:07:52.332 "r_mbytes_per_sec": 0, 00:07:52.332 "w_mbytes_per_sec": 0 00:07:52.332 }, 00:07:52.332 "claimed": true, 00:07:52.332 "claim_type": "exclusive_write", 00:07:52.332 "zoned": false, 00:07:52.332 "supported_io_types": { 00:07:52.332 "read": true, 00:07:52.332 "write": true, 00:07:52.332 "unmap": true, 00:07:52.332 "flush": true, 00:07:52.332 "reset": true, 00:07:52.332 "nvme_admin": false, 00:07:52.332 "nvme_io": false, 00:07:52.332 "nvme_io_md": false, 00:07:52.332 "write_zeroes": true, 00:07:52.332 "zcopy": true, 00:07:52.332 "get_zone_info": false, 00:07:52.332 "zone_management": false, 00:07:52.332 "zone_append": false, 00:07:52.332 "compare": false, 00:07:52.332 "compare_and_write": false, 00:07:52.332 "abort": true, 00:07:52.332 "seek_hole": false, 00:07:52.332 "seek_data": false, 00:07:52.332 "copy": true, 00:07:52.332 "nvme_iov_md": false 00:07:52.332 }, 00:07:52.332 "memory_domains": [ 00:07:52.332 { 00:07:52.332 "dma_device_id": "system", 00:07:52.332 "dma_device_type": 1 00:07:52.332 }, 00:07:52.332 { 00:07:52.332 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:52.332 "dma_device_type": 2 00:07:52.332 } 00:07:52.332 ], 00:07:52.332 "driver_specific": {} 00:07:52.332 } 00:07:52.332 ] 00:07:52.332 17:29:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:52.332 17:29:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:07:52.332 17:29:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 2 00:07:52.332 17:29:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:52.332 17:29:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:52.332 17:29:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:52.332 17:29:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:52.332 17:29:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:52.332 17:29:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:52.332 17:29:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:52.332 17:29:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:52.332 17:29:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:52.332 17:29:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:52.332 17:29:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:52.332 17:29:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:52.332 17:29:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:52.332 17:29:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:52.332 17:29:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:52.332 "name": "Existed_Raid", 00:07:52.332 "uuid": "8f320e03-d6b0-4aa1-8d18-40d6b53b2830", 00:07:52.332 "strip_size_kb": 64, 00:07:52.332 "state": "configuring", 00:07:52.332 "raid_level": "concat", 00:07:52.332 "superblock": true, 00:07:52.332 "num_base_bdevs": 2, 00:07:52.332 "num_base_bdevs_discovered": 1, 00:07:52.332 "num_base_bdevs_operational": 2, 00:07:52.332 "base_bdevs_list": [ 00:07:52.332 { 00:07:52.332 "name": "BaseBdev1", 00:07:52.332 "uuid": "19907062-2875-4c5e-a5db-ae4ae1e75b7a", 00:07:52.332 "is_configured": true, 00:07:52.332 "data_offset": 2048, 00:07:52.332 "data_size": 63488 00:07:52.332 }, 00:07:52.332 { 00:07:52.332 "name": "BaseBdev2", 00:07:52.332 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:52.332 "is_configured": false, 00:07:52.332 "data_offset": 0, 00:07:52.332 "data_size": 0 00:07:52.332 } 00:07:52.332 ] 00:07:52.332 }' 00:07:52.332 17:29:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:52.332 17:29:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:52.592 17:29:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:07:52.592 17:29:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:52.592 17:29:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:52.592 [2024-11-27 17:29:23.773181] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:07:52.592 [2024-11-27 17:29:23.773228] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:07:52.592 17:29:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:52.592 17:29:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:07:52.592 17:29:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:52.592 17:29:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:52.852 [2024-11-27 17:29:23.781229] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:52.852 [2024-11-27 17:29:23.783351] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:07:52.852 [2024-11-27 17:29:23.783395] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:07:52.852 17:29:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:52.852 17:29:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:07:52.852 17:29:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:07:52.852 17:29:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 2 00:07:52.852 17:29:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:52.852 17:29:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:52.852 17:29:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:52.852 17:29:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:52.852 17:29:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:52.852 17:29:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:52.852 17:29:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:52.852 17:29:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:52.852 17:29:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:52.852 17:29:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:52.852 17:29:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:52.852 17:29:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:52.852 17:29:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:52.852 17:29:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:52.852 17:29:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:52.852 "name": "Existed_Raid", 00:07:52.852 "uuid": "52cbab3a-795c-48cd-96a5-40cb2959cd5f", 00:07:52.852 "strip_size_kb": 64, 00:07:52.852 "state": "configuring", 00:07:52.852 "raid_level": "concat", 00:07:52.852 "superblock": true, 00:07:52.852 "num_base_bdevs": 2, 00:07:52.852 "num_base_bdevs_discovered": 1, 00:07:52.852 "num_base_bdevs_operational": 2, 00:07:52.852 "base_bdevs_list": [ 00:07:52.852 { 00:07:52.852 "name": "BaseBdev1", 00:07:52.852 "uuid": "19907062-2875-4c5e-a5db-ae4ae1e75b7a", 00:07:52.852 "is_configured": true, 00:07:52.852 "data_offset": 2048, 00:07:52.852 "data_size": 63488 00:07:52.852 }, 00:07:52.852 { 00:07:52.852 "name": "BaseBdev2", 00:07:52.852 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:52.852 "is_configured": false, 00:07:52.852 "data_offset": 0, 00:07:52.852 "data_size": 0 00:07:52.852 } 00:07:52.852 ] 00:07:52.852 }' 00:07:52.852 17:29:23 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:52.852 17:29:23 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:53.113 17:29:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:07:53.113 17:29:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:53.113 17:29:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:53.113 [2024-11-27 17:29:24.239781] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:07:53.113 [2024-11-27 17:29:24.240601] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:07:53.113 [2024-11-27 17:29:24.240668] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:07:53.113 BaseBdev2 00:07:53.113 [2024-11-27 17:29:24.241575] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:07:53.113 17:29:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:53.113 [2024-11-27 17:29:24.242101] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:07:53.113 [2024-11-27 17:29:24.242193] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:07:53.113 17:29:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:07:53.113 [2024-11-27 17:29:24.242664] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:53.113 17:29:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:07:53.113 17:29:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:07:53.113 17:29:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:07:53.113 17:29:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:07:53.113 17:29:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:07:53.113 17:29:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:07:53.113 17:29:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:53.113 17:29:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:53.113 17:29:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:53.113 17:29:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:07:53.113 17:29:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:53.113 17:29:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:53.113 [ 00:07:53.113 { 00:07:53.113 "name": "BaseBdev2", 00:07:53.113 "aliases": [ 00:07:53.113 "2f741eda-0b23-4e95-be91-4f1958c7bed3" 00:07:53.113 ], 00:07:53.113 "product_name": "Malloc disk", 00:07:53.113 "block_size": 512, 00:07:53.113 "num_blocks": 65536, 00:07:53.113 "uuid": "2f741eda-0b23-4e95-be91-4f1958c7bed3", 00:07:53.113 "assigned_rate_limits": { 00:07:53.113 "rw_ios_per_sec": 0, 00:07:53.113 "rw_mbytes_per_sec": 0, 00:07:53.113 "r_mbytes_per_sec": 0, 00:07:53.113 "w_mbytes_per_sec": 0 00:07:53.113 }, 00:07:53.113 "claimed": true, 00:07:53.113 "claim_type": "exclusive_write", 00:07:53.113 "zoned": false, 00:07:53.113 "supported_io_types": { 00:07:53.113 "read": true, 00:07:53.113 "write": true, 00:07:53.113 "unmap": true, 00:07:53.113 "flush": true, 00:07:53.113 "reset": true, 00:07:53.113 "nvme_admin": false, 00:07:53.113 "nvme_io": false, 00:07:53.113 "nvme_io_md": false, 00:07:53.113 "write_zeroes": true, 00:07:53.113 "zcopy": true, 00:07:53.113 "get_zone_info": false, 00:07:53.113 "zone_management": false, 00:07:53.113 "zone_append": false, 00:07:53.113 "compare": false, 00:07:53.113 "compare_and_write": false, 00:07:53.113 "abort": true, 00:07:53.113 "seek_hole": false, 00:07:53.113 "seek_data": false, 00:07:53.113 "copy": true, 00:07:53.113 "nvme_iov_md": false 00:07:53.113 }, 00:07:53.113 "memory_domains": [ 00:07:53.113 { 00:07:53.113 "dma_device_id": "system", 00:07:53.113 "dma_device_type": 1 00:07:53.113 }, 00:07:53.113 { 00:07:53.113 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:53.113 "dma_device_type": 2 00:07:53.113 } 00:07:53.113 ], 00:07:53.113 "driver_specific": {} 00:07:53.113 } 00:07:53.113 ] 00:07:53.113 17:29:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:53.113 17:29:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:07:53.113 17:29:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:07:53.113 17:29:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:07:53.113 17:29:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online concat 64 2 00:07:53.113 17:29:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:53.113 17:29:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:53.113 17:29:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:53.113 17:29:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:53.113 17:29:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:53.113 17:29:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:53.113 17:29:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:53.113 17:29:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:53.113 17:29:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:53.113 17:29:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:53.113 17:29:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:53.113 17:29:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:53.113 17:29:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:53.373 17:29:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:53.373 17:29:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:53.373 "name": "Existed_Raid", 00:07:53.373 "uuid": "52cbab3a-795c-48cd-96a5-40cb2959cd5f", 00:07:53.373 "strip_size_kb": 64, 00:07:53.373 "state": "online", 00:07:53.373 "raid_level": "concat", 00:07:53.373 "superblock": true, 00:07:53.373 "num_base_bdevs": 2, 00:07:53.373 "num_base_bdevs_discovered": 2, 00:07:53.373 "num_base_bdevs_operational": 2, 00:07:53.373 "base_bdevs_list": [ 00:07:53.373 { 00:07:53.373 "name": "BaseBdev1", 00:07:53.373 "uuid": "19907062-2875-4c5e-a5db-ae4ae1e75b7a", 00:07:53.373 "is_configured": true, 00:07:53.373 "data_offset": 2048, 00:07:53.373 "data_size": 63488 00:07:53.373 }, 00:07:53.373 { 00:07:53.373 "name": "BaseBdev2", 00:07:53.373 "uuid": "2f741eda-0b23-4e95-be91-4f1958c7bed3", 00:07:53.373 "is_configured": true, 00:07:53.373 "data_offset": 2048, 00:07:53.373 "data_size": 63488 00:07:53.373 } 00:07:53.373 ] 00:07:53.373 }' 00:07:53.373 17:29:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:53.373 17:29:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:53.633 17:29:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:07:53.633 17:29:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:07:53.633 17:29:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:07:53.633 17:29:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:07:53.633 17:29:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:07:53.633 17:29:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:07:53.633 17:29:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:07:53.633 17:29:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:07:53.633 17:29:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:53.633 17:29:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:53.633 [2024-11-27 17:29:24.703197] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:53.633 17:29:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:53.633 17:29:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:07:53.633 "name": "Existed_Raid", 00:07:53.633 "aliases": [ 00:07:53.633 "52cbab3a-795c-48cd-96a5-40cb2959cd5f" 00:07:53.633 ], 00:07:53.633 "product_name": "Raid Volume", 00:07:53.633 "block_size": 512, 00:07:53.633 "num_blocks": 126976, 00:07:53.633 "uuid": "52cbab3a-795c-48cd-96a5-40cb2959cd5f", 00:07:53.633 "assigned_rate_limits": { 00:07:53.633 "rw_ios_per_sec": 0, 00:07:53.633 "rw_mbytes_per_sec": 0, 00:07:53.633 "r_mbytes_per_sec": 0, 00:07:53.633 "w_mbytes_per_sec": 0 00:07:53.633 }, 00:07:53.633 "claimed": false, 00:07:53.633 "zoned": false, 00:07:53.633 "supported_io_types": { 00:07:53.633 "read": true, 00:07:53.633 "write": true, 00:07:53.633 "unmap": true, 00:07:53.633 "flush": true, 00:07:53.633 "reset": true, 00:07:53.633 "nvme_admin": false, 00:07:53.633 "nvme_io": false, 00:07:53.633 "nvme_io_md": false, 00:07:53.633 "write_zeroes": true, 00:07:53.633 "zcopy": false, 00:07:53.633 "get_zone_info": false, 00:07:53.633 "zone_management": false, 00:07:53.633 "zone_append": false, 00:07:53.633 "compare": false, 00:07:53.633 "compare_and_write": false, 00:07:53.633 "abort": false, 00:07:53.633 "seek_hole": false, 00:07:53.633 "seek_data": false, 00:07:53.633 "copy": false, 00:07:53.633 "nvme_iov_md": false 00:07:53.633 }, 00:07:53.633 "memory_domains": [ 00:07:53.633 { 00:07:53.633 "dma_device_id": "system", 00:07:53.633 "dma_device_type": 1 00:07:53.633 }, 00:07:53.633 { 00:07:53.633 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:53.633 "dma_device_type": 2 00:07:53.633 }, 00:07:53.633 { 00:07:53.633 "dma_device_id": "system", 00:07:53.633 "dma_device_type": 1 00:07:53.633 }, 00:07:53.633 { 00:07:53.633 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:53.633 "dma_device_type": 2 00:07:53.633 } 00:07:53.633 ], 00:07:53.633 "driver_specific": { 00:07:53.633 "raid": { 00:07:53.633 "uuid": "52cbab3a-795c-48cd-96a5-40cb2959cd5f", 00:07:53.633 "strip_size_kb": 64, 00:07:53.633 "state": "online", 00:07:53.633 "raid_level": "concat", 00:07:53.633 "superblock": true, 00:07:53.633 "num_base_bdevs": 2, 00:07:53.633 "num_base_bdevs_discovered": 2, 00:07:53.633 "num_base_bdevs_operational": 2, 00:07:53.633 "base_bdevs_list": [ 00:07:53.633 { 00:07:53.633 "name": "BaseBdev1", 00:07:53.633 "uuid": "19907062-2875-4c5e-a5db-ae4ae1e75b7a", 00:07:53.633 "is_configured": true, 00:07:53.633 "data_offset": 2048, 00:07:53.633 "data_size": 63488 00:07:53.633 }, 00:07:53.633 { 00:07:53.633 "name": "BaseBdev2", 00:07:53.633 "uuid": "2f741eda-0b23-4e95-be91-4f1958c7bed3", 00:07:53.633 "is_configured": true, 00:07:53.633 "data_offset": 2048, 00:07:53.633 "data_size": 63488 00:07:53.633 } 00:07:53.633 ] 00:07:53.633 } 00:07:53.633 } 00:07:53.633 }' 00:07:53.633 17:29:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:07:53.633 17:29:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:07:53.633 BaseBdev2' 00:07:53.633 17:29:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:53.633 17:29:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:07:53.633 17:29:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:53.633 17:29:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:07:53.633 17:29:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:53.633 17:29:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:53.633 17:29:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:53.894 17:29:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:53.894 17:29:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:53.894 17:29:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:53.894 17:29:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:53.894 17:29:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:07:53.894 17:29:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:53.894 17:29:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:53.894 17:29:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:53.894 17:29:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:53.894 17:29:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:53.894 17:29:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:53.894 17:29:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:07:53.894 17:29:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:53.894 17:29:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:53.894 [2024-11-27 17:29:24.902642] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:07:53.894 [2024-11-27 17:29:24.902671] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:53.894 [2024-11-27 17:29:24.902716] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:53.894 17:29:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:53.894 17:29:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:07:53.894 17:29:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy concat 00:07:53.894 17:29:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:07:53.894 17:29:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # return 1 00:07:53.894 17:29:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:07:53.894 17:29:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline concat 64 1 00:07:53.894 17:29:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:07:53.894 17:29:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:07:53.894 17:29:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:53.894 17:29:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:53.894 17:29:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:07:53.894 17:29:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:53.894 17:29:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:53.894 17:29:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:53.894 17:29:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:53.894 17:29:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:53.894 17:29:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:53.894 17:29:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:53.894 17:29:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:07:53.894 17:29:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:53.894 17:29:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:53.894 "name": "Existed_Raid", 00:07:53.894 "uuid": "52cbab3a-795c-48cd-96a5-40cb2959cd5f", 00:07:53.894 "strip_size_kb": 64, 00:07:53.894 "state": "offline", 00:07:53.894 "raid_level": "concat", 00:07:53.894 "superblock": true, 00:07:53.894 "num_base_bdevs": 2, 00:07:53.894 "num_base_bdevs_discovered": 1, 00:07:53.894 "num_base_bdevs_operational": 1, 00:07:53.894 "base_bdevs_list": [ 00:07:53.894 { 00:07:53.894 "name": null, 00:07:53.894 "uuid": "00000000-0000-0000-0000-000000000000", 00:07:53.894 "is_configured": false, 00:07:53.894 "data_offset": 0, 00:07:53.894 "data_size": 63488 00:07:53.894 }, 00:07:53.894 { 00:07:53.894 "name": "BaseBdev2", 00:07:53.894 "uuid": "2f741eda-0b23-4e95-be91-4f1958c7bed3", 00:07:53.894 "is_configured": true, 00:07:53.894 "data_offset": 2048, 00:07:53.894 "data_size": 63488 00:07:53.894 } 00:07:53.894 ] 00:07:53.894 }' 00:07:53.894 17:29:24 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:53.894 17:29:24 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:54.154 17:29:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:07:54.154 17:29:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:07:54.154 17:29:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:54.154 17:29:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:54.154 17:29:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:54.154 17:29:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:07:54.154 17:29:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:54.154 17:29:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:07:54.154 17:29:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:07:54.154 17:29:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:07:54.154 17:29:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:54.154 17:29:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:54.414 [2024-11-27 17:29:25.346521] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:07:54.414 [2024-11-27 17:29:25.346630] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:07:54.414 17:29:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:54.414 17:29:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:07:54.414 17:29:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:07:54.414 17:29:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:54.414 17:29:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:07:54.414 17:29:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:54.414 17:29:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:54.414 17:29:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:54.414 17:29:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:07:54.414 17:29:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:07:54.414 17:29:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:07:54.414 17:29:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 73083 00:07:54.414 17:29:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@950 -- # '[' -z 73083 ']' 00:07:54.414 17:29:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # kill -0 73083 00:07:54.414 17:29:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@955 -- # uname 00:07:54.414 17:29:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:07:54.414 17:29:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 73083 00:07:54.414 17:29:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:07:54.414 17:29:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:07:54.414 17:29:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@968 -- # echo 'killing process with pid 73083' 00:07:54.414 killing process with pid 73083 00:07:54.414 17:29:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@969 -- # kill 73083 00:07:54.414 [2024-11-27 17:29:25.463318] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:54.414 17:29:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@974 -- # wait 73083 00:07:54.414 [2024-11-27 17:29:25.464896] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:54.675 17:29:25 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:07:54.675 00:07:54.675 real 0m3.971s 00:07:54.675 user 0m6.039s 00:07:54.675 sys 0m0.863s 00:07:54.675 17:29:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1126 -- # xtrace_disable 00:07:54.675 ************************************ 00:07:54.675 END TEST raid_state_function_test_sb 00:07:54.675 ************************************ 00:07:54.675 17:29:25 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:07:54.935 17:29:25 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test concat 2 00:07:54.935 17:29:25 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:07:54.935 17:29:25 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:07:54.935 17:29:25 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:54.935 ************************************ 00:07:54.935 START TEST raid_superblock_test 00:07:54.935 ************************************ 00:07:54.935 17:29:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1125 -- # raid_superblock_test concat 2 00:07:54.935 17:29:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=concat 00:07:54.935 17:29:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=2 00:07:54.935 17:29:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:07:54.935 17:29:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:07:54.935 17:29:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:07:54.935 17:29:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:07:54.935 17:29:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:07:54.935 17:29:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:07:54.935 17:29:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:07:54.935 17:29:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:07:54.935 17:29:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:07:54.935 17:29:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:07:54.935 17:29:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:07:54.935 17:29:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' concat '!=' raid1 ']' 00:07:54.936 17:29:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@405 -- # strip_size=64 00:07:54.936 17:29:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@406 -- # strip_size_create_arg='-z 64' 00:07:54.936 17:29:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=73324 00:07:54.936 17:29:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:07:54.936 17:29:25 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 73324 00:07:54.936 17:29:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@831 -- # '[' -z 73324 ']' 00:07:54.936 17:29:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:54.936 17:29:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:07:54.936 17:29:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:54.936 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:54.936 17:29:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:07:54.936 17:29:25 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:54.936 [2024-11-27 17:29:26.000453] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:07:54.936 [2024-11-27 17:29:26.000703] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid73324 ] 00:07:55.195 [2024-11-27 17:29:26.146543] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:55.195 [2024-11-27 17:29:26.220353] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:07:55.195 [2024-11-27 17:29:26.295793] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:55.195 [2024-11-27 17:29:26.295932] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:55.765 17:29:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:07:55.765 17:29:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@864 -- # return 0 00:07:55.765 17:29:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:07:55.765 17:29:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:07:55.765 17:29:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:07:55.765 17:29:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:07:55.765 17:29:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:07:55.765 17:29:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:07:55.765 17:29:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:07:55.765 17:29:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:07:55.765 17:29:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:07:55.765 17:29:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:55.765 17:29:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:55.765 malloc1 00:07:55.765 17:29:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:55.765 17:29:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:07:55.765 17:29:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:55.765 17:29:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:55.765 [2024-11-27 17:29:26.833541] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:07:55.765 [2024-11-27 17:29:26.833637] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:55.765 [2024-11-27 17:29:26.833686] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:07:55.765 [2024-11-27 17:29:26.833719] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:55.765 [2024-11-27 17:29:26.836127] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:55.765 [2024-11-27 17:29:26.836212] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:07:55.765 pt1 00:07:55.765 17:29:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:55.765 17:29:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:07:55.765 17:29:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:07:55.765 17:29:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:07:55.765 17:29:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:07:55.765 17:29:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:07:55.765 17:29:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:07:55.765 17:29:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:07:55.765 17:29:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:07:55.765 17:29:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:07:55.765 17:29:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:55.765 17:29:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:55.765 malloc2 00:07:55.765 17:29:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:55.765 17:29:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:07:55.765 17:29:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:55.765 17:29:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:55.765 [2024-11-27 17:29:26.884162] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:07:55.765 [2024-11-27 17:29:26.884375] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:55.765 [2024-11-27 17:29:26.884425] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:07:55.766 [2024-11-27 17:29:26.884455] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:55.766 [2024-11-27 17:29:26.888814] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:55.766 [2024-11-27 17:29:26.888875] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:07:55.766 pt2 00:07:55.766 17:29:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:55.766 17:29:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:07:55.766 17:29:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:07:55.766 17:29:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''pt1 pt2'\''' -n raid_bdev1 -s 00:07:55.766 17:29:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:55.766 17:29:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:55.766 [2024-11-27 17:29:26.897118] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:07:55.766 [2024-11-27 17:29:26.899778] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:07:55.766 [2024-11-27 17:29:26.900017] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:07:55.766 [2024-11-27 17:29:26.900043] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:07:55.766 [2024-11-27 17:29:26.900425] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:07:55.766 [2024-11-27 17:29:26.900598] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:07:55.766 [2024-11-27 17:29:26.900612] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:07:55.766 [2024-11-27 17:29:26.900773] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:55.766 17:29:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:55.766 17:29:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online concat 64 2 00:07:55.766 17:29:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:55.766 17:29:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:55.766 17:29:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:55.766 17:29:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:55.766 17:29:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:55.766 17:29:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:55.766 17:29:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:55.766 17:29:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:55.766 17:29:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:55.766 17:29:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:55.766 17:29:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:55.766 17:29:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:55.766 17:29:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:55.766 17:29:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:56.026 17:29:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:56.026 "name": "raid_bdev1", 00:07:56.026 "uuid": "f65feec2-a19d-47e0-8f38-e025fd6e768c", 00:07:56.026 "strip_size_kb": 64, 00:07:56.026 "state": "online", 00:07:56.026 "raid_level": "concat", 00:07:56.026 "superblock": true, 00:07:56.026 "num_base_bdevs": 2, 00:07:56.026 "num_base_bdevs_discovered": 2, 00:07:56.026 "num_base_bdevs_operational": 2, 00:07:56.026 "base_bdevs_list": [ 00:07:56.026 { 00:07:56.026 "name": "pt1", 00:07:56.026 "uuid": "00000000-0000-0000-0000-000000000001", 00:07:56.026 "is_configured": true, 00:07:56.026 "data_offset": 2048, 00:07:56.026 "data_size": 63488 00:07:56.026 }, 00:07:56.026 { 00:07:56.026 "name": "pt2", 00:07:56.026 "uuid": "00000000-0000-0000-0000-000000000002", 00:07:56.026 "is_configured": true, 00:07:56.026 "data_offset": 2048, 00:07:56.026 "data_size": 63488 00:07:56.026 } 00:07:56.026 ] 00:07:56.026 }' 00:07:56.026 17:29:26 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:56.026 17:29:26 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:56.286 17:29:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:07:56.286 17:29:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:07:56.286 17:29:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:07:56.286 17:29:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:07:56.286 17:29:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:07:56.286 17:29:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:07:56.286 17:29:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:07:56.286 17:29:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:07:56.286 17:29:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:56.286 17:29:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:56.286 [2024-11-27 17:29:27.332585] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:56.286 17:29:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:56.286 17:29:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:07:56.286 "name": "raid_bdev1", 00:07:56.286 "aliases": [ 00:07:56.286 "f65feec2-a19d-47e0-8f38-e025fd6e768c" 00:07:56.286 ], 00:07:56.286 "product_name": "Raid Volume", 00:07:56.286 "block_size": 512, 00:07:56.286 "num_blocks": 126976, 00:07:56.286 "uuid": "f65feec2-a19d-47e0-8f38-e025fd6e768c", 00:07:56.286 "assigned_rate_limits": { 00:07:56.286 "rw_ios_per_sec": 0, 00:07:56.286 "rw_mbytes_per_sec": 0, 00:07:56.286 "r_mbytes_per_sec": 0, 00:07:56.286 "w_mbytes_per_sec": 0 00:07:56.286 }, 00:07:56.286 "claimed": false, 00:07:56.286 "zoned": false, 00:07:56.286 "supported_io_types": { 00:07:56.286 "read": true, 00:07:56.286 "write": true, 00:07:56.286 "unmap": true, 00:07:56.286 "flush": true, 00:07:56.286 "reset": true, 00:07:56.286 "nvme_admin": false, 00:07:56.286 "nvme_io": false, 00:07:56.286 "nvme_io_md": false, 00:07:56.286 "write_zeroes": true, 00:07:56.286 "zcopy": false, 00:07:56.286 "get_zone_info": false, 00:07:56.286 "zone_management": false, 00:07:56.286 "zone_append": false, 00:07:56.286 "compare": false, 00:07:56.286 "compare_and_write": false, 00:07:56.286 "abort": false, 00:07:56.286 "seek_hole": false, 00:07:56.286 "seek_data": false, 00:07:56.286 "copy": false, 00:07:56.286 "nvme_iov_md": false 00:07:56.286 }, 00:07:56.286 "memory_domains": [ 00:07:56.286 { 00:07:56.286 "dma_device_id": "system", 00:07:56.286 "dma_device_type": 1 00:07:56.286 }, 00:07:56.286 { 00:07:56.286 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:56.286 "dma_device_type": 2 00:07:56.286 }, 00:07:56.286 { 00:07:56.286 "dma_device_id": "system", 00:07:56.286 "dma_device_type": 1 00:07:56.286 }, 00:07:56.286 { 00:07:56.286 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:56.286 "dma_device_type": 2 00:07:56.286 } 00:07:56.286 ], 00:07:56.286 "driver_specific": { 00:07:56.286 "raid": { 00:07:56.286 "uuid": "f65feec2-a19d-47e0-8f38-e025fd6e768c", 00:07:56.286 "strip_size_kb": 64, 00:07:56.286 "state": "online", 00:07:56.286 "raid_level": "concat", 00:07:56.286 "superblock": true, 00:07:56.286 "num_base_bdevs": 2, 00:07:56.286 "num_base_bdevs_discovered": 2, 00:07:56.286 "num_base_bdevs_operational": 2, 00:07:56.286 "base_bdevs_list": [ 00:07:56.286 { 00:07:56.286 "name": "pt1", 00:07:56.286 "uuid": "00000000-0000-0000-0000-000000000001", 00:07:56.286 "is_configured": true, 00:07:56.286 "data_offset": 2048, 00:07:56.286 "data_size": 63488 00:07:56.286 }, 00:07:56.286 { 00:07:56.286 "name": "pt2", 00:07:56.286 "uuid": "00000000-0000-0000-0000-000000000002", 00:07:56.286 "is_configured": true, 00:07:56.286 "data_offset": 2048, 00:07:56.286 "data_size": 63488 00:07:56.286 } 00:07:56.286 ] 00:07:56.286 } 00:07:56.286 } 00:07:56.286 }' 00:07:56.286 17:29:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:07:56.286 17:29:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:07:56.286 pt2' 00:07:56.286 17:29:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:56.286 17:29:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:07:56.286 17:29:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:56.286 17:29:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:56.286 17:29:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:07:56.286 17:29:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:56.286 17:29:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:56.286 17:29:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:56.546 17:29:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:56.546 17:29:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:56.546 17:29:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:56.546 17:29:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:07:56.546 17:29:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:56.546 17:29:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:56.546 17:29:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:56.546 17:29:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:56.546 17:29:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:56.546 17:29:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:56.546 17:29:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:07:56.546 17:29:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:07:56.546 17:29:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:56.546 17:29:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:56.546 [2024-11-27 17:29:27.556108] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:56.546 17:29:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:56.546 17:29:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=f65feec2-a19d-47e0-8f38-e025fd6e768c 00:07:56.546 17:29:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z f65feec2-a19d-47e0-8f38-e025fd6e768c ']' 00:07:56.546 17:29:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:07:56.547 17:29:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:56.547 17:29:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:56.547 [2024-11-27 17:29:27.599780] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:07:56.547 [2024-11-27 17:29:27.599848] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:07:56.547 [2024-11-27 17:29:27.599953] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:56.547 [2024-11-27 17:29:27.600046] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:56.547 [2024-11-27 17:29:27.600109] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:07:56.547 17:29:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:56.547 17:29:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:56.547 17:29:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:07:56.547 17:29:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:56.547 17:29:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:56.547 17:29:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:56.547 17:29:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:07:56.547 17:29:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:07:56.547 17:29:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:07:56.547 17:29:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:07:56.547 17:29:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:56.547 17:29:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:56.547 17:29:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:56.547 17:29:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:07:56.547 17:29:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:07:56.547 17:29:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:56.547 17:29:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:56.547 17:29:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:56.547 17:29:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:07:56.547 17:29:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:07:56.547 17:29:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:56.547 17:29:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:56.547 17:29:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:56.547 17:29:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:07:56.547 17:29:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:07:56.807 17:29:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@650 -- # local es=0 00:07:56.807 17:29:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:07:56.807 17:29:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:07:56.807 17:29:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:07:56.807 17:29:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:07:56.807 17:29:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:07:56.807 17:29:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:07:56.807 17:29:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:56.807 17:29:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:56.807 [2024-11-27 17:29:27.751541] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:07:56.807 [2024-11-27 17:29:27.753693] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:07:56.807 [2024-11-27 17:29:27.753807] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:07:56.807 [2024-11-27 17:29:27.753861] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:07:56.807 [2024-11-27 17:29:27.753880] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:07:56.807 [2024-11-27 17:29:27.753889] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state configuring 00:07:56.807 request: 00:07:56.807 { 00:07:56.807 "name": "raid_bdev1", 00:07:56.807 "raid_level": "concat", 00:07:56.807 "base_bdevs": [ 00:07:56.807 "malloc1", 00:07:56.807 "malloc2" 00:07:56.807 ], 00:07:56.807 "strip_size_kb": 64, 00:07:56.807 "superblock": false, 00:07:56.807 "method": "bdev_raid_create", 00:07:56.807 "req_id": 1 00:07:56.807 } 00:07:56.807 Got JSON-RPC error response 00:07:56.807 response: 00:07:56.807 { 00:07:56.807 "code": -17, 00:07:56.807 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:07:56.807 } 00:07:56.807 17:29:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:07:56.807 17:29:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@653 -- # es=1 00:07:56.807 17:29:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:07:56.807 17:29:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:07:56.807 17:29:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:07:56.807 17:29:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:56.807 17:29:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:07:56.807 17:29:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:56.807 17:29:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:56.807 17:29:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:56.807 17:29:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:07:56.807 17:29:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:07:56.807 17:29:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:07:56.807 17:29:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:56.807 17:29:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:56.807 [2024-11-27 17:29:27.815384] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:07:56.807 [2024-11-27 17:29:27.815492] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:56.807 [2024-11-27 17:29:27.815522] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:07:56.807 [2024-11-27 17:29:27.815531] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:56.807 [2024-11-27 17:29:27.817933] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:56.807 [2024-11-27 17:29:27.817966] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:07:56.807 [2024-11-27 17:29:27.818029] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:07:56.807 [2024-11-27 17:29:27.818058] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:07:56.807 pt1 00:07:56.807 17:29:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:56.807 17:29:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring concat 64 2 00:07:56.807 17:29:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:56.807 17:29:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:07:56.807 17:29:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:56.807 17:29:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:56.807 17:29:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:56.807 17:29:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:56.807 17:29:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:56.807 17:29:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:56.807 17:29:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:56.808 17:29:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:56.808 17:29:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:56.808 17:29:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:56.808 17:29:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:56.808 17:29:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:56.808 17:29:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:56.808 "name": "raid_bdev1", 00:07:56.808 "uuid": "f65feec2-a19d-47e0-8f38-e025fd6e768c", 00:07:56.808 "strip_size_kb": 64, 00:07:56.808 "state": "configuring", 00:07:56.808 "raid_level": "concat", 00:07:56.808 "superblock": true, 00:07:56.808 "num_base_bdevs": 2, 00:07:56.808 "num_base_bdevs_discovered": 1, 00:07:56.808 "num_base_bdevs_operational": 2, 00:07:56.808 "base_bdevs_list": [ 00:07:56.808 { 00:07:56.808 "name": "pt1", 00:07:56.808 "uuid": "00000000-0000-0000-0000-000000000001", 00:07:56.808 "is_configured": true, 00:07:56.808 "data_offset": 2048, 00:07:56.808 "data_size": 63488 00:07:56.808 }, 00:07:56.808 { 00:07:56.808 "name": null, 00:07:56.808 "uuid": "00000000-0000-0000-0000-000000000002", 00:07:56.808 "is_configured": false, 00:07:56.808 "data_offset": 2048, 00:07:56.808 "data_size": 63488 00:07:56.808 } 00:07:56.808 ] 00:07:56.808 }' 00:07:56.808 17:29:27 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:56.808 17:29:27 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:57.068 17:29:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 2 -gt 2 ']' 00:07:57.068 17:29:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:07:57.068 17:29:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:07:57.068 17:29:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:07:57.068 17:29:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:57.068 17:29:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:57.068 [2024-11-27 17:29:28.254732] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:07:57.068 [2024-11-27 17:29:28.254825] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:57.068 [2024-11-27 17:29:28.254885] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:07:57.068 [2024-11-27 17:29:28.254916] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:57.068 [2024-11-27 17:29:28.255353] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:57.068 [2024-11-27 17:29:28.255412] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:07:57.068 [2024-11-27 17:29:28.255502] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:07:57.068 [2024-11-27 17:29:28.255547] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:07:57.068 [2024-11-27 17:29:28.255663] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:07:57.068 [2024-11-27 17:29:28.255700] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:07:57.068 [2024-11-27 17:29:28.255986] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:07:57.068 [2024-11-27 17:29:28.256129] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:07:57.068 [2024-11-27 17:29:28.256183] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:07:57.068 [2024-11-27 17:29:28.256316] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:57.328 pt2 00:07:57.328 17:29:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:57.328 17:29:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:07:57.328 17:29:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:07:57.328 17:29:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online concat 64 2 00:07:57.328 17:29:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:57.328 17:29:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:57.328 17:29:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:57.328 17:29:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:57.328 17:29:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:57.328 17:29:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:57.328 17:29:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:57.328 17:29:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:57.328 17:29:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:57.328 17:29:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:57.328 17:29:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:57.328 17:29:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:57.328 17:29:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:57.328 17:29:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:57.328 17:29:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:57.328 "name": "raid_bdev1", 00:07:57.328 "uuid": "f65feec2-a19d-47e0-8f38-e025fd6e768c", 00:07:57.328 "strip_size_kb": 64, 00:07:57.328 "state": "online", 00:07:57.328 "raid_level": "concat", 00:07:57.328 "superblock": true, 00:07:57.328 "num_base_bdevs": 2, 00:07:57.328 "num_base_bdevs_discovered": 2, 00:07:57.328 "num_base_bdevs_operational": 2, 00:07:57.329 "base_bdevs_list": [ 00:07:57.329 { 00:07:57.329 "name": "pt1", 00:07:57.329 "uuid": "00000000-0000-0000-0000-000000000001", 00:07:57.329 "is_configured": true, 00:07:57.329 "data_offset": 2048, 00:07:57.329 "data_size": 63488 00:07:57.329 }, 00:07:57.329 { 00:07:57.329 "name": "pt2", 00:07:57.329 "uuid": "00000000-0000-0000-0000-000000000002", 00:07:57.329 "is_configured": true, 00:07:57.329 "data_offset": 2048, 00:07:57.329 "data_size": 63488 00:07:57.329 } 00:07:57.329 ] 00:07:57.329 }' 00:07:57.329 17:29:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:57.329 17:29:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:57.589 17:29:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:07:57.589 17:29:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:07:57.589 17:29:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:07:57.589 17:29:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:07:57.589 17:29:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:07:57.589 17:29:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:07:57.589 17:29:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:07:57.589 17:29:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:07:57.589 17:29:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:57.589 17:29:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:57.589 [2024-11-27 17:29:28.638304] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:57.589 17:29:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:57.589 17:29:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:07:57.589 "name": "raid_bdev1", 00:07:57.589 "aliases": [ 00:07:57.589 "f65feec2-a19d-47e0-8f38-e025fd6e768c" 00:07:57.589 ], 00:07:57.589 "product_name": "Raid Volume", 00:07:57.589 "block_size": 512, 00:07:57.589 "num_blocks": 126976, 00:07:57.589 "uuid": "f65feec2-a19d-47e0-8f38-e025fd6e768c", 00:07:57.589 "assigned_rate_limits": { 00:07:57.589 "rw_ios_per_sec": 0, 00:07:57.589 "rw_mbytes_per_sec": 0, 00:07:57.589 "r_mbytes_per_sec": 0, 00:07:57.589 "w_mbytes_per_sec": 0 00:07:57.589 }, 00:07:57.589 "claimed": false, 00:07:57.589 "zoned": false, 00:07:57.589 "supported_io_types": { 00:07:57.589 "read": true, 00:07:57.589 "write": true, 00:07:57.589 "unmap": true, 00:07:57.589 "flush": true, 00:07:57.589 "reset": true, 00:07:57.589 "nvme_admin": false, 00:07:57.589 "nvme_io": false, 00:07:57.589 "nvme_io_md": false, 00:07:57.589 "write_zeroes": true, 00:07:57.589 "zcopy": false, 00:07:57.589 "get_zone_info": false, 00:07:57.589 "zone_management": false, 00:07:57.589 "zone_append": false, 00:07:57.589 "compare": false, 00:07:57.589 "compare_and_write": false, 00:07:57.589 "abort": false, 00:07:57.589 "seek_hole": false, 00:07:57.589 "seek_data": false, 00:07:57.589 "copy": false, 00:07:57.589 "nvme_iov_md": false 00:07:57.589 }, 00:07:57.589 "memory_domains": [ 00:07:57.589 { 00:07:57.589 "dma_device_id": "system", 00:07:57.589 "dma_device_type": 1 00:07:57.589 }, 00:07:57.589 { 00:07:57.589 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:57.589 "dma_device_type": 2 00:07:57.589 }, 00:07:57.589 { 00:07:57.589 "dma_device_id": "system", 00:07:57.589 "dma_device_type": 1 00:07:57.589 }, 00:07:57.589 { 00:07:57.589 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:57.589 "dma_device_type": 2 00:07:57.589 } 00:07:57.589 ], 00:07:57.589 "driver_specific": { 00:07:57.589 "raid": { 00:07:57.589 "uuid": "f65feec2-a19d-47e0-8f38-e025fd6e768c", 00:07:57.589 "strip_size_kb": 64, 00:07:57.589 "state": "online", 00:07:57.589 "raid_level": "concat", 00:07:57.589 "superblock": true, 00:07:57.589 "num_base_bdevs": 2, 00:07:57.589 "num_base_bdevs_discovered": 2, 00:07:57.589 "num_base_bdevs_operational": 2, 00:07:57.589 "base_bdevs_list": [ 00:07:57.589 { 00:07:57.589 "name": "pt1", 00:07:57.589 "uuid": "00000000-0000-0000-0000-000000000001", 00:07:57.589 "is_configured": true, 00:07:57.589 "data_offset": 2048, 00:07:57.589 "data_size": 63488 00:07:57.589 }, 00:07:57.589 { 00:07:57.589 "name": "pt2", 00:07:57.589 "uuid": "00000000-0000-0000-0000-000000000002", 00:07:57.589 "is_configured": true, 00:07:57.589 "data_offset": 2048, 00:07:57.589 "data_size": 63488 00:07:57.589 } 00:07:57.589 ] 00:07:57.589 } 00:07:57.589 } 00:07:57.589 }' 00:07:57.589 17:29:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:07:57.589 17:29:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:07:57.589 pt2' 00:07:57.589 17:29:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:57.589 17:29:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:07:57.589 17:29:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:57.589 17:29:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:07:57.589 17:29:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:57.589 17:29:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:57.589 17:29:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:57.589 17:29:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:57.850 17:29:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:57.850 17:29:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:57.850 17:29:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:07:57.850 17:29:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:07:57.850 17:29:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:57.850 17:29:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:57.850 17:29:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:07:57.850 17:29:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:57.850 17:29:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:07:57.850 17:29:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:07:57.850 17:29:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:07:57.850 17:29:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:07:57.850 17:29:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:57.850 17:29:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:57.850 [2024-11-27 17:29:28.857888] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:07:57.850 17:29:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:57.850 17:29:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' f65feec2-a19d-47e0-8f38-e025fd6e768c '!=' f65feec2-a19d-47e0-8f38-e025fd6e768c ']' 00:07:57.850 17:29:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy concat 00:07:57.850 17:29:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:07:57.850 17:29:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # return 1 00:07:57.850 17:29:28 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 73324 00:07:57.850 17:29:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@950 -- # '[' -z 73324 ']' 00:07:57.850 17:29:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # kill -0 73324 00:07:57.850 17:29:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@955 -- # uname 00:07:57.850 17:29:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:07:57.850 17:29:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 73324 00:07:57.850 killing process with pid 73324 00:07:57.850 17:29:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:07:57.850 17:29:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:07:57.850 17:29:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 73324' 00:07:57.850 17:29:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@969 -- # kill 73324 00:07:57.850 [2024-11-27 17:29:28.911073] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:07:57.850 [2024-11-27 17:29:28.911165] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:07:57.850 [2024-11-27 17:29:28.911215] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:07:57.850 [2024-11-27 17:29:28.911223] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:07:57.850 17:29:28 bdev_raid.raid_superblock_test -- common/autotest_common.sh@974 -- # wait 73324 00:07:57.850 [2024-11-27 17:29:28.952373] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:07:58.425 ************************************ 00:07:58.425 END TEST raid_superblock_test 00:07:58.425 ************************************ 00:07:58.425 17:29:29 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:07:58.425 00:07:58.425 real 0m3.412s 00:07:58.425 user 0m5.039s 00:07:58.425 sys 0m0.810s 00:07:58.425 17:29:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:07:58.425 17:29:29 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:07:58.425 17:29:29 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test concat 2 read 00:07:58.425 17:29:29 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:07:58.425 17:29:29 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:07:58.425 17:29:29 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:07:58.425 ************************************ 00:07:58.425 START TEST raid_read_error_test 00:07:58.425 ************************************ 00:07:58.425 17:29:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1125 -- # raid_io_error_test concat 2 read 00:07:58.425 17:29:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=concat 00:07:58.425 17:29:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=2 00:07:58.426 17:29:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:07:58.426 17:29:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:07:58.426 17:29:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:58.426 17:29:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:07:58.426 17:29:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:07:58.426 17:29:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:58.426 17:29:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:07:58.426 17:29:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:07:58.426 17:29:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:07:58.426 17:29:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:07:58.426 17:29:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:07:58.426 17:29:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:07:58.426 17:29:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:07:58.426 17:29:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:07:58.426 17:29:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:07:58.426 17:29:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:07:58.426 17:29:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' concat '!=' raid1 ']' 00:07:58.426 17:29:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:07:58.426 17:29:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:07:58.426 17:29:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:07:58.426 17:29:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.lOsq63ZZGO 00:07:58.426 17:29:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=73524 00:07:58.426 17:29:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:07:58.426 17:29:29 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 73524 00:07:58.426 17:29:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@831 -- # '[' -z 73524 ']' 00:07:58.426 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:58.426 17:29:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:58.426 17:29:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:07:58.426 17:29:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:58.426 17:29:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:07:58.426 17:29:29 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:58.426 [2024-11-27 17:29:29.494040] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:07:58.426 [2024-11-27 17:29:29.494175] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid73524 ] 00:07:58.686 [2024-11-27 17:29:29.621091] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:58.686 [2024-11-27 17:29:29.687098] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:07:58.686 [2024-11-27 17:29:29.762269] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:58.686 [2024-11-27 17:29:29.762302] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:07:59.256 17:29:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:07:59.256 17:29:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@864 -- # return 0 00:07:59.256 17:29:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:07:59.256 17:29:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:07:59.256 17:29:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:59.256 17:29:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:59.256 BaseBdev1_malloc 00:07:59.256 17:29:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:59.256 17:29:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:07:59.256 17:29:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:59.256 17:29:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:59.256 true 00:07:59.256 17:29:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:59.256 17:29:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:07:59.256 17:29:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:59.256 17:29:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:59.256 [2024-11-27 17:29:30.359430] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:07:59.256 [2024-11-27 17:29:30.359493] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:59.256 [2024-11-27 17:29:30.359522] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:07:59.256 [2024-11-27 17:29:30.359531] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:59.256 [2024-11-27 17:29:30.361945] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:59.256 [2024-11-27 17:29:30.361990] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:07:59.256 BaseBdev1 00:07:59.256 17:29:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:59.256 17:29:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:07:59.256 17:29:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:07:59.256 17:29:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:59.256 17:29:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:59.256 BaseBdev2_malloc 00:07:59.256 17:29:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:59.256 17:29:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:07:59.256 17:29:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:59.256 17:29:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:59.256 true 00:07:59.256 17:29:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:59.256 17:29:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:07:59.256 17:29:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:59.256 17:29:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:59.256 [2024-11-27 17:29:30.417493] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:07:59.256 [2024-11-27 17:29:30.417551] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:59.256 [2024-11-27 17:29:30.417572] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:07:59.256 [2024-11-27 17:29:30.417582] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:59.256 [2024-11-27 17:29:30.420013] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:59.256 [2024-11-27 17:29:30.420106] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:07:59.256 BaseBdev2 00:07:59.257 17:29:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:59.257 17:29:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 -s 00:07:59.257 17:29:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:59.257 17:29:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:59.257 [2024-11-27 17:29:30.429566] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:07:59.257 [2024-11-27 17:29:30.431746] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:07:59.257 [2024-11-27 17:29:30.431948] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:07:59.257 [2024-11-27 17:29:30.431961] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:07:59.257 [2024-11-27 17:29:30.432269] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:07:59.257 [2024-11-27 17:29:30.432410] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:07:59.257 [2024-11-27 17:29:30.432429] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:07:59.257 [2024-11-27 17:29:30.432573] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:07:59.257 17:29:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:59.257 17:29:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online concat 64 2 00:07:59.257 17:29:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:07:59.257 17:29:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:07:59.257 17:29:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:07:59.257 17:29:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:07:59.257 17:29:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:07:59.257 17:29:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:07:59.257 17:29:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:07:59.257 17:29:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:07:59.257 17:29:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:07:59.257 17:29:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:07:59.257 17:29:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:07:59.257 17:29:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:59.257 17:29:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:59.516 17:29:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:59.517 17:29:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:07:59.517 "name": "raid_bdev1", 00:07:59.517 "uuid": "15b368cd-8229-436f-9c05-56857acd11ef", 00:07:59.517 "strip_size_kb": 64, 00:07:59.517 "state": "online", 00:07:59.517 "raid_level": "concat", 00:07:59.517 "superblock": true, 00:07:59.517 "num_base_bdevs": 2, 00:07:59.517 "num_base_bdevs_discovered": 2, 00:07:59.517 "num_base_bdevs_operational": 2, 00:07:59.517 "base_bdevs_list": [ 00:07:59.517 { 00:07:59.517 "name": "BaseBdev1", 00:07:59.517 "uuid": "3001f230-7d7e-5a0b-ad5b-c94d98fa91f3", 00:07:59.517 "is_configured": true, 00:07:59.517 "data_offset": 2048, 00:07:59.517 "data_size": 63488 00:07:59.517 }, 00:07:59.517 { 00:07:59.517 "name": "BaseBdev2", 00:07:59.517 "uuid": "b6090661-d0f2-5295-9b8d-6b0ed0656937", 00:07:59.517 "is_configured": true, 00:07:59.517 "data_offset": 2048, 00:07:59.517 "data_size": 63488 00:07:59.517 } 00:07:59.517 ] 00:07:59.517 }' 00:07:59.517 17:29:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:07:59.517 17:29:30 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:07:59.777 17:29:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:07:59.777 17:29:30 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:07:59.777 [2024-11-27 17:29:30.949063] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:08:00.717 17:29:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:08:00.717 17:29:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:00.717 17:29:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:00.717 17:29:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:00.717 17:29:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:08:00.717 17:29:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ concat = \r\a\i\d\1 ]] 00:08:00.717 17:29:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=2 00:08:00.717 17:29:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online concat 64 2 00:08:00.717 17:29:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:00.717 17:29:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:00.717 17:29:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:00.717 17:29:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:00.717 17:29:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:00.717 17:29:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:00.717 17:29:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:00.717 17:29:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:00.717 17:29:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:00.717 17:29:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:00.717 17:29:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:00.717 17:29:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:00.717 17:29:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:00.977 17:29:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:00.977 17:29:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:00.977 "name": "raid_bdev1", 00:08:00.977 "uuid": "15b368cd-8229-436f-9c05-56857acd11ef", 00:08:00.977 "strip_size_kb": 64, 00:08:00.977 "state": "online", 00:08:00.977 "raid_level": "concat", 00:08:00.977 "superblock": true, 00:08:00.977 "num_base_bdevs": 2, 00:08:00.977 "num_base_bdevs_discovered": 2, 00:08:00.977 "num_base_bdevs_operational": 2, 00:08:00.977 "base_bdevs_list": [ 00:08:00.977 { 00:08:00.977 "name": "BaseBdev1", 00:08:00.977 "uuid": "3001f230-7d7e-5a0b-ad5b-c94d98fa91f3", 00:08:00.977 "is_configured": true, 00:08:00.977 "data_offset": 2048, 00:08:00.977 "data_size": 63488 00:08:00.977 }, 00:08:00.977 { 00:08:00.977 "name": "BaseBdev2", 00:08:00.977 "uuid": "b6090661-d0f2-5295-9b8d-6b0ed0656937", 00:08:00.977 "is_configured": true, 00:08:00.977 "data_offset": 2048, 00:08:00.977 "data_size": 63488 00:08:00.977 } 00:08:00.977 ] 00:08:00.977 }' 00:08:00.977 17:29:31 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:00.977 17:29:31 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:01.238 17:29:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:08:01.238 17:29:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:01.238 17:29:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:01.238 [2024-11-27 17:29:32.313403] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:08:01.238 [2024-11-27 17:29:32.313502] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:01.238 [2024-11-27 17:29:32.316099] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:01.238 [2024-11-27 17:29:32.316219] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:01.238 [2024-11-27 17:29:32.316277] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:01.238 [2024-11-27 17:29:32.316336] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:08:01.238 { 00:08:01.238 "results": [ 00:08:01.238 { 00:08:01.238 "job": "raid_bdev1", 00:08:01.238 "core_mask": "0x1", 00:08:01.238 "workload": "randrw", 00:08:01.238 "percentage": 50, 00:08:01.238 "status": "finished", 00:08:01.238 "queue_depth": 1, 00:08:01.238 "io_size": 131072, 00:08:01.238 "runtime": 1.365031, 00:08:01.238 "iops": 15814.292862213386, 00:08:01.238 "mibps": 1976.7866077766732, 00:08:01.238 "io_failed": 1, 00:08:01.238 "io_timeout": 0, 00:08:01.238 "avg_latency_us": 88.43814734532286, 00:08:01.238 "min_latency_us": 24.593886462882097, 00:08:01.238 "max_latency_us": 1366.5257641921398 00:08:01.238 } 00:08:01.238 ], 00:08:01.238 "core_count": 1 00:08:01.238 } 00:08:01.238 17:29:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:01.238 17:29:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 73524 00:08:01.238 17:29:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@950 -- # '[' -z 73524 ']' 00:08:01.238 17:29:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # kill -0 73524 00:08:01.238 17:29:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@955 -- # uname 00:08:01.238 17:29:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:08:01.238 17:29:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 73524 00:08:01.238 17:29:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:08:01.238 17:29:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:08:01.238 17:29:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 73524' 00:08:01.238 killing process with pid 73524 00:08:01.238 17:29:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@969 -- # kill 73524 00:08:01.238 [2024-11-27 17:29:32.354072] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:01.238 17:29:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@974 -- # wait 73524 00:08:01.238 [2024-11-27 17:29:32.381528] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:01.808 17:29:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:08:01.808 17:29:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:08:01.808 17:29:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.lOsq63ZZGO 00:08:01.808 17:29:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.73 00:08:01.808 17:29:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy concat 00:08:01.808 17:29:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:08:01.808 17:29:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:08:01.808 17:29:32 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.73 != \0\.\0\0 ]] 00:08:01.808 00:08:01.808 real 0m3.360s 00:08:01.808 user 0m4.112s 00:08:01.808 sys 0m0.585s 00:08:01.808 ************************************ 00:08:01.808 END TEST raid_read_error_test 00:08:01.808 ************************************ 00:08:01.808 17:29:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:08:01.808 17:29:32 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:01.808 17:29:32 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test concat 2 write 00:08:01.808 17:29:32 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:08:01.808 17:29:32 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:08:01.808 17:29:32 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:01.808 ************************************ 00:08:01.808 START TEST raid_write_error_test 00:08:01.808 ************************************ 00:08:01.808 17:29:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1125 -- # raid_io_error_test concat 2 write 00:08:01.808 17:29:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=concat 00:08:01.808 17:29:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=2 00:08:01.808 17:29:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:08:01.808 17:29:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:08:01.808 17:29:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:01.808 17:29:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:08:01.808 17:29:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:08:01.808 17:29:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:01.808 17:29:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:08:01.808 17:29:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:08:01.808 17:29:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:01.808 17:29:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:08:01.808 17:29:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:08:01.808 17:29:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:08:01.808 17:29:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:08:01.808 17:29:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:08:01.808 17:29:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:08:01.808 17:29:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:08:01.808 17:29:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' concat '!=' raid1 ']' 00:08:01.808 17:29:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:08:01.808 17:29:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:08:01.808 17:29:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:08:01.808 17:29:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.izH3qPDV7y 00:08:01.808 17:29:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=73659 00:08:01.808 17:29:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:08:01.808 17:29:32 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 73659 00:08:01.808 17:29:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@831 -- # '[' -z 73659 ']' 00:08:01.808 17:29:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:01.808 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:01.808 17:29:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:08:01.808 17:29:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:01.808 17:29:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:08:01.808 17:29:32 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:01.808 [2024-11-27 17:29:32.924406] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:08:01.808 [2024-11-27 17:29:32.924533] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid73659 ] 00:08:02.068 [2024-11-27 17:29:33.069050] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:02.068 [2024-11-27 17:29:33.136019] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:08:02.068 [2024-11-27 17:29:33.210976] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:02.069 [2024-11-27 17:29:33.211015] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:02.639 17:29:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:08:02.639 17:29:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@864 -- # return 0 00:08:02.639 17:29:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:08:02.639 17:29:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:08:02.639 17:29:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:02.639 17:29:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:02.639 BaseBdev1_malloc 00:08:02.639 17:29:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:02.639 17:29:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:08:02.639 17:29:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:02.639 17:29:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:02.639 true 00:08:02.639 17:29:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:02.639 17:29:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:08:02.639 17:29:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:02.639 17:29:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:02.639 [2024-11-27 17:29:33.773675] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:08:02.639 [2024-11-27 17:29:33.773734] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:02.639 [2024-11-27 17:29:33.773756] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:08:02.640 [2024-11-27 17:29:33.773765] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:02.640 [2024-11-27 17:29:33.776104] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:02.640 [2024-11-27 17:29:33.776155] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:08:02.640 BaseBdev1 00:08:02.640 17:29:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:02.640 17:29:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:08:02.640 17:29:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:08:02.640 17:29:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:02.640 17:29:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:02.640 BaseBdev2_malloc 00:08:02.640 17:29:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:02.640 17:29:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:08:02.640 17:29:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:02.640 17:29:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:02.900 true 00:08:02.900 17:29:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:02.900 17:29:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:08:02.900 17:29:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:02.900 17:29:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:02.900 [2024-11-27 17:29:33.837020] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:08:02.900 [2024-11-27 17:29:33.837098] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:02.900 [2024-11-27 17:29:33.837129] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:08:02.900 [2024-11-27 17:29:33.837158] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:02.900 [2024-11-27 17:29:33.840554] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:02.900 [2024-11-27 17:29:33.840597] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:08:02.900 BaseBdev2 00:08:02.900 17:29:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:02.900 17:29:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 -s 00:08:02.900 17:29:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:02.900 17:29:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:02.900 [2024-11-27 17:29:33.849269] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:02.900 [2024-11-27 17:29:33.851532] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:02.900 [2024-11-27 17:29:33.851738] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:08:02.900 [2024-11-27 17:29:33.851760] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:08:02.900 [2024-11-27 17:29:33.852042] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:08:02.900 [2024-11-27 17:29:33.852206] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:08:02.900 [2024-11-27 17:29:33.852234] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:08:02.900 [2024-11-27 17:29:33.852390] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:02.900 17:29:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:02.900 17:29:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online concat 64 2 00:08:02.900 17:29:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:02.900 17:29:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:02.900 17:29:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:02.900 17:29:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:02.900 17:29:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:02.900 17:29:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:02.900 17:29:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:02.900 17:29:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:02.900 17:29:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:02.900 17:29:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:02.900 17:29:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:02.901 17:29:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:02.901 17:29:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:02.901 17:29:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:02.901 17:29:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:02.901 "name": "raid_bdev1", 00:08:02.901 "uuid": "ec1f7abd-e0ec-47b7-ab8d-9db2e41c5d1e", 00:08:02.901 "strip_size_kb": 64, 00:08:02.901 "state": "online", 00:08:02.901 "raid_level": "concat", 00:08:02.901 "superblock": true, 00:08:02.901 "num_base_bdevs": 2, 00:08:02.901 "num_base_bdevs_discovered": 2, 00:08:02.901 "num_base_bdevs_operational": 2, 00:08:02.901 "base_bdevs_list": [ 00:08:02.901 { 00:08:02.901 "name": "BaseBdev1", 00:08:02.901 "uuid": "f2cc42e4-fd26-5809-b6d2-f37d7c1adefa", 00:08:02.901 "is_configured": true, 00:08:02.901 "data_offset": 2048, 00:08:02.901 "data_size": 63488 00:08:02.901 }, 00:08:02.901 { 00:08:02.901 "name": "BaseBdev2", 00:08:02.901 "uuid": "acd55a60-3c64-50df-ba90-7ad9abe4b247", 00:08:02.901 "is_configured": true, 00:08:02.901 "data_offset": 2048, 00:08:02.901 "data_size": 63488 00:08:02.901 } 00:08:02.901 ] 00:08:02.901 }' 00:08:02.901 17:29:33 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:02.901 17:29:33 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:03.161 17:29:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:08:03.161 17:29:34 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:08:03.421 [2024-11-27 17:29:34.364785] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:08:04.365 17:29:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:08:04.365 17:29:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:04.365 17:29:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:04.365 17:29:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:04.365 17:29:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:08:04.365 17:29:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ concat = \r\a\i\d\1 ]] 00:08:04.365 17:29:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=2 00:08:04.365 17:29:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online concat 64 2 00:08:04.365 17:29:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:04.365 17:29:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:04.365 17:29:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:04.365 17:29:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:04.365 17:29:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:04.365 17:29:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:04.365 17:29:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:04.365 17:29:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:04.365 17:29:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:04.365 17:29:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:04.365 17:29:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:04.365 17:29:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:04.365 17:29:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:04.365 17:29:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:04.365 17:29:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:04.365 "name": "raid_bdev1", 00:08:04.365 "uuid": "ec1f7abd-e0ec-47b7-ab8d-9db2e41c5d1e", 00:08:04.365 "strip_size_kb": 64, 00:08:04.365 "state": "online", 00:08:04.365 "raid_level": "concat", 00:08:04.365 "superblock": true, 00:08:04.365 "num_base_bdevs": 2, 00:08:04.365 "num_base_bdevs_discovered": 2, 00:08:04.365 "num_base_bdevs_operational": 2, 00:08:04.365 "base_bdevs_list": [ 00:08:04.365 { 00:08:04.365 "name": "BaseBdev1", 00:08:04.365 "uuid": "f2cc42e4-fd26-5809-b6d2-f37d7c1adefa", 00:08:04.365 "is_configured": true, 00:08:04.365 "data_offset": 2048, 00:08:04.365 "data_size": 63488 00:08:04.365 }, 00:08:04.365 { 00:08:04.365 "name": "BaseBdev2", 00:08:04.365 "uuid": "acd55a60-3c64-50df-ba90-7ad9abe4b247", 00:08:04.365 "is_configured": true, 00:08:04.365 "data_offset": 2048, 00:08:04.365 "data_size": 63488 00:08:04.365 } 00:08:04.365 ] 00:08:04.365 }' 00:08:04.365 17:29:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:04.365 17:29:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:04.625 17:29:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:08:04.625 17:29:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:04.625 17:29:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:04.625 [2024-11-27 17:29:35.737163] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:08:04.625 [2024-11-27 17:29:35.737200] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:04.625 [2024-11-27 17:29:35.739723] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:04.625 [2024-11-27 17:29:35.739774] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:04.625 [2024-11-27 17:29:35.739814] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:04.625 [2024-11-27 17:29:35.739829] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:08:04.625 { 00:08:04.625 "results": [ 00:08:04.625 { 00:08:04.625 "job": "raid_bdev1", 00:08:04.625 "core_mask": "0x1", 00:08:04.625 "workload": "randrw", 00:08:04.625 "percentage": 50, 00:08:04.625 "status": "finished", 00:08:04.625 "queue_depth": 1, 00:08:04.625 "io_size": 131072, 00:08:04.625 "runtime": 1.373001, 00:08:04.625 "iops": 15932.253508919513, 00:08:04.625 "mibps": 1991.5316886149392, 00:08:04.625 "io_failed": 1, 00:08:04.625 "io_timeout": 0, 00:08:04.625 "avg_latency_us": 87.69562336663736, 00:08:04.625 "min_latency_us": 24.370305676855896, 00:08:04.625 "max_latency_us": 1366.5257641921398 00:08:04.625 } 00:08:04.625 ], 00:08:04.625 "core_count": 1 00:08:04.625 } 00:08:04.625 17:29:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:04.625 17:29:35 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 73659 00:08:04.625 17:29:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@950 -- # '[' -z 73659 ']' 00:08:04.625 17:29:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # kill -0 73659 00:08:04.625 17:29:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@955 -- # uname 00:08:04.625 17:29:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:08:04.625 17:29:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 73659 00:08:04.625 17:29:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:08:04.625 17:29:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:08:04.625 killing process with pid 73659 00:08:04.625 17:29:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 73659' 00:08:04.625 17:29:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@969 -- # kill 73659 00:08:04.625 17:29:35 bdev_raid.raid_write_error_test -- common/autotest_common.sh@974 -- # wait 73659 00:08:04.625 [2024-11-27 17:29:35.783819] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:04.625 [2024-11-27 17:29:35.813207] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:05.196 17:29:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.izH3qPDV7y 00:08:05.196 17:29:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:08:05.196 17:29:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:08:05.196 17:29:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.73 00:08:05.196 17:29:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy concat 00:08:05.196 17:29:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:08:05.196 17:29:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:08:05.196 17:29:36 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.73 != \0\.\0\0 ]] 00:08:05.196 00:08:05.196 real 0m3.355s 00:08:05.196 user 0m4.107s 00:08:05.196 sys 0m0.594s 00:08:05.196 17:29:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:08:05.196 17:29:36 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:05.196 ************************************ 00:08:05.196 END TEST raid_write_error_test 00:08:05.196 ************************************ 00:08:05.196 17:29:36 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:08:05.196 17:29:36 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test raid1 2 false 00:08:05.196 17:29:36 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:08:05.196 17:29:36 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:08:05.196 17:29:36 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:05.196 ************************************ 00:08:05.196 START TEST raid_state_function_test 00:08:05.196 ************************************ 00:08:05.196 17:29:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1125 -- # raid_state_function_test raid1 2 false 00:08:05.196 17:29:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:08:05.196 17:29:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:08:05.196 17:29:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:08:05.196 17:29:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:08:05.196 17:29:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:08:05.196 17:29:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:05.196 17:29:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:08:05.196 17:29:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:05.196 17:29:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:05.196 17:29:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:08:05.196 17:29:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:05.196 17:29:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:05.196 17:29:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:08:05.196 17:29:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:08:05.196 17:29:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:08:05.196 17:29:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:08:05.196 17:29:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:08:05.196 17:29:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:08:05.196 17:29:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:08:05.196 17:29:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:08:05.196 17:29:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:08:05.196 17:29:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:08:05.196 17:29:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=73786 00:08:05.196 17:29:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:08:05.196 Process raid pid: 73786 00:08:05.196 17:29:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 73786' 00:08:05.196 17:29:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 73786 00:08:05.196 17:29:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@831 -- # '[' -z 73786 ']' 00:08:05.196 17:29:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:05.197 17:29:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:08:05.197 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:05.197 17:29:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:05.197 17:29:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:08:05.197 17:29:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:05.197 [2024-11-27 17:29:36.358083] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:08:05.197 [2024-11-27 17:29:36.358395] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:08:05.456 [2024-11-27 17:29:36.505945] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:05.456 [2024-11-27 17:29:36.576557] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:08:05.717 [2024-11-27 17:29:36.653202] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:05.717 [2024-11-27 17:29:36.653245] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:06.287 17:29:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:08:06.287 17:29:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@864 -- # return 0 00:08:06.287 17:29:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:08:06.287 17:29:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:06.287 17:29:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:06.287 [2024-11-27 17:29:37.181355] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:06.287 [2024-11-27 17:29:37.181406] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:06.287 [2024-11-27 17:29:37.181428] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:06.287 [2024-11-27 17:29:37.181439] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:06.287 17:29:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:06.287 17:29:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:08:06.287 17:29:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:06.287 17:29:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:06.287 17:29:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:06.287 17:29:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:06.287 17:29:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:06.287 17:29:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:06.287 17:29:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:06.287 17:29:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:06.287 17:29:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:06.287 17:29:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:06.287 17:29:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:06.287 17:29:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:06.287 17:29:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:06.287 17:29:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:06.287 17:29:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:06.287 "name": "Existed_Raid", 00:08:06.287 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:06.287 "strip_size_kb": 0, 00:08:06.287 "state": "configuring", 00:08:06.287 "raid_level": "raid1", 00:08:06.287 "superblock": false, 00:08:06.287 "num_base_bdevs": 2, 00:08:06.287 "num_base_bdevs_discovered": 0, 00:08:06.287 "num_base_bdevs_operational": 2, 00:08:06.287 "base_bdevs_list": [ 00:08:06.287 { 00:08:06.287 "name": "BaseBdev1", 00:08:06.287 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:06.287 "is_configured": false, 00:08:06.287 "data_offset": 0, 00:08:06.287 "data_size": 0 00:08:06.287 }, 00:08:06.287 { 00:08:06.287 "name": "BaseBdev2", 00:08:06.287 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:06.287 "is_configured": false, 00:08:06.287 "data_offset": 0, 00:08:06.287 "data_size": 0 00:08:06.287 } 00:08:06.287 ] 00:08:06.287 }' 00:08:06.287 17:29:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:06.287 17:29:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:06.549 17:29:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:06.549 17:29:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:06.549 17:29:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:06.549 [2024-11-27 17:29:37.644396] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:06.549 [2024-11-27 17:29:37.644447] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:08:06.549 17:29:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:06.549 17:29:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:08:06.549 17:29:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:06.549 17:29:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:06.549 [2024-11-27 17:29:37.656389] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:06.549 [2024-11-27 17:29:37.656425] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:06.549 [2024-11-27 17:29:37.656459] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:06.549 [2024-11-27 17:29:37.656470] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:06.549 17:29:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:06.549 17:29:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:08:06.549 17:29:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:06.549 17:29:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:06.549 [2024-11-27 17:29:37.683232] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:06.549 BaseBdev1 00:08:06.549 17:29:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:06.549 17:29:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:08:06.549 17:29:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:08:06.549 17:29:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:08:06.549 17:29:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:08:06.549 17:29:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:08:06.549 17:29:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:08:06.549 17:29:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:08:06.549 17:29:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:06.549 17:29:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:06.549 17:29:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:06.549 17:29:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:08:06.549 17:29:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:06.549 17:29:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:06.549 [ 00:08:06.549 { 00:08:06.549 "name": "BaseBdev1", 00:08:06.549 "aliases": [ 00:08:06.549 "3bf6d492-8fff-4782-86a2-b1a59dca68af" 00:08:06.549 ], 00:08:06.549 "product_name": "Malloc disk", 00:08:06.549 "block_size": 512, 00:08:06.549 "num_blocks": 65536, 00:08:06.549 "uuid": "3bf6d492-8fff-4782-86a2-b1a59dca68af", 00:08:06.549 "assigned_rate_limits": { 00:08:06.549 "rw_ios_per_sec": 0, 00:08:06.549 "rw_mbytes_per_sec": 0, 00:08:06.549 "r_mbytes_per_sec": 0, 00:08:06.549 "w_mbytes_per_sec": 0 00:08:06.549 }, 00:08:06.549 "claimed": true, 00:08:06.549 "claim_type": "exclusive_write", 00:08:06.549 "zoned": false, 00:08:06.549 "supported_io_types": { 00:08:06.549 "read": true, 00:08:06.549 "write": true, 00:08:06.549 "unmap": true, 00:08:06.549 "flush": true, 00:08:06.549 "reset": true, 00:08:06.549 "nvme_admin": false, 00:08:06.549 "nvme_io": false, 00:08:06.549 "nvme_io_md": false, 00:08:06.549 "write_zeroes": true, 00:08:06.549 "zcopy": true, 00:08:06.549 "get_zone_info": false, 00:08:06.549 "zone_management": false, 00:08:06.549 "zone_append": false, 00:08:06.549 "compare": false, 00:08:06.549 "compare_and_write": false, 00:08:06.549 "abort": true, 00:08:06.549 "seek_hole": false, 00:08:06.549 "seek_data": false, 00:08:06.549 "copy": true, 00:08:06.549 "nvme_iov_md": false 00:08:06.549 }, 00:08:06.549 "memory_domains": [ 00:08:06.549 { 00:08:06.549 "dma_device_id": "system", 00:08:06.549 "dma_device_type": 1 00:08:06.549 }, 00:08:06.549 { 00:08:06.549 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:06.549 "dma_device_type": 2 00:08:06.549 } 00:08:06.549 ], 00:08:06.549 "driver_specific": {} 00:08:06.549 } 00:08:06.549 ] 00:08:06.549 17:29:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:06.549 17:29:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:08:06.549 17:29:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:08:06.549 17:29:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:06.549 17:29:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:06.549 17:29:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:06.549 17:29:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:06.549 17:29:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:06.549 17:29:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:06.549 17:29:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:06.549 17:29:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:06.549 17:29:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:06.549 17:29:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:06.549 17:29:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:06.549 17:29:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:06.549 17:29:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:06.810 17:29:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:06.810 17:29:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:06.810 "name": "Existed_Raid", 00:08:06.810 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:06.810 "strip_size_kb": 0, 00:08:06.810 "state": "configuring", 00:08:06.810 "raid_level": "raid1", 00:08:06.810 "superblock": false, 00:08:06.810 "num_base_bdevs": 2, 00:08:06.810 "num_base_bdevs_discovered": 1, 00:08:06.810 "num_base_bdevs_operational": 2, 00:08:06.810 "base_bdevs_list": [ 00:08:06.810 { 00:08:06.810 "name": "BaseBdev1", 00:08:06.810 "uuid": "3bf6d492-8fff-4782-86a2-b1a59dca68af", 00:08:06.810 "is_configured": true, 00:08:06.810 "data_offset": 0, 00:08:06.810 "data_size": 65536 00:08:06.810 }, 00:08:06.810 { 00:08:06.810 "name": "BaseBdev2", 00:08:06.810 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:06.810 "is_configured": false, 00:08:06.810 "data_offset": 0, 00:08:06.810 "data_size": 0 00:08:06.810 } 00:08:06.810 ] 00:08:06.810 }' 00:08:06.810 17:29:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:06.810 17:29:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:07.070 17:29:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:07.070 17:29:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:07.070 17:29:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:07.070 [2024-11-27 17:29:38.158481] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:07.070 [2024-11-27 17:29:38.158538] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:08:07.070 17:29:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:07.070 17:29:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:08:07.070 17:29:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:07.070 17:29:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:07.070 [2024-11-27 17:29:38.170500] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:07.070 [2024-11-27 17:29:38.172629] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:07.070 [2024-11-27 17:29:38.172672] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:07.070 17:29:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:07.070 17:29:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:08:07.070 17:29:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:07.070 17:29:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:08:07.070 17:29:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:07.070 17:29:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:07.070 17:29:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:07.070 17:29:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:07.070 17:29:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:07.070 17:29:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:07.070 17:29:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:07.070 17:29:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:07.070 17:29:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:07.070 17:29:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:07.070 17:29:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:07.070 17:29:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:07.070 17:29:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:07.070 17:29:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:07.070 17:29:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:07.070 "name": "Existed_Raid", 00:08:07.070 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:07.070 "strip_size_kb": 0, 00:08:07.070 "state": "configuring", 00:08:07.070 "raid_level": "raid1", 00:08:07.070 "superblock": false, 00:08:07.070 "num_base_bdevs": 2, 00:08:07.070 "num_base_bdevs_discovered": 1, 00:08:07.070 "num_base_bdevs_operational": 2, 00:08:07.070 "base_bdevs_list": [ 00:08:07.070 { 00:08:07.070 "name": "BaseBdev1", 00:08:07.070 "uuid": "3bf6d492-8fff-4782-86a2-b1a59dca68af", 00:08:07.070 "is_configured": true, 00:08:07.070 "data_offset": 0, 00:08:07.070 "data_size": 65536 00:08:07.070 }, 00:08:07.070 { 00:08:07.070 "name": "BaseBdev2", 00:08:07.070 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:07.070 "is_configured": false, 00:08:07.070 "data_offset": 0, 00:08:07.070 "data_size": 0 00:08:07.070 } 00:08:07.070 ] 00:08:07.070 }' 00:08:07.070 17:29:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:07.071 17:29:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:07.638 17:29:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:08:07.638 17:29:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:07.638 17:29:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:07.638 [2024-11-27 17:29:38.621333] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:07.638 [2024-11-27 17:29:38.621498] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:08:07.638 [2024-11-27 17:29:38.621537] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:08:07.638 [2024-11-27 17:29:38.622504] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:08:07.638 [2024-11-27 17:29:38.623055] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:08:07.638 [2024-11-27 17:29:38.623167] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:08:07.638 [2024-11-27 17:29:38.623749] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:07.638 BaseBdev2 00:08:07.638 17:29:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:07.638 17:29:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:08:07.638 17:29:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:08:07.638 17:29:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:08:07.638 17:29:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:08:07.638 17:29:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:08:07.638 17:29:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:08:07.638 17:29:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:08:07.638 17:29:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:07.638 17:29:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:07.638 17:29:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:07.638 17:29:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:08:07.638 17:29:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:07.638 17:29:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:07.638 [ 00:08:07.638 { 00:08:07.638 "name": "BaseBdev2", 00:08:07.638 "aliases": [ 00:08:07.638 "97e61142-d752-4c8a-b209-5bbf75ab1672" 00:08:07.638 ], 00:08:07.638 "product_name": "Malloc disk", 00:08:07.638 "block_size": 512, 00:08:07.638 "num_blocks": 65536, 00:08:07.638 "uuid": "97e61142-d752-4c8a-b209-5bbf75ab1672", 00:08:07.638 "assigned_rate_limits": { 00:08:07.638 "rw_ios_per_sec": 0, 00:08:07.638 "rw_mbytes_per_sec": 0, 00:08:07.638 "r_mbytes_per_sec": 0, 00:08:07.638 "w_mbytes_per_sec": 0 00:08:07.638 }, 00:08:07.638 "claimed": true, 00:08:07.638 "claim_type": "exclusive_write", 00:08:07.638 "zoned": false, 00:08:07.638 "supported_io_types": { 00:08:07.638 "read": true, 00:08:07.638 "write": true, 00:08:07.638 "unmap": true, 00:08:07.638 "flush": true, 00:08:07.638 "reset": true, 00:08:07.638 "nvme_admin": false, 00:08:07.638 "nvme_io": false, 00:08:07.638 "nvme_io_md": false, 00:08:07.638 "write_zeroes": true, 00:08:07.638 "zcopy": true, 00:08:07.638 "get_zone_info": false, 00:08:07.638 "zone_management": false, 00:08:07.638 "zone_append": false, 00:08:07.638 "compare": false, 00:08:07.638 "compare_and_write": false, 00:08:07.638 "abort": true, 00:08:07.638 "seek_hole": false, 00:08:07.638 "seek_data": false, 00:08:07.638 "copy": true, 00:08:07.638 "nvme_iov_md": false 00:08:07.638 }, 00:08:07.638 "memory_domains": [ 00:08:07.638 { 00:08:07.638 "dma_device_id": "system", 00:08:07.639 "dma_device_type": 1 00:08:07.639 }, 00:08:07.639 { 00:08:07.639 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:07.639 "dma_device_type": 2 00:08:07.639 } 00:08:07.639 ], 00:08:07.639 "driver_specific": {} 00:08:07.639 } 00:08:07.639 ] 00:08:07.639 17:29:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:07.639 17:29:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:08:07.639 17:29:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:08:07.639 17:29:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:07.639 17:29:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 2 00:08:07.639 17:29:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:07.639 17:29:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:07.639 17:29:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:07.639 17:29:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:07.639 17:29:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:07.639 17:29:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:07.639 17:29:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:07.639 17:29:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:07.639 17:29:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:07.639 17:29:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:07.639 17:29:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:07.639 17:29:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:07.639 17:29:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:07.639 17:29:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:07.639 17:29:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:07.639 "name": "Existed_Raid", 00:08:07.639 "uuid": "e5b41d73-0627-499e-bfd3-2e2e7e8b1e86", 00:08:07.639 "strip_size_kb": 0, 00:08:07.639 "state": "online", 00:08:07.639 "raid_level": "raid1", 00:08:07.639 "superblock": false, 00:08:07.639 "num_base_bdevs": 2, 00:08:07.639 "num_base_bdevs_discovered": 2, 00:08:07.639 "num_base_bdevs_operational": 2, 00:08:07.639 "base_bdevs_list": [ 00:08:07.639 { 00:08:07.639 "name": "BaseBdev1", 00:08:07.639 "uuid": "3bf6d492-8fff-4782-86a2-b1a59dca68af", 00:08:07.639 "is_configured": true, 00:08:07.639 "data_offset": 0, 00:08:07.639 "data_size": 65536 00:08:07.639 }, 00:08:07.639 { 00:08:07.639 "name": "BaseBdev2", 00:08:07.639 "uuid": "97e61142-d752-4c8a-b209-5bbf75ab1672", 00:08:07.639 "is_configured": true, 00:08:07.639 "data_offset": 0, 00:08:07.639 "data_size": 65536 00:08:07.639 } 00:08:07.639 ] 00:08:07.639 }' 00:08:07.639 17:29:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:07.639 17:29:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:08.206 17:29:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:08:08.206 17:29:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:08:08.206 17:29:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:08.206 17:29:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:08.206 17:29:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:08:08.206 17:29:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:08.206 17:29:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:08:08.206 17:29:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:08.206 17:29:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:08.206 17:29:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:08.206 [2024-11-27 17:29:39.128677] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:08.206 17:29:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:08.206 17:29:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:08.206 "name": "Existed_Raid", 00:08:08.206 "aliases": [ 00:08:08.206 "e5b41d73-0627-499e-bfd3-2e2e7e8b1e86" 00:08:08.206 ], 00:08:08.206 "product_name": "Raid Volume", 00:08:08.206 "block_size": 512, 00:08:08.206 "num_blocks": 65536, 00:08:08.206 "uuid": "e5b41d73-0627-499e-bfd3-2e2e7e8b1e86", 00:08:08.206 "assigned_rate_limits": { 00:08:08.206 "rw_ios_per_sec": 0, 00:08:08.206 "rw_mbytes_per_sec": 0, 00:08:08.206 "r_mbytes_per_sec": 0, 00:08:08.206 "w_mbytes_per_sec": 0 00:08:08.206 }, 00:08:08.206 "claimed": false, 00:08:08.206 "zoned": false, 00:08:08.206 "supported_io_types": { 00:08:08.206 "read": true, 00:08:08.206 "write": true, 00:08:08.206 "unmap": false, 00:08:08.206 "flush": false, 00:08:08.206 "reset": true, 00:08:08.206 "nvme_admin": false, 00:08:08.206 "nvme_io": false, 00:08:08.206 "nvme_io_md": false, 00:08:08.206 "write_zeroes": true, 00:08:08.206 "zcopy": false, 00:08:08.206 "get_zone_info": false, 00:08:08.206 "zone_management": false, 00:08:08.206 "zone_append": false, 00:08:08.206 "compare": false, 00:08:08.206 "compare_and_write": false, 00:08:08.206 "abort": false, 00:08:08.206 "seek_hole": false, 00:08:08.206 "seek_data": false, 00:08:08.206 "copy": false, 00:08:08.206 "nvme_iov_md": false 00:08:08.206 }, 00:08:08.206 "memory_domains": [ 00:08:08.206 { 00:08:08.206 "dma_device_id": "system", 00:08:08.206 "dma_device_type": 1 00:08:08.206 }, 00:08:08.206 { 00:08:08.206 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:08.206 "dma_device_type": 2 00:08:08.206 }, 00:08:08.206 { 00:08:08.206 "dma_device_id": "system", 00:08:08.206 "dma_device_type": 1 00:08:08.206 }, 00:08:08.206 { 00:08:08.206 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:08.206 "dma_device_type": 2 00:08:08.206 } 00:08:08.206 ], 00:08:08.206 "driver_specific": { 00:08:08.206 "raid": { 00:08:08.206 "uuid": "e5b41d73-0627-499e-bfd3-2e2e7e8b1e86", 00:08:08.206 "strip_size_kb": 0, 00:08:08.206 "state": "online", 00:08:08.206 "raid_level": "raid1", 00:08:08.206 "superblock": false, 00:08:08.206 "num_base_bdevs": 2, 00:08:08.206 "num_base_bdevs_discovered": 2, 00:08:08.206 "num_base_bdevs_operational": 2, 00:08:08.206 "base_bdevs_list": [ 00:08:08.206 { 00:08:08.206 "name": "BaseBdev1", 00:08:08.206 "uuid": "3bf6d492-8fff-4782-86a2-b1a59dca68af", 00:08:08.206 "is_configured": true, 00:08:08.206 "data_offset": 0, 00:08:08.206 "data_size": 65536 00:08:08.206 }, 00:08:08.206 { 00:08:08.206 "name": "BaseBdev2", 00:08:08.206 "uuid": "97e61142-d752-4c8a-b209-5bbf75ab1672", 00:08:08.207 "is_configured": true, 00:08:08.207 "data_offset": 0, 00:08:08.207 "data_size": 65536 00:08:08.207 } 00:08:08.207 ] 00:08:08.207 } 00:08:08.207 } 00:08:08.207 }' 00:08:08.207 17:29:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:08.207 17:29:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:08:08.207 BaseBdev2' 00:08:08.207 17:29:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:08.207 17:29:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:08.207 17:29:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:08.207 17:29:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:08:08.207 17:29:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:08.207 17:29:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:08.207 17:29:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:08.207 17:29:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:08.207 17:29:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:08.207 17:29:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:08.207 17:29:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:08.207 17:29:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:08:08.207 17:29:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:08.207 17:29:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:08.207 17:29:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:08.207 17:29:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:08.207 17:29:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:08.207 17:29:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:08.207 17:29:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:08:08.207 17:29:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:08.207 17:29:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:08.207 [2024-11-27 17:29:39.360039] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:08:08.207 17:29:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:08.207 17:29:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:08:08.207 17:29:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:08:08.207 17:29:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:08:08.207 17:29:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@199 -- # return 0 00:08:08.207 17:29:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:08:08.207 17:29:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 1 00:08:08.207 17:29:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:08.207 17:29:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:08.207 17:29:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:08.207 17:29:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:08.207 17:29:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:08:08.207 17:29:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:08.207 17:29:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:08.207 17:29:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:08.207 17:29:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:08.207 17:29:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:08.207 17:29:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:08.207 17:29:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:08.207 17:29:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:08.466 17:29:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:08.466 17:29:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:08.466 "name": "Existed_Raid", 00:08:08.466 "uuid": "e5b41d73-0627-499e-bfd3-2e2e7e8b1e86", 00:08:08.466 "strip_size_kb": 0, 00:08:08.466 "state": "online", 00:08:08.466 "raid_level": "raid1", 00:08:08.466 "superblock": false, 00:08:08.466 "num_base_bdevs": 2, 00:08:08.466 "num_base_bdevs_discovered": 1, 00:08:08.466 "num_base_bdevs_operational": 1, 00:08:08.466 "base_bdevs_list": [ 00:08:08.466 { 00:08:08.466 "name": null, 00:08:08.466 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:08.466 "is_configured": false, 00:08:08.466 "data_offset": 0, 00:08:08.466 "data_size": 65536 00:08:08.466 }, 00:08:08.466 { 00:08:08.466 "name": "BaseBdev2", 00:08:08.466 "uuid": "97e61142-d752-4c8a-b209-5bbf75ab1672", 00:08:08.466 "is_configured": true, 00:08:08.466 "data_offset": 0, 00:08:08.466 "data_size": 65536 00:08:08.466 } 00:08:08.466 ] 00:08:08.466 }' 00:08:08.466 17:29:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:08.466 17:29:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:08.726 17:29:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:08:08.726 17:29:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:08.726 17:29:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:08.726 17:29:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:08.726 17:29:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:08.726 17:29:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:08:08.726 17:29:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:08.726 17:29:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:08:08.726 17:29:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:08:08.726 17:29:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:08:08.726 17:29:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:08.726 17:29:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:08.726 [2024-11-27 17:29:39.887941] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:08:08.726 [2024-11-27 17:29:39.888126] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:08.726 [2024-11-27 17:29:39.909081] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:08.726 [2024-11-27 17:29:39.909231] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:08.726 [2024-11-27 17:29:39.909275] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:08:08.726 17:29:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:08.726 17:29:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:08:08.726 17:29:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:08.985 17:29:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:08.985 17:29:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:08:08.985 17:29:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:08.985 17:29:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:08.985 17:29:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:08.985 17:29:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:08:08.985 17:29:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:08:08.985 17:29:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:08:08.985 17:29:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 73786 00:08:08.985 17:29:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@950 -- # '[' -z 73786 ']' 00:08:08.985 17:29:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # kill -0 73786 00:08:08.985 17:29:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@955 -- # uname 00:08:08.985 17:29:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:08:08.985 17:29:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 73786 00:08:08.985 17:29:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:08:08.985 17:29:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:08:08.985 killing process with pid 73786 00:08:08.985 17:29:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 73786' 00:08:08.985 17:29:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@969 -- # kill 73786 00:08:08.985 [2024-11-27 17:29:40.008352] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:08.985 17:29:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@974 -- # wait 73786 00:08:08.985 [2024-11-27 17:29:40.009949] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:09.244 17:29:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:08:09.244 00:08:09.244 real 0m4.115s 00:08:09.244 user 0m6.267s 00:08:09.244 sys 0m0.911s 00:08:09.244 17:29:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:08:09.244 17:29:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:09.244 ************************************ 00:08:09.244 END TEST raid_state_function_test 00:08:09.244 ************************************ 00:08:09.244 17:29:40 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test raid1 2 true 00:08:09.244 17:29:40 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:08:09.244 17:29:40 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:08:09.244 17:29:40 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:09.503 ************************************ 00:08:09.503 START TEST raid_state_function_test_sb 00:08:09.503 ************************************ 00:08:09.503 17:29:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1125 -- # raid_state_function_test raid1 2 true 00:08:09.503 17:29:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:08:09.503 17:29:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:08:09.503 17:29:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:08:09.503 17:29:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:08:09.503 17:29:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:08:09.503 17:29:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:09.503 17:29:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:08:09.503 17:29:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:09.503 17:29:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:09.503 17:29:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:08:09.503 17:29:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:09.503 17:29:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:09.503 17:29:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:08:09.503 17:29:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:08:09.503 17:29:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:08:09.503 17:29:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:08:09.503 17:29:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:08:09.503 17:29:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:08:09.503 17:29:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:08:09.503 17:29:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:08:09.503 17:29:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:08:09.503 17:29:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:08:09.503 Process raid pid: 74028 00:08:09.503 17:29:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=74028 00:08:09.503 17:29:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:08:09.503 17:29:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 74028' 00:08:09.503 17:29:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 74028 00:08:09.503 17:29:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@831 -- # '[' -z 74028 ']' 00:08:09.503 17:29:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:09.503 17:29:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@836 -- # local max_retries=100 00:08:09.503 17:29:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:09.503 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:09.503 17:29:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # xtrace_disable 00:08:09.503 17:29:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:09.503 [2024-11-27 17:29:40.545013] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:08:09.503 [2024-11-27 17:29:40.545217] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:08:09.503 [2024-11-27 17:29:40.687155] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:09.761 [2024-11-27 17:29:40.754448] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:08:09.761 [2024-11-27 17:29:40.830112] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:09.761 [2024-11-27 17:29:40.830280] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:10.329 17:29:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:08:10.329 17:29:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@864 -- # return 0 00:08:10.329 17:29:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:08:10.329 17:29:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:10.329 17:29:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:10.329 [2024-11-27 17:29:41.363017] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:10.329 [2024-11-27 17:29:41.363072] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:10.329 [2024-11-27 17:29:41.363097] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:10.329 [2024-11-27 17:29:41.363108] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:10.329 17:29:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:10.329 17:29:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:08:10.329 17:29:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:10.329 17:29:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:10.329 17:29:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:10.329 17:29:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:10.329 17:29:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:10.329 17:29:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:10.329 17:29:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:10.329 17:29:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:10.329 17:29:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:10.329 17:29:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:10.329 17:29:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:10.329 17:29:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:10.329 17:29:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:10.329 17:29:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:10.329 17:29:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:10.329 "name": "Existed_Raid", 00:08:10.329 "uuid": "c332cb67-3369-4c98-88e3-b50d57867914", 00:08:10.329 "strip_size_kb": 0, 00:08:10.329 "state": "configuring", 00:08:10.329 "raid_level": "raid1", 00:08:10.329 "superblock": true, 00:08:10.329 "num_base_bdevs": 2, 00:08:10.329 "num_base_bdevs_discovered": 0, 00:08:10.329 "num_base_bdevs_operational": 2, 00:08:10.329 "base_bdevs_list": [ 00:08:10.329 { 00:08:10.329 "name": "BaseBdev1", 00:08:10.329 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:10.329 "is_configured": false, 00:08:10.329 "data_offset": 0, 00:08:10.329 "data_size": 0 00:08:10.329 }, 00:08:10.329 { 00:08:10.329 "name": "BaseBdev2", 00:08:10.329 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:10.329 "is_configured": false, 00:08:10.329 "data_offset": 0, 00:08:10.329 "data_size": 0 00:08:10.329 } 00:08:10.329 ] 00:08:10.329 }' 00:08:10.329 17:29:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:10.329 17:29:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:10.899 17:29:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:10.899 17:29:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:10.899 17:29:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:10.899 [2024-11-27 17:29:41.818236] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:10.899 [2024-11-27 17:29:41.818353] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:08:10.899 17:29:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:10.899 17:29:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:08:10.899 17:29:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:10.899 17:29:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:10.899 [2024-11-27 17:29:41.830231] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:10.899 [2024-11-27 17:29:41.830319] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:10.899 [2024-11-27 17:29:41.830360] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:10.899 [2024-11-27 17:29:41.830383] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:10.899 17:29:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:10.899 17:29:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:08:10.899 17:29:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:10.899 17:29:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:10.899 [2024-11-27 17:29:41.857179] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:10.899 BaseBdev1 00:08:10.899 17:29:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:10.899 17:29:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:08:10.899 17:29:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:08:10.899 17:29:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:08:10.899 17:29:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:08:10.899 17:29:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:08:10.899 17:29:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:08:10.899 17:29:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:08:10.899 17:29:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:10.899 17:29:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:10.899 17:29:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:10.899 17:29:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:08:10.899 17:29:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:10.899 17:29:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:10.899 [ 00:08:10.899 { 00:08:10.899 "name": "BaseBdev1", 00:08:10.899 "aliases": [ 00:08:10.899 "9569dab9-30c2-43ff-8733-abc1b127ec75" 00:08:10.899 ], 00:08:10.899 "product_name": "Malloc disk", 00:08:10.899 "block_size": 512, 00:08:10.899 "num_blocks": 65536, 00:08:10.899 "uuid": "9569dab9-30c2-43ff-8733-abc1b127ec75", 00:08:10.899 "assigned_rate_limits": { 00:08:10.899 "rw_ios_per_sec": 0, 00:08:10.899 "rw_mbytes_per_sec": 0, 00:08:10.899 "r_mbytes_per_sec": 0, 00:08:10.899 "w_mbytes_per_sec": 0 00:08:10.899 }, 00:08:10.899 "claimed": true, 00:08:10.899 "claim_type": "exclusive_write", 00:08:10.899 "zoned": false, 00:08:10.899 "supported_io_types": { 00:08:10.899 "read": true, 00:08:10.899 "write": true, 00:08:10.899 "unmap": true, 00:08:10.899 "flush": true, 00:08:10.899 "reset": true, 00:08:10.899 "nvme_admin": false, 00:08:10.899 "nvme_io": false, 00:08:10.899 "nvme_io_md": false, 00:08:10.899 "write_zeroes": true, 00:08:10.899 "zcopy": true, 00:08:10.899 "get_zone_info": false, 00:08:10.899 "zone_management": false, 00:08:10.899 "zone_append": false, 00:08:10.899 "compare": false, 00:08:10.899 "compare_and_write": false, 00:08:10.899 "abort": true, 00:08:10.899 "seek_hole": false, 00:08:10.899 "seek_data": false, 00:08:10.899 "copy": true, 00:08:10.899 "nvme_iov_md": false 00:08:10.899 }, 00:08:10.899 "memory_domains": [ 00:08:10.899 { 00:08:10.899 "dma_device_id": "system", 00:08:10.899 "dma_device_type": 1 00:08:10.899 }, 00:08:10.899 { 00:08:10.899 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:10.899 "dma_device_type": 2 00:08:10.899 } 00:08:10.899 ], 00:08:10.899 "driver_specific": {} 00:08:10.899 } 00:08:10.899 ] 00:08:10.899 17:29:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:10.899 17:29:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:08:10.899 17:29:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:08:10.899 17:29:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:10.899 17:29:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:10.899 17:29:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:10.899 17:29:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:10.899 17:29:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:10.899 17:29:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:10.899 17:29:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:10.899 17:29:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:10.899 17:29:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:10.899 17:29:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:10.899 17:29:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:10.899 17:29:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:10.899 17:29:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:10.899 17:29:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:10.899 17:29:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:10.899 "name": "Existed_Raid", 00:08:10.899 "uuid": "56c05dfb-d89e-43d7-9d14-acd5422d4e46", 00:08:10.899 "strip_size_kb": 0, 00:08:10.899 "state": "configuring", 00:08:10.899 "raid_level": "raid1", 00:08:10.899 "superblock": true, 00:08:10.899 "num_base_bdevs": 2, 00:08:10.899 "num_base_bdevs_discovered": 1, 00:08:10.899 "num_base_bdevs_operational": 2, 00:08:10.899 "base_bdevs_list": [ 00:08:10.899 { 00:08:10.899 "name": "BaseBdev1", 00:08:10.899 "uuid": "9569dab9-30c2-43ff-8733-abc1b127ec75", 00:08:10.899 "is_configured": true, 00:08:10.899 "data_offset": 2048, 00:08:10.899 "data_size": 63488 00:08:10.899 }, 00:08:10.899 { 00:08:10.899 "name": "BaseBdev2", 00:08:10.899 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:10.899 "is_configured": false, 00:08:10.899 "data_offset": 0, 00:08:10.899 "data_size": 0 00:08:10.899 } 00:08:10.899 ] 00:08:10.899 }' 00:08:10.899 17:29:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:10.899 17:29:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:11.160 17:29:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:11.160 17:29:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:11.160 17:29:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:11.160 [2024-11-27 17:29:42.292428] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:11.160 [2024-11-27 17:29:42.292475] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:08:11.160 17:29:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:11.160 17:29:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:08:11.160 17:29:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:11.160 17:29:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:11.160 [2024-11-27 17:29:42.304462] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:11.160 [2024-11-27 17:29:42.306617] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:11.160 [2024-11-27 17:29:42.306698] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:11.160 17:29:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:11.160 17:29:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:08:11.160 17:29:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:11.160 17:29:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:08:11.160 17:29:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:11.160 17:29:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:11.160 17:29:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:11.160 17:29:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:11.160 17:29:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:11.160 17:29:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:11.160 17:29:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:11.160 17:29:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:11.160 17:29:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:11.160 17:29:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:11.160 17:29:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:11.160 17:29:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:11.160 17:29:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:11.160 17:29:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:11.420 17:29:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:11.420 "name": "Existed_Raid", 00:08:11.420 "uuid": "306477b4-0e7b-4a85-bc4c-cd3aadf226bf", 00:08:11.420 "strip_size_kb": 0, 00:08:11.420 "state": "configuring", 00:08:11.420 "raid_level": "raid1", 00:08:11.420 "superblock": true, 00:08:11.420 "num_base_bdevs": 2, 00:08:11.420 "num_base_bdevs_discovered": 1, 00:08:11.420 "num_base_bdevs_operational": 2, 00:08:11.420 "base_bdevs_list": [ 00:08:11.420 { 00:08:11.420 "name": "BaseBdev1", 00:08:11.420 "uuid": "9569dab9-30c2-43ff-8733-abc1b127ec75", 00:08:11.420 "is_configured": true, 00:08:11.420 "data_offset": 2048, 00:08:11.420 "data_size": 63488 00:08:11.420 }, 00:08:11.420 { 00:08:11.420 "name": "BaseBdev2", 00:08:11.420 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:11.420 "is_configured": false, 00:08:11.420 "data_offset": 0, 00:08:11.420 "data_size": 0 00:08:11.420 } 00:08:11.420 ] 00:08:11.420 }' 00:08:11.420 17:29:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:11.420 17:29:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:11.681 17:29:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:08:11.681 17:29:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:11.681 17:29:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:11.681 [2024-11-27 17:29:42.770101] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:11.681 [2024-11-27 17:29:42.770734] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:08:11.681 [2024-11-27 17:29:42.770897] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:08:11.681 BaseBdev2 00:08:11.681 [2024-11-27 17:29:42.771788] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:08:11.681 17:29:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:11.681 [2024-11-27 17:29:42.772317] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:08:11.681 [2024-11-27 17:29:42.772396] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:08:11.681 [2024-11-27 17:29:42.772772] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:11.681 17:29:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:08:11.681 17:29:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:08:11.681 17:29:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:08:11.681 17:29:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:08:11.681 17:29:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:08:11.681 17:29:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:08:11.681 17:29:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:08:11.681 17:29:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:11.681 17:29:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:11.681 17:29:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:11.681 17:29:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:08:11.681 17:29:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:11.681 17:29:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:11.681 [ 00:08:11.681 { 00:08:11.681 "name": "BaseBdev2", 00:08:11.681 "aliases": [ 00:08:11.681 "c9d9161c-9c9e-46e1-92b6-d1cab26f3def" 00:08:11.681 ], 00:08:11.681 "product_name": "Malloc disk", 00:08:11.681 "block_size": 512, 00:08:11.681 "num_blocks": 65536, 00:08:11.681 "uuid": "c9d9161c-9c9e-46e1-92b6-d1cab26f3def", 00:08:11.681 "assigned_rate_limits": { 00:08:11.681 "rw_ios_per_sec": 0, 00:08:11.681 "rw_mbytes_per_sec": 0, 00:08:11.681 "r_mbytes_per_sec": 0, 00:08:11.681 "w_mbytes_per_sec": 0 00:08:11.681 }, 00:08:11.681 "claimed": true, 00:08:11.681 "claim_type": "exclusive_write", 00:08:11.681 "zoned": false, 00:08:11.681 "supported_io_types": { 00:08:11.681 "read": true, 00:08:11.681 "write": true, 00:08:11.681 "unmap": true, 00:08:11.681 "flush": true, 00:08:11.681 "reset": true, 00:08:11.681 "nvme_admin": false, 00:08:11.681 "nvme_io": false, 00:08:11.681 "nvme_io_md": false, 00:08:11.681 "write_zeroes": true, 00:08:11.681 "zcopy": true, 00:08:11.681 "get_zone_info": false, 00:08:11.681 "zone_management": false, 00:08:11.681 "zone_append": false, 00:08:11.681 "compare": false, 00:08:11.681 "compare_and_write": false, 00:08:11.681 "abort": true, 00:08:11.681 "seek_hole": false, 00:08:11.681 "seek_data": false, 00:08:11.681 "copy": true, 00:08:11.681 "nvme_iov_md": false 00:08:11.681 }, 00:08:11.681 "memory_domains": [ 00:08:11.681 { 00:08:11.681 "dma_device_id": "system", 00:08:11.681 "dma_device_type": 1 00:08:11.681 }, 00:08:11.681 { 00:08:11.681 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:11.681 "dma_device_type": 2 00:08:11.681 } 00:08:11.681 ], 00:08:11.681 "driver_specific": {} 00:08:11.681 } 00:08:11.681 ] 00:08:11.681 17:29:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:11.681 17:29:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:08:11.681 17:29:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:08:11.681 17:29:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:11.681 17:29:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 2 00:08:11.681 17:29:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:11.681 17:29:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:11.681 17:29:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:11.681 17:29:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:11.681 17:29:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:11.681 17:29:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:11.681 17:29:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:11.681 17:29:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:11.681 17:29:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:11.681 17:29:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:11.681 17:29:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:11.681 17:29:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:11.681 17:29:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:11.681 17:29:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:11.681 17:29:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:11.681 "name": "Existed_Raid", 00:08:11.681 "uuid": "306477b4-0e7b-4a85-bc4c-cd3aadf226bf", 00:08:11.681 "strip_size_kb": 0, 00:08:11.681 "state": "online", 00:08:11.681 "raid_level": "raid1", 00:08:11.681 "superblock": true, 00:08:11.681 "num_base_bdevs": 2, 00:08:11.681 "num_base_bdevs_discovered": 2, 00:08:11.681 "num_base_bdevs_operational": 2, 00:08:11.681 "base_bdevs_list": [ 00:08:11.681 { 00:08:11.681 "name": "BaseBdev1", 00:08:11.681 "uuid": "9569dab9-30c2-43ff-8733-abc1b127ec75", 00:08:11.681 "is_configured": true, 00:08:11.681 "data_offset": 2048, 00:08:11.681 "data_size": 63488 00:08:11.681 }, 00:08:11.681 { 00:08:11.681 "name": "BaseBdev2", 00:08:11.681 "uuid": "c9d9161c-9c9e-46e1-92b6-d1cab26f3def", 00:08:11.681 "is_configured": true, 00:08:11.681 "data_offset": 2048, 00:08:11.681 "data_size": 63488 00:08:11.681 } 00:08:11.681 ] 00:08:11.681 }' 00:08:11.681 17:29:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:11.681 17:29:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:12.251 17:29:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:08:12.251 17:29:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:08:12.251 17:29:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:12.251 17:29:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:12.251 17:29:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:08:12.251 17:29:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:12.251 17:29:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:08:12.251 17:29:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:12.251 17:29:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:12.251 17:29:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:12.251 [2024-11-27 17:29:43.217565] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:12.251 17:29:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:12.251 17:29:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:12.251 "name": "Existed_Raid", 00:08:12.251 "aliases": [ 00:08:12.251 "306477b4-0e7b-4a85-bc4c-cd3aadf226bf" 00:08:12.251 ], 00:08:12.251 "product_name": "Raid Volume", 00:08:12.251 "block_size": 512, 00:08:12.251 "num_blocks": 63488, 00:08:12.251 "uuid": "306477b4-0e7b-4a85-bc4c-cd3aadf226bf", 00:08:12.251 "assigned_rate_limits": { 00:08:12.251 "rw_ios_per_sec": 0, 00:08:12.251 "rw_mbytes_per_sec": 0, 00:08:12.251 "r_mbytes_per_sec": 0, 00:08:12.251 "w_mbytes_per_sec": 0 00:08:12.251 }, 00:08:12.251 "claimed": false, 00:08:12.251 "zoned": false, 00:08:12.251 "supported_io_types": { 00:08:12.251 "read": true, 00:08:12.251 "write": true, 00:08:12.251 "unmap": false, 00:08:12.251 "flush": false, 00:08:12.251 "reset": true, 00:08:12.251 "nvme_admin": false, 00:08:12.251 "nvme_io": false, 00:08:12.251 "nvme_io_md": false, 00:08:12.251 "write_zeroes": true, 00:08:12.251 "zcopy": false, 00:08:12.251 "get_zone_info": false, 00:08:12.251 "zone_management": false, 00:08:12.251 "zone_append": false, 00:08:12.251 "compare": false, 00:08:12.251 "compare_and_write": false, 00:08:12.251 "abort": false, 00:08:12.251 "seek_hole": false, 00:08:12.251 "seek_data": false, 00:08:12.251 "copy": false, 00:08:12.251 "nvme_iov_md": false 00:08:12.251 }, 00:08:12.251 "memory_domains": [ 00:08:12.251 { 00:08:12.251 "dma_device_id": "system", 00:08:12.251 "dma_device_type": 1 00:08:12.251 }, 00:08:12.251 { 00:08:12.251 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:12.251 "dma_device_type": 2 00:08:12.251 }, 00:08:12.251 { 00:08:12.251 "dma_device_id": "system", 00:08:12.251 "dma_device_type": 1 00:08:12.251 }, 00:08:12.251 { 00:08:12.251 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:12.251 "dma_device_type": 2 00:08:12.251 } 00:08:12.251 ], 00:08:12.251 "driver_specific": { 00:08:12.251 "raid": { 00:08:12.251 "uuid": "306477b4-0e7b-4a85-bc4c-cd3aadf226bf", 00:08:12.251 "strip_size_kb": 0, 00:08:12.251 "state": "online", 00:08:12.251 "raid_level": "raid1", 00:08:12.251 "superblock": true, 00:08:12.251 "num_base_bdevs": 2, 00:08:12.251 "num_base_bdevs_discovered": 2, 00:08:12.251 "num_base_bdevs_operational": 2, 00:08:12.251 "base_bdevs_list": [ 00:08:12.251 { 00:08:12.251 "name": "BaseBdev1", 00:08:12.251 "uuid": "9569dab9-30c2-43ff-8733-abc1b127ec75", 00:08:12.251 "is_configured": true, 00:08:12.251 "data_offset": 2048, 00:08:12.251 "data_size": 63488 00:08:12.251 }, 00:08:12.251 { 00:08:12.251 "name": "BaseBdev2", 00:08:12.251 "uuid": "c9d9161c-9c9e-46e1-92b6-d1cab26f3def", 00:08:12.251 "is_configured": true, 00:08:12.251 "data_offset": 2048, 00:08:12.251 "data_size": 63488 00:08:12.251 } 00:08:12.251 ] 00:08:12.251 } 00:08:12.251 } 00:08:12.251 }' 00:08:12.251 17:29:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:12.251 17:29:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:08:12.251 BaseBdev2' 00:08:12.251 17:29:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:12.251 17:29:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:12.251 17:29:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:12.251 17:29:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:12.251 17:29:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:08:12.251 17:29:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:12.251 17:29:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:12.251 17:29:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:12.251 17:29:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:12.251 17:29:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:12.251 17:29:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:12.251 17:29:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:12.251 17:29:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:08:12.251 17:29:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:12.251 17:29:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:12.251 17:29:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:12.512 17:29:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:12.512 17:29:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:12.512 17:29:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:08:12.512 17:29:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:12.512 17:29:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:12.512 [2024-11-27 17:29:43.452960] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:08:12.512 17:29:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:12.512 17:29:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:08:12.512 17:29:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:08:12.512 17:29:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:08:12.512 17:29:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@199 -- # return 0 00:08:12.512 17:29:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:08:12.512 17:29:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 1 00:08:12.512 17:29:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:12.512 17:29:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:12.512 17:29:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:12.512 17:29:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:12.512 17:29:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:08:12.512 17:29:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:12.512 17:29:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:12.512 17:29:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:12.512 17:29:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:12.512 17:29:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:12.512 17:29:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:12.512 17:29:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:12.512 17:29:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:12.512 17:29:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:12.512 17:29:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:12.512 "name": "Existed_Raid", 00:08:12.512 "uuid": "306477b4-0e7b-4a85-bc4c-cd3aadf226bf", 00:08:12.512 "strip_size_kb": 0, 00:08:12.512 "state": "online", 00:08:12.512 "raid_level": "raid1", 00:08:12.512 "superblock": true, 00:08:12.512 "num_base_bdevs": 2, 00:08:12.512 "num_base_bdevs_discovered": 1, 00:08:12.512 "num_base_bdevs_operational": 1, 00:08:12.512 "base_bdevs_list": [ 00:08:12.512 { 00:08:12.512 "name": null, 00:08:12.512 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:12.512 "is_configured": false, 00:08:12.512 "data_offset": 0, 00:08:12.512 "data_size": 63488 00:08:12.512 }, 00:08:12.512 { 00:08:12.512 "name": "BaseBdev2", 00:08:12.512 "uuid": "c9d9161c-9c9e-46e1-92b6-d1cab26f3def", 00:08:12.512 "is_configured": true, 00:08:12.512 "data_offset": 2048, 00:08:12.512 "data_size": 63488 00:08:12.512 } 00:08:12.512 ] 00:08:12.512 }' 00:08:12.512 17:29:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:12.512 17:29:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:12.772 17:29:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:08:12.772 17:29:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:12.772 17:29:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:12.772 17:29:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:08:12.772 17:29:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:12.772 17:29:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:12.772 17:29:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:12.772 17:29:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:08:12.772 17:29:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:08:12.772 17:29:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:08:12.772 17:29:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:12.772 17:29:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:12.772 [2024-11-27 17:29:43.956722] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:08:12.772 [2024-11-27 17:29:43.956825] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:13.032 [2024-11-27 17:29:43.977925] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:13.032 [2024-11-27 17:29:43.978046] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:13.032 [2024-11-27 17:29:43.978093] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:08:13.032 17:29:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:13.032 17:29:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:08:13.032 17:29:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:13.032 17:29:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:13.032 17:29:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:13.032 17:29:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:13.032 17:29:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:08:13.032 17:29:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:13.032 17:29:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:08:13.032 17:29:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:08:13.032 17:29:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:08:13.032 17:29:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 74028 00:08:13.032 17:29:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@950 -- # '[' -z 74028 ']' 00:08:13.032 17:29:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # kill -0 74028 00:08:13.032 17:29:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@955 -- # uname 00:08:13.032 17:29:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:08:13.032 17:29:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 74028 00:08:13.032 17:29:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:08:13.032 17:29:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:08:13.032 17:29:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@968 -- # echo 'killing process with pid 74028' 00:08:13.032 killing process with pid 74028 00:08:13.032 17:29:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@969 -- # kill 74028 00:08:13.032 [2024-11-27 17:29:44.067975] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:13.032 17:29:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@974 -- # wait 74028 00:08:13.032 [2024-11-27 17:29:44.069576] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:13.291 17:29:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:08:13.291 00:08:13.291 real 0m3.989s 00:08:13.291 user 0m6.090s 00:08:13.291 sys 0m0.855s 00:08:13.291 17:29:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1126 -- # xtrace_disable 00:08:13.291 ************************************ 00:08:13.291 END TEST raid_state_function_test_sb 00:08:13.291 ************************************ 00:08:13.291 17:29:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:13.551 17:29:44 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test raid1 2 00:08:13.551 17:29:44 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:08:13.551 17:29:44 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:08:13.551 17:29:44 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:13.551 ************************************ 00:08:13.551 START TEST raid_superblock_test 00:08:13.551 ************************************ 00:08:13.551 17:29:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1125 -- # raid_superblock_test raid1 2 00:08:13.551 17:29:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=raid1 00:08:13.551 17:29:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=2 00:08:13.551 17:29:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:08:13.551 17:29:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:08:13.551 17:29:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:08:13.551 17:29:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:08:13.551 17:29:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:08:13.551 17:29:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:08:13.551 17:29:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:08:13.551 17:29:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:08:13.551 17:29:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:08:13.551 17:29:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:08:13.551 17:29:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:08:13.551 17:29:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' raid1 '!=' raid1 ']' 00:08:13.551 17:29:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@408 -- # strip_size=0 00:08:13.551 17:29:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=74269 00:08:13.551 17:29:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:08:13.551 17:29:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 74269 00:08:13.551 17:29:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@831 -- # '[' -z 74269 ']' 00:08:13.551 17:29:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:13.551 17:29:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:08:13.551 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:13.551 17:29:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:13.551 17:29:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:08:13.551 17:29:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:13.551 [2024-11-27 17:29:44.600376] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:08:13.551 [2024-11-27 17:29:44.600519] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid74269 ] 00:08:13.551 [2024-11-27 17:29:44.727449] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:13.810 [2024-11-27 17:29:44.797214] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:08:13.810 [2024-11-27 17:29:44.875343] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:13.810 [2024-11-27 17:29:44.875390] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:14.379 17:29:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:08:14.379 17:29:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@864 -- # return 0 00:08:14.379 17:29:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:08:14.379 17:29:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:08:14.379 17:29:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:08:14.379 17:29:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:08:14.379 17:29:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:08:14.379 17:29:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:08:14.379 17:29:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:08:14.379 17:29:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:08:14.379 17:29:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:08:14.379 17:29:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:14.379 17:29:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:14.379 malloc1 00:08:14.379 17:29:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:14.379 17:29:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:08:14.379 17:29:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:14.379 17:29:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:14.379 [2024-11-27 17:29:45.454602] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:08:14.379 [2024-11-27 17:29:45.454728] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:14.379 [2024-11-27 17:29:45.454777] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:08:14.379 [2024-11-27 17:29:45.454840] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:14.379 [2024-11-27 17:29:45.457225] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:14.379 [2024-11-27 17:29:45.457291] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:08:14.379 pt1 00:08:14.379 17:29:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:14.379 17:29:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:08:14.379 17:29:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:08:14.379 17:29:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:08:14.379 17:29:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:08:14.379 17:29:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:08:14.379 17:29:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:08:14.379 17:29:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:08:14.379 17:29:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:08:14.379 17:29:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:08:14.379 17:29:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:14.379 17:29:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:14.379 malloc2 00:08:14.379 17:29:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:14.379 17:29:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:08:14.379 17:29:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:14.379 17:29:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:14.379 [2024-11-27 17:29:45.502264] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:08:14.379 [2024-11-27 17:29:45.502448] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:14.379 [2024-11-27 17:29:45.502493] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:08:14.379 [2024-11-27 17:29:45.502519] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:14.380 [2024-11-27 17:29:45.507274] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:14.380 pt2 00:08:14.380 [2024-11-27 17:29:45.507417] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:08:14.380 17:29:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:14.380 17:29:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:08:14.380 17:29:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:08:14.380 17:29:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''pt1 pt2'\''' -n raid_bdev1 -s 00:08:14.380 17:29:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:14.380 17:29:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:14.380 [2024-11-27 17:29:45.515702] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:08:14.380 [2024-11-27 17:29:45.518739] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:08:14.380 [2024-11-27 17:29:45.518997] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:08:14.380 [2024-11-27 17:29:45.519017] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:08:14.380 [2024-11-27 17:29:45.519317] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:08:14.380 [2024-11-27 17:29:45.519475] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:08:14.380 [2024-11-27 17:29:45.519493] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:08:14.380 [2024-11-27 17:29:45.519668] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:14.380 17:29:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:14.380 17:29:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:08:14.380 17:29:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:14.380 17:29:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:14.380 17:29:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:14.380 17:29:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:14.380 17:29:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:14.380 17:29:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:14.380 17:29:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:14.380 17:29:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:14.380 17:29:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:14.380 17:29:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:14.380 17:29:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:14.380 17:29:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:14.380 17:29:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:14.380 17:29:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:14.639 17:29:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:14.639 "name": "raid_bdev1", 00:08:14.639 "uuid": "a967b97a-4454-4f98-a55d-5b6e90f97dd7", 00:08:14.639 "strip_size_kb": 0, 00:08:14.639 "state": "online", 00:08:14.639 "raid_level": "raid1", 00:08:14.639 "superblock": true, 00:08:14.639 "num_base_bdevs": 2, 00:08:14.639 "num_base_bdevs_discovered": 2, 00:08:14.639 "num_base_bdevs_operational": 2, 00:08:14.639 "base_bdevs_list": [ 00:08:14.639 { 00:08:14.639 "name": "pt1", 00:08:14.639 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:14.639 "is_configured": true, 00:08:14.639 "data_offset": 2048, 00:08:14.639 "data_size": 63488 00:08:14.639 }, 00:08:14.639 { 00:08:14.639 "name": "pt2", 00:08:14.639 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:14.639 "is_configured": true, 00:08:14.639 "data_offset": 2048, 00:08:14.639 "data_size": 63488 00:08:14.639 } 00:08:14.639 ] 00:08:14.639 }' 00:08:14.639 17:29:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:14.639 17:29:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:14.899 17:29:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:08:14.899 17:29:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:08:14.899 17:29:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:14.899 17:29:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:14.899 17:29:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:08:14.899 17:29:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:14.899 17:29:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:08:14.899 17:29:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:14.899 17:29:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:14.899 17:29:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:14.899 [2024-11-27 17:29:45.979250] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:14.899 17:29:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:14.899 17:29:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:14.899 "name": "raid_bdev1", 00:08:14.899 "aliases": [ 00:08:14.899 "a967b97a-4454-4f98-a55d-5b6e90f97dd7" 00:08:14.899 ], 00:08:14.899 "product_name": "Raid Volume", 00:08:14.899 "block_size": 512, 00:08:14.899 "num_blocks": 63488, 00:08:14.899 "uuid": "a967b97a-4454-4f98-a55d-5b6e90f97dd7", 00:08:14.899 "assigned_rate_limits": { 00:08:14.899 "rw_ios_per_sec": 0, 00:08:14.899 "rw_mbytes_per_sec": 0, 00:08:14.899 "r_mbytes_per_sec": 0, 00:08:14.899 "w_mbytes_per_sec": 0 00:08:14.899 }, 00:08:14.899 "claimed": false, 00:08:14.899 "zoned": false, 00:08:14.899 "supported_io_types": { 00:08:14.899 "read": true, 00:08:14.899 "write": true, 00:08:14.899 "unmap": false, 00:08:14.899 "flush": false, 00:08:14.899 "reset": true, 00:08:14.899 "nvme_admin": false, 00:08:14.899 "nvme_io": false, 00:08:14.899 "nvme_io_md": false, 00:08:14.899 "write_zeroes": true, 00:08:14.899 "zcopy": false, 00:08:14.899 "get_zone_info": false, 00:08:14.899 "zone_management": false, 00:08:14.899 "zone_append": false, 00:08:14.899 "compare": false, 00:08:14.899 "compare_and_write": false, 00:08:14.899 "abort": false, 00:08:14.899 "seek_hole": false, 00:08:14.899 "seek_data": false, 00:08:14.899 "copy": false, 00:08:14.899 "nvme_iov_md": false 00:08:14.899 }, 00:08:14.899 "memory_domains": [ 00:08:14.899 { 00:08:14.899 "dma_device_id": "system", 00:08:14.899 "dma_device_type": 1 00:08:14.899 }, 00:08:14.899 { 00:08:14.899 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:14.899 "dma_device_type": 2 00:08:14.899 }, 00:08:14.899 { 00:08:14.899 "dma_device_id": "system", 00:08:14.899 "dma_device_type": 1 00:08:14.899 }, 00:08:14.899 { 00:08:14.900 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:14.900 "dma_device_type": 2 00:08:14.900 } 00:08:14.900 ], 00:08:14.900 "driver_specific": { 00:08:14.900 "raid": { 00:08:14.900 "uuid": "a967b97a-4454-4f98-a55d-5b6e90f97dd7", 00:08:14.900 "strip_size_kb": 0, 00:08:14.900 "state": "online", 00:08:14.900 "raid_level": "raid1", 00:08:14.900 "superblock": true, 00:08:14.900 "num_base_bdevs": 2, 00:08:14.900 "num_base_bdevs_discovered": 2, 00:08:14.900 "num_base_bdevs_operational": 2, 00:08:14.900 "base_bdevs_list": [ 00:08:14.900 { 00:08:14.900 "name": "pt1", 00:08:14.900 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:14.900 "is_configured": true, 00:08:14.900 "data_offset": 2048, 00:08:14.900 "data_size": 63488 00:08:14.900 }, 00:08:14.900 { 00:08:14.900 "name": "pt2", 00:08:14.900 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:14.900 "is_configured": true, 00:08:14.900 "data_offset": 2048, 00:08:14.900 "data_size": 63488 00:08:14.900 } 00:08:14.900 ] 00:08:14.900 } 00:08:14.900 } 00:08:14.900 }' 00:08:14.900 17:29:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:14.900 17:29:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:08:14.900 pt2' 00:08:14.900 17:29:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:15.160 17:29:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:15.160 17:29:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:15.160 17:29:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:15.160 17:29:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:08:15.160 17:29:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:15.160 17:29:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:15.160 17:29:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:15.160 17:29:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:15.160 17:29:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:15.160 17:29:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:15.160 17:29:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:08:15.160 17:29:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:15.160 17:29:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:15.160 17:29:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:15.160 17:29:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:15.160 17:29:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:15.160 17:29:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:15.160 17:29:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:08:15.160 17:29:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:08:15.160 17:29:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:15.160 17:29:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:15.160 [2024-11-27 17:29:46.238819] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:15.160 17:29:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:15.160 17:29:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=a967b97a-4454-4f98-a55d-5b6e90f97dd7 00:08:15.160 17:29:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z a967b97a-4454-4f98-a55d-5b6e90f97dd7 ']' 00:08:15.160 17:29:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:08:15.160 17:29:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:15.160 17:29:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:15.160 [2024-11-27 17:29:46.282500] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:08:15.160 [2024-11-27 17:29:46.282525] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:15.160 [2024-11-27 17:29:46.282588] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:15.160 [2024-11-27 17:29:46.282645] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:15.160 [2024-11-27 17:29:46.282654] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:08:15.160 17:29:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:15.160 17:29:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:08:15.160 17:29:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:15.160 17:29:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:15.160 17:29:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:15.160 17:29:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:15.160 17:29:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:08:15.160 17:29:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:08:15.160 17:29:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:08:15.160 17:29:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:08:15.160 17:29:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:15.160 17:29:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:15.420 17:29:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:15.420 17:29:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:08:15.420 17:29:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:08:15.420 17:29:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:15.420 17:29:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:15.420 17:29:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:15.420 17:29:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:08:15.420 17:29:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:08:15.420 17:29:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:15.420 17:29:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:15.420 17:29:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:15.420 17:29:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:08:15.420 17:29:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:08:15.420 17:29:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@650 -- # local es=0 00:08:15.420 17:29:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:08:15.420 17:29:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:08:15.420 17:29:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:08:15.420 17:29:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:08:15.420 17:29:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:08:15.420 17:29:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:08:15.420 17:29:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:15.420 17:29:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:15.420 [2024-11-27 17:29:46.426266] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:08:15.420 [2024-11-27 17:29:46.428446] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:08:15.421 [2024-11-27 17:29:46.428560] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:08:15.421 [2024-11-27 17:29:46.428638] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:08:15.421 [2024-11-27 17:29:46.428702] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:08:15.421 [2024-11-27 17:29:46.428731] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state configuring 00:08:15.421 request: 00:08:15.421 { 00:08:15.421 "name": "raid_bdev1", 00:08:15.421 "raid_level": "raid1", 00:08:15.421 "base_bdevs": [ 00:08:15.421 "malloc1", 00:08:15.421 "malloc2" 00:08:15.421 ], 00:08:15.421 "superblock": false, 00:08:15.421 "method": "bdev_raid_create", 00:08:15.421 "req_id": 1 00:08:15.421 } 00:08:15.421 Got JSON-RPC error response 00:08:15.421 response: 00:08:15.421 { 00:08:15.421 "code": -17, 00:08:15.421 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:08:15.421 } 00:08:15.421 17:29:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:08:15.421 17:29:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@653 -- # es=1 00:08:15.421 17:29:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:08:15.421 17:29:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:08:15.421 17:29:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:08:15.421 17:29:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:15.421 17:29:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:15.421 17:29:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:08:15.421 17:29:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:15.421 17:29:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:15.421 17:29:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:08:15.421 17:29:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:08:15.421 17:29:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:08:15.421 17:29:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:15.421 17:29:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:15.421 [2024-11-27 17:29:46.490165] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:08:15.421 [2024-11-27 17:29:46.490270] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:15.421 [2024-11-27 17:29:46.490310] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:08:15.421 [2024-11-27 17:29:46.490343] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:15.421 [2024-11-27 17:29:46.492786] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:15.421 [2024-11-27 17:29:46.492850] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:08:15.421 [2024-11-27 17:29:46.492953] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:08:15.421 [2024-11-27 17:29:46.493031] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:08:15.421 pt1 00:08:15.421 17:29:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:15.421 17:29:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 2 00:08:15.421 17:29:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:15.421 17:29:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:15.421 17:29:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:15.421 17:29:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:15.421 17:29:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:15.421 17:29:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:15.421 17:29:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:15.421 17:29:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:15.421 17:29:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:15.421 17:29:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:15.421 17:29:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:15.421 17:29:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:15.421 17:29:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:15.421 17:29:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:15.421 17:29:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:15.421 "name": "raid_bdev1", 00:08:15.421 "uuid": "a967b97a-4454-4f98-a55d-5b6e90f97dd7", 00:08:15.421 "strip_size_kb": 0, 00:08:15.421 "state": "configuring", 00:08:15.421 "raid_level": "raid1", 00:08:15.421 "superblock": true, 00:08:15.421 "num_base_bdevs": 2, 00:08:15.421 "num_base_bdevs_discovered": 1, 00:08:15.421 "num_base_bdevs_operational": 2, 00:08:15.421 "base_bdevs_list": [ 00:08:15.421 { 00:08:15.421 "name": "pt1", 00:08:15.421 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:15.421 "is_configured": true, 00:08:15.421 "data_offset": 2048, 00:08:15.421 "data_size": 63488 00:08:15.421 }, 00:08:15.421 { 00:08:15.421 "name": null, 00:08:15.421 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:15.421 "is_configured": false, 00:08:15.421 "data_offset": 2048, 00:08:15.421 "data_size": 63488 00:08:15.421 } 00:08:15.421 ] 00:08:15.421 }' 00:08:15.421 17:29:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:15.421 17:29:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:15.991 17:29:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 2 -gt 2 ']' 00:08:15.991 17:29:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:08:15.991 17:29:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:08:15.991 17:29:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:08:15.991 17:29:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:15.991 17:29:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:15.991 [2024-11-27 17:29:46.961347] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:08:15.991 [2024-11-27 17:29:46.961394] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:15.991 [2024-11-27 17:29:46.961414] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:08:15.991 [2024-11-27 17:29:46.961422] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:15.991 [2024-11-27 17:29:46.961804] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:15.991 [2024-11-27 17:29:46.961819] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:08:15.991 [2024-11-27 17:29:46.961879] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:08:15.991 [2024-11-27 17:29:46.961898] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:08:15.991 [2024-11-27 17:29:46.961998] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:08:15.991 [2024-11-27 17:29:46.962007] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:08:15.991 [2024-11-27 17:29:46.962270] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:08:15.991 [2024-11-27 17:29:46.962394] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:08:15.991 [2024-11-27 17:29:46.962409] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:08:15.991 [2024-11-27 17:29:46.962507] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:15.991 pt2 00:08:15.991 17:29:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:15.991 17:29:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:08:15.991 17:29:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:08:15.991 17:29:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:08:15.991 17:29:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:15.991 17:29:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:15.991 17:29:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:15.991 17:29:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:15.991 17:29:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:15.991 17:29:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:15.991 17:29:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:15.991 17:29:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:15.991 17:29:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:15.991 17:29:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:15.991 17:29:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:15.991 17:29:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:15.991 17:29:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:15.991 17:29:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:15.991 17:29:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:15.991 "name": "raid_bdev1", 00:08:15.991 "uuid": "a967b97a-4454-4f98-a55d-5b6e90f97dd7", 00:08:15.991 "strip_size_kb": 0, 00:08:15.991 "state": "online", 00:08:15.991 "raid_level": "raid1", 00:08:15.991 "superblock": true, 00:08:15.991 "num_base_bdevs": 2, 00:08:15.991 "num_base_bdevs_discovered": 2, 00:08:15.991 "num_base_bdevs_operational": 2, 00:08:15.991 "base_bdevs_list": [ 00:08:15.991 { 00:08:15.991 "name": "pt1", 00:08:15.991 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:15.991 "is_configured": true, 00:08:15.991 "data_offset": 2048, 00:08:15.991 "data_size": 63488 00:08:15.991 }, 00:08:15.991 { 00:08:15.991 "name": "pt2", 00:08:15.991 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:15.991 "is_configured": true, 00:08:15.991 "data_offset": 2048, 00:08:15.991 "data_size": 63488 00:08:15.991 } 00:08:15.991 ] 00:08:15.991 }' 00:08:15.991 17:29:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:15.991 17:29:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:16.251 17:29:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:08:16.251 17:29:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:08:16.251 17:29:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:16.251 17:29:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:16.251 17:29:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:08:16.251 17:29:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:16.251 17:29:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:08:16.251 17:29:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:16.251 17:29:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:16.251 17:29:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:16.251 [2024-11-27 17:29:47.396863] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:16.251 17:29:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:16.251 17:29:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:16.251 "name": "raid_bdev1", 00:08:16.251 "aliases": [ 00:08:16.251 "a967b97a-4454-4f98-a55d-5b6e90f97dd7" 00:08:16.251 ], 00:08:16.251 "product_name": "Raid Volume", 00:08:16.251 "block_size": 512, 00:08:16.251 "num_blocks": 63488, 00:08:16.251 "uuid": "a967b97a-4454-4f98-a55d-5b6e90f97dd7", 00:08:16.251 "assigned_rate_limits": { 00:08:16.251 "rw_ios_per_sec": 0, 00:08:16.251 "rw_mbytes_per_sec": 0, 00:08:16.251 "r_mbytes_per_sec": 0, 00:08:16.251 "w_mbytes_per_sec": 0 00:08:16.251 }, 00:08:16.251 "claimed": false, 00:08:16.251 "zoned": false, 00:08:16.251 "supported_io_types": { 00:08:16.251 "read": true, 00:08:16.251 "write": true, 00:08:16.251 "unmap": false, 00:08:16.251 "flush": false, 00:08:16.251 "reset": true, 00:08:16.251 "nvme_admin": false, 00:08:16.251 "nvme_io": false, 00:08:16.251 "nvme_io_md": false, 00:08:16.251 "write_zeroes": true, 00:08:16.251 "zcopy": false, 00:08:16.251 "get_zone_info": false, 00:08:16.251 "zone_management": false, 00:08:16.251 "zone_append": false, 00:08:16.251 "compare": false, 00:08:16.251 "compare_and_write": false, 00:08:16.251 "abort": false, 00:08:16.251 "seek_hole": false, 00:08:16.251 "seek_data": false, 00:08:16.251 "copy": false, 00:08:16.251 "nvme_iov_md": false 00:08:16.251 }, 00:08:16.251 "memory_domains": [ 00:08:16.251 { 00:08:16.251 "dma_device_id": "system", 00:08:16.251 "dma_device_type": 1 00:08:16.251 }, 00:08:16.251 { 00:08:16.251 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:16.251 "dma_device_type": 2 00:08:16.251 }, 00:08:16.251 { 00:08:16.251 "dma_device_id": "system", 00:08:16.251 "dma_device_type": 1 00:08:16.251 }, 00:08:16.251 { 00:08:16.251 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:16.251 "dma_device_type": 2 00:08:16.251 } 00:08:16.251 ], 00:08:16.251 "driver_specific": { 00:08:16.251 "raid": { 00:08:16.251 "uuid": "a967b97a-4454-4f98-a55d-5b6e90f97dd7", 00:08:16.251 "strip_size_kb": 0, 00:08:16.251 "state": "online", 00:08:16.251 "raid_level": "raid1", 00:08:16.251 "superblock": true, 00:08:16.251 "num_base_bdevs": 2, 00:08:16.251 "num_base_bdevs_discovered": 2, 00:08:16.251 "num_base_bdevs_operational": 2, 00:08:16.251 "base_bdevs_list": [ 00:08:16.251 { 00:08:16.251 "name": "pt1", 00:08:16.251 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:16.251 "is_configured": true, 00:08:16.251 "data_offset": 2048, 00:08:16.251 "data_size": 63488 00:08:16.251 }, 00:08:16.251 { 00:08:16.251 "name": "pt2", 00:08:16.251 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:16.251 "is_configured": true, 00:08:16.251 "data_offset": 2048, 00:08:16.251 "data_size": 63488 00:08:16.251 } 00:08:16.251 ] 00:08:16.251 } 00:08:16.251 } 00:08:16.251 }' 00:08:16.251 17:29:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:16.511 17:29:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:08:16.511 pt2' 00:08:16.511 17:29:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:16.511 17:29:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:16.511 17:29:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:16.511 17:29:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:08:16.511 17:29:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:16.511 17:29:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:16.511 17:29:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:16.511 17:29:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:16.511 17:29:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:16.511 17:29:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:16.511 17:29:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:16.511 17:29:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:08:16.511 17:29:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:16.511 17:29:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:16.511 17:29:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:16.511 17:29:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:16.511 17:29:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:16.511 17:29:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:16.512 17:29:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:08:16.512 17:29:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:16.512 17:29:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:16.512 17:29:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:08:16.512 [2024-11-27 17:29:47.620493] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:16.512 17:29:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:16.512 17:29:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' a967b97a-4454-4f98-a55d-5b6e90f97dd7 '!=' a967b97a-4454-4f98-a55d-5b6e90f97dd7 ']' 00:08:16.512 17:29:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy raid1 00:08:16.512 17:29:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:08:16.512 17:29:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@199 -- # return 0 00:08:16.512 17:29:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@493 -- # rpc_cmd bdev_passthru_delete pt1 00:08:16.512 17:29:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:16.512 17:29:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:16.512 [2024-11-27 17:29:47.672248] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: pt1 00:08:16.512 17:29:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:16.512 17:29:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@496 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:08:16.512 17:29:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:16.512 17:29:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:16.512 17:29:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:16.512 17:29:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:16.512 17:29:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:08:16.512 17:29:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:16.512 17:29:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:16.512 17:29:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:16.512 17:29:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:16.512 17:29:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:16.512 17:29:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:16.512 17:29:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:16.512 17:29:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:16.772 17:29:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:16.772 17:29:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:16.772 "name": "raid_bdev1", 00:08:16.772 "uuid": "a967b97a-4454-4f98-a55d-5b6e90f97dd7", 00:08:16.772 "strip_size_kb": 0, 00:08:16.772 "state": "online", 00:08:16.772 "raid_level": "raid1", 00:08:16.772 "superblock": true, 00:08:16.772 "num_base_bdevs": 2, 00:08:16.772 "num_base_bdevs_discovered": 1, 00:08:16.772 "num_base_bdevs_operational": 1, 00:08:16.772 "base_bdevs_list": [ 00:08:16.772 { 00:08:16.772 "name": null, 00:08:16.772 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:16.772 "is_configured": false, 00:08:16.772 "data_offset": 0, 00:08:16.772 "data_size": 63488 00:08:16.772 }, 00:08:16.772 { 00:08:16.772 "name": "pt2", 00:08:16.772 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:16.772 "is_configured": true, 00:08:16.772 "data_offset": 2048, 00:08:16.772 "data_size": 63488 00:08:16.772 } 00:08:16.772 ] 00:08:16.772 }' 00:08:16.772 17:29:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:16.772 17:29:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:17.032 17:29:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@499 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:08:17.032 17:29:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:17.032 17:29:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:17.032 [2024-11-27 17:29:48.087442] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:08:17.032 [2024-11-27 17:29:48.087480] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:17.032 [2024-11-27 17:29:48.087553] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:17.032 [2024-11-27 17:29:48.087602] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:17.032 [2024-11-27 17:29:48.087611] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:08:17.032 17:29:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:17.032 17:29:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:17.032 17:29:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:17.032 17:29:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:17.032 17:29:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # jq -r '.[]' 00:08:17.032 17:29:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:17.032 17:29:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # raid_bdev= 00:08:17.032 17:29:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@501 -- # '[' -n '' ']' 00:08:17.032 17:29:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i = 1 )) 00:08:17.032 17:29:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:08:17.032 17:29:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt2 00:08:17.032 17:29:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:17.032 17:29:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:17.033 17:29:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:17.033 17:29:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:08:17.033 17:29:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:08:17.033 17:29:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i = 1 )) 00:08:17.033 17:29:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:08:17.033 17:29:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@519 -- # i=1 00:08:17.033 17:29:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@520 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:08:17.033 17:29:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:17.033 17:29:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:17.033 [2024-11-27 17:29:48.159308] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:08:17.033 [2024-11-27 17:29:48.159364] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:17.033 [2024-11-27 17:29:48.159401] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008780 00:08:17.033 [2024-11-27 17:29:48.159410] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:17.033 [2024-11-27 17:29:48.161871] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:17.033 [2024-11-27 17:29:48.161907] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:08:17.033 [2024-11-27 17:29:48.161987] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:08:17.033 [2024-11-27 17:29:48.162021] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:08:17.033 [2024-11-27 17:29:48.162100] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:08:17.033 [2024-11-27 17:29:48.162107] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:08:17.033 [2024-11-27 17:29:48.162365] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:08:17.033 [2024-11-27 17:29:48.162501] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:08:17.033 [2024-11-27 17:29:48.162512] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001c80 00:08:17.033 [2024-11-27 17:29:48.162616] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:17.033 pt2 00:08:17.033 17:29:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:17.033 17:29:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@523 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:08:17.033 17:29:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:17.033 17:29:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:17.033 17:29:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:17.033 17:29:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:17.033 17:29:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:08:17.033 17:29:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:17.033 17:29:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:17.033 17:29:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:17.033 17:29:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:17.033 17:29:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:17.033 17:29:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:17.033 17:29:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:17.033 17:29:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:17.033 17:29:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:17.033 17:29:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:17.033 "name": "raid_bdev1", 00:08:17.033 "uuid": "a967b97a-4454-4f98-a55d-5b6e90f97dd7", 00:08:17.033 "strip_size_kb": 0, 00:08:17.033 "state": "online", 00:08:17.033 "raid_level": "raid1", 00:08:17.033 "superblock": true, 00:08:17.033 "num_base_bdevs": 2, 00:08:17.033 "num_base_bdevs_discovered": 1, 00:08:17.033 "num_base_bdevs_operational": 1, 00:08:17.033 "base_bdevs_list": [ 00:08:17.033 { 00:08:17.033 "name": null, 00:08:17.033 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:17.033 "is_configured": false, 00:08:17.033 "data_offset": 2048, 00:08:17.033 "data_size": 63488 00:08:17.033 }, 00:08:17.033 { 00:08:17.033 "name": "pt2", 00:08:17.033 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:17.033 "is_configured": true, 00:08:17.033 "data_offset": 2048, 00:08:17.033 "data_size": 63488 00:08:17.033 } 00:08:17.033 ] 00:08:17.033 }' 00:08:17.033 17:29:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:17.033 17:29:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:17.603 17:29:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@526 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:08:17.603 17:29:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:17.603 17:29:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:17.603 [2024-11-27 17:29:48.622600] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:08:17.603 [2024-11-27 17:29:48.622665] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:17.603 [2024-11-27 17:29:48.622758] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:17.603 [2024-11-27 17:29:48.622839] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:17.603 [2024-11-27 17:29:48.622903] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name raid_bdev1, state offline 00:08:17.603 17:29:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:17.603 17:29:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:17.603 17:29:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # jq -r '.[]' 00:08:17.603 17:29:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:17.603 17:29:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:17.603 17:29:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:17.603 17:29:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # raid_bdev= 00:08:17.603 17:29:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@528 -- # '[' -n '' ']' 00:08:17.603 17:29:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@532 -- # '[' 2 -gt 2 ']' 00:08:17.603 17:29:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@540 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:08:17.603 17:29:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:17.603 17:29:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:17.603 [2024-11-27 17:29:48.682532] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:08:17.603 [2024-11-27 17:29:48.682637] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:17.603 [2024-11-27 17:29:48.682669] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008d80 00:08:17.603 [2024-11-27 17:29:48.682705] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:17.603 [2024-11-27 17:29:48.685102] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:17.603 [2024-11-27 17:29:48.685200] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:08:17.603 [2024-11-27 17:29:48.685292] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:08:17.603 [2024-11-27 17:29:48.685369] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:08:17.603 [2024-11-27 17:29:48.685498] bdev_raid.c:3675:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev pt2 (4) greater than existing raid bdev raid_bdev1 (2) 00:08:17.603 [2024-11-27 17:29:48.685555] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:08:17.603 [2024-11-27 17:29:48.685615] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002000 name raid_bdev1, state configuring 00:08:17.603 [2024-11-27 17:29:48.685689] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:08:17.603 [2024-11-27 17:29:48.685786] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000002380 00:08:17.603 [2024-11-27 17:29:48.685822] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:08:17.603 [2024-11-27 17:29:48.686056] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:08:17.603 [2024-11-27 17:29:48.686220] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000002380 00:08:17.603 [2024-11-27 17:29:48.686260] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000002380 00:08:17.603 [2024-11-27 17:29:48.686413] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:17.603 pt1 00:08:17.603 17:29:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:17.603 17:29:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@542 -- # '[' 2 -gt 2 ']' 00:08:17.603 17:29:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@554 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:08:17.603 17:29:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:17.603 17:29:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:17.603 17:29:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:17.603 17:29:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:17.603 17:29:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:08:17.603 17:29:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:17.603 17:29:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:17.603 17:29:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:17.603 17:29:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:17.603 17:29:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:17.603 17:29:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:17.603 17:29:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:17.604 17:29:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:17.604 17:29:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:17.604 17:29:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:17.604 "name": "raid_bdev1", 00:08:17.604 "uuid": "a967b97a-4454-4f98-a55d-5b6e90f97dd7", 00:08:17.604 "strip_size_kb": 0, 00:08:17.604 "state": "online", 00:08:17.604 "raid_level": "raid1", 00:08:17.604 "superblock": true, 00:08:17.604 "num_base_bdevs": 2, 00:08:17.604 "num_base_bdevs_discovered": 1, 00:08:17.604 "num_base_bdevs_operational": 1, 00:08:17.604 "base_bdevs_list": [ 00:08:17.604 { 00:08:17.604 "name": null, 00:08:17.604 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:17.604 "is_configured": false, 00:08:17.604 "data_offset": 2048, 00:08:17.604 "data_size": 63488 00:08:17.604 }, 00:08:17.604 { 00:08:17.604 "name": "pt2", 00:08:17.604 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:17.604 "is_configured": true, 00:08:17.604 "data_offset": 2048, 00:08:17.604 "data_size": 63488 00:08:17.604 } 00:08:17.604 ] 00:08:17.604 }' 00:08:17.604 17:29:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:17.604 17:29:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:17.863 17:29:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # rpc_cmd bdev_raid_get_bdevs online 00:08:17.863 17:29:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:17.863 17:29:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:17.863 17:29:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:08:17.863 17:29:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:18.124 17:29:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # [[ false == \f\a\l\s\e ]] 00:08:18.124 17:29:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # jq -r '.[] | .uuid' 00:08:18.124 17:29:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:08:18.124 17:29:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:18.124 17:29:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:18.124 [2024-11-27 17:29:49.102032] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:18.124 17:29:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:18.124 17:29:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # '[' a967b97a-4454-4f98-a55d-5b6e90f97dd7 '!=' a967b97a-4454-4f98-a55d-5b6e90f97dd7 ']' 00:08:18.124 17:29:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 74269 00:08:18.124 17:29:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@950 -- # '[' -z 74269 ']' 00:08:18.124 17:29:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # kill -0 74269 00:08:18.124 17:29:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@955 -- # uname 00:08:18.124 17:29:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:08:18.124 17:29:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 74269 00:08:18.124 killing process with pid 74269 00:08:18.124 17:29:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:08:18.124 17:29:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:08:18.124 17:29:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 74269' 00:08:18.124 17:29:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@969 -- # kill 74269 00:08:18.124 [2024-11-27 17:29:49.166481] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:18.124 [2024-11-27 17:29:49.166549] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:18.124 [2024-11-27 17:29:49.166592] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:18.125 [2024-11-27 17:29:49.166600] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002380 name raid_bdev1, state offline 00:08:18.125 17:29:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@974 -- # wait 74269 00:08:18.125 [2024-11-27 17:29:49.207819] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:18.387 17:29:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:08:18.387 00:08:18.387 real 0m5.065s 00:08:18.387 user 0m8.053s 00:08:18.387 sys 0m1.091s 00:08:18.387 ************************************ 00:08:18.387 END TEST raid_superblock_test 00:08:18.387 ************************************ 00:08:18.387 17:29:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:08:18.387 17:29:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:18.647 17:29:49 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test raid1 2 read 00:08:18.647 17:29:49 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:08:18.647 17:29:49 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:08:18.647 17:29:49 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:18.647 ************************************ 00:08:18.647 START TEST raid_read_error_test 00:08:18.647 ************************************ 00:08:18.647 17:29:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1125 -- # raid_io_error_test raid1 2 read 00:08:18.647 17:29:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid1 00:08:18.647 17:29:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=2 00:08:18.647 17:29:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:08:18.647 17:29:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:08:18.647 17:29:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:18.647 17:29:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:08:18.647 17:29:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:08:18.647 17:29:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:18.647 17:29:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:08:18.647 17:29:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:08:18.647 17:29:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:18.647 17:29:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:08:18.647 17:29:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:08:18.647 17:29:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:08:18.647 17:29:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:08:18.647 17:29:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:08:18.647 17:29:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:08:18.647 17:29:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:08:18.647 17:29:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid1 '!=' raid1 ']' 00:08:18.647 17:29:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@804 -- # strip_size=0 00:08:18.647 17:29:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:08:18.647 17:29:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.jFFZoOpUX8 00:08:18.647 17:29:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=74587 00:08:18.647 17:29:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:08:18.647 17:29:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 74587 00:08:18.647 17:29:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@831 -- # '[' -z 74587 ']' 00:08:18.647 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:18.647 17:29:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:18.647 17:29:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:08:18.647 17:29:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:18.647 17:29:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:08:18.647 17:29:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:18.647 [2024-11-27 17:29:49.754266] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:08:18.647 [2024-11-27 17:29:49.754419] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid74587 ] 00:08:18.906 [2024-11-27 17:29:49.901000] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:18.906 [2024-11-27 17:29:49.969346] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:08:18.906 [2024-11-27 17:29:50.044776] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:18.906 [2024-11-27 17:29:50.044818] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:19.475 17:29:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:08:19.475 17:29:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@864 -- # return 0 00:08:19.475 17:29:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:08:19.475 17:29:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:08:19.475 17:29:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:19.475 17:29:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:19.475 BaseBdev1_malloc 00:08:19.475 17:29:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:19.475 17:29:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:08:19.475 17:29:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:19.475 17:29:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:19.475 true 00:08:19.475 17:29:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:19.475 17:29:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:08:19.475 17:29:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:19.475 17:29:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:19.475 [2024-11-27 17:29:50.619504] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:08:19.475 [2024-11-27 17:29:50.619576] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:19.475 [2024-11-27 17:29:50.619605] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:08:19.475 [2024-11-27 17:29:50.619621] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:19.475 [2024-11-27 17:29:50.622140] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:19.475 [2024-11-27 17:29:50.622205] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:08:19.475 BaseBdev1 00:08:19.475 17:29:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:19.475 17:29:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:08:19.475 17:29:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:08:19.475 17:29:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:19.475 17:29:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:19.475 BaseBdev2_malloc 00:08:19.475 17:29:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:19.475 17:29:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:08:19.475 17:29:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:19.475 17:29:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:19.735 true 00:08:19.735 17:29:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:19.735 17:29:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:08:19.735 17:29:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:19.735 17:29:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:19.735 [2024-11-27 17:29:50.681823] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:08:19.735 [2024-11-27 17:29:50.681908] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:19.735 [2024-11-27 17:29:50.681941] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:08:19.735 [2024-11-27 17:29:50.681954] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:19.735 [2024-11-27 17:29:50.685518] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:19.735 [2024-11-27 17:29:50.685629] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:08:19.735 BaseBdev2 00:08:19.735 17:29:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:19.735 17:29:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 -s 00:08:19.735 17:29:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:19.735 17:29:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:19.735 [2024-11-27 17:29:50.693938] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:19.735 [2024-11-27 17:29:50.696246] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:19.735 [2024-11-27 17:29:50.696502] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:08:19.735 [2024-11-27 17:29:50.696519] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:08:19.735 [2024-11-27 17:29:50.696801] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:08:19.735 [2024-11-27 17:29:50.696947] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:08:19.735 [2024-11-27 17:29:50.696966] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:08:19.735 [2024-11-27 17:29:50.697095] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:19.735 17:29:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:19.735 17:29:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:08:19.735 17:29:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:19.735 17:29:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:19.735 17:29:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:19.735 17:29:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:19.735 17:29:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:19.735 17:29:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:19.735 17:29:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:19.735 17:29:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:19.735 17:29:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:19.735 17:29:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:19.735 17:29:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:19.735 17:29:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:19.735 17:29:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:19.735 17:29:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:19.735 17:29:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:19.735 "name": "raid_bdev1", 00:08:19.735 "uuid": "f584868e-f02b-40e5-a130-2ece7e415869", 00:08:19.735 "strip_size_kb": 0, 00:08:19.735 "state": "online", 00:08:19.735 "raid_level": "raid1", 00:08:19.735 "superblock": true, 00:08:19.735 "num_base_bdevs": 2, 00:08:19.735 "num_base_bdevs_discovered": 2, 00:08:19.735 "num_base_bdevs_operational": 2, 00:08:19.735 "base_bdevs_list": [ 00:08:19.735 { 00:08:19.735 "name": "BaseBdev1", 00:08:19.735 "uuid": "7ada55dc-41b2-5f26-9398-e570504a0664", 00:08:19.735 "is_configured": true, 00:08:19.735 "data_offset": 2048, 00:08:19.735 "data_size": 63488 00:08:19.735 }, 00:08:19.735 { 00:08:19.735 "name": "BaseBdev2", 00:08:19.735 "uuid": "e3dfc28d-b7bd-5d36-8036-1619cbbe1e70", 00:08:19.735 "is_configured": true, 00:08:19.735 "data_offset": 2048, 00:08:19.735 "data_size": 63488 00:08:19.735 } 00:08:19.735 ] 00:08:19.735 }' 00:08:19.736 17:29:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:19.736 17:29:50 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:19.995 17:29:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:08:19.995 17:29:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:08:20.255 [2024-11-27 17:29:51.217520] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:08:21.196 17:29:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:08:21.196 17:29:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:21.196 17:29:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:21.196 17:29:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:21.196 17:29:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:08:21.196 17:29:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid1 = \r\a\i\d\1 ]] 00:08:21.196 17:29:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ read = \w\r\i\t\e ]] 00:08:21.196 17:29:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=2 00:08:21.196 17:29:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:08:21.196 17:29:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:21.196 17:29:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:21.196 17:29:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:21.196 17:29:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:21.196 17:29:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:21.196 17:29:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:21.196 17:29:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:21.196 17:29:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:21.196 17:29:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:21.196 17:29:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:21.196 17:29:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:21.196 17:29:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:21.196 17:29:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:21.196 17:29:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:21.196 17:29:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:21.196 "name": "raid_bdev1", 00:08:21.196 "uuid": "f584868e-f02b-40e5-a130-2ece7e415869", 00:08:21.196 "strip_size_kb": 0, 00:08:21.196 "state": "online", 00:08:21.196 "raid_level": "raid1", 00:08:21.196 "superblock": true, 00:08:21.196 "num_base_bdevs": 2, 00:08:21.196 "num_base_bdevs_discovered": 2, 00:08:21.196 "num_base_bdevs_operational": 2, 00:08:21.196 "base_bdevs_list": [ 00:08:21.196 { 00:08:21.196 "name": "BaseBdev1", 00:08:21.196 "uuid": "7ada55dc-41b2-5f26-9398-e570504a0664", 00:08:21.196 "is_configured": true, 00:08:21.196 "data_offset": 2048, 00:08:21.196 "data_size": 63488 00:08:21.196 }, 00:08:21.196 { 00:08:21.196 "name": "BaseBdev2", 00:08:21.196 "uuid": "e3dfc28d-b7bd-5d36-8036-1619cbbe1e70", 00:08:21.196 "is_configured": true, 00:08:21.196 "data_offset": 2048, 00:08:21.196 "data_size": 63488 00:08:21.196 } 00:08:21.196 ] 00:08:21.196 }' 00:08:21.196 17:29:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:21.196 17:29:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:21.457 17:29:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:08:21.457 17:29:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:21.457 17:29:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:21.457 [2024-11-27 17:29:52.595769] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:08:21.457 [2024-11-27 17:29:52.595807] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:21.457 [2024-11-27 17:29:52.598440] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:21.457 [2024-11-27 17:29:52.598522] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:21.457 [2024-11-27 17:29:52.598651] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:21.457 [2024-11-27 17:29:52.598708] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:08:21.457 { 00:08:21.457 "results": [ 00:08:21.457 { 00:08:21.457 "job": "raid_bdev1", 00:08:21.457 "core_mask": "0x1", 00:08:21.457 "workload": "randrw", 00:08:21.457 "percentage": 50, 00:08:21.457 "status": "finished", 00:08:21.457 "queue_depth": 1, 00:08:21.457 "io_size": 131072, 00:08:21.457 "runtime": 1.378806, 00:08:21.457 "iops": 16187.193847430313, 00:08:21.457 "mibps": 2023.3992309287892, 00:08:21.457 "io_failed": 0, 00:08:21.457 "io_timeout": 0, 00:08:21.457 "avg_latency_us": 59.3001278210685, 00:08:21.457 "min_latency_us": 22.022707423580787, 00:08:21.457 "max_latency_us": 1452.380786026201 00:08:21.457 } 00:08:21.457 ], 00:08:21.457 "core_count": 1 00:08:21.457 } 00:08:21.457 17:29:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:21.457 17:29:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 74587 00:08:21.457 17:29:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@950 -- # '[' -z 74587 ']' 00:08:21.457 17:29:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # kill -0 74587 00:08:21.457 17:29:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@955 -- # uname 00:08:21.457 17:29:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:08:21.457 17:29:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 74587 00:08:21.457 killing process with pid 74587 00:08:21.457 17:29:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:08:21.457 17:29:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:08:21.457 17:29:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 74587' 00:08:21.457 17:29:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@969 -- # kill 74587 00:08:21.457 [2024-11-27 17:29:52.637967] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:21.457 17:29:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@974 -- # wait 74587 00:08:21.717 [2024-11-27 17:29:52.667779] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:21.978 17:29:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.jFFZoOpUX8 00:08:21.978 17:29:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:08:21.978 17:29:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:08:21.978 17:29:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.00 00:08:21.978 17:29:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid1 00:08:21.978 17:29:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:08:21.978 17:29:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@199 -- # return 0 00:08:21.978 17:29:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@847 -- # [[ 0.00 = \0\.\0\0 ]] 00:08:21.978 00:08:21.978 real 0m3.395s 00:08:21.978 user 0m4.165s 00:08:21.978 sys 0m0.614s 00:08:21.978 17:29:53 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:08:21.978 ************************************ 00:08:21.978 END TEST raid_read_error_test 00:08:21.978 ************************************ 00:08:21.978 17:29:53 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:21.978 17:29:53 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test raid1 2 write 00:08:21.978 17:29:53 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:08:21.978 17:29:53 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:08:21.978 17:29:53 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:21.978 ************************************ 00:08:21.978 START TEST raid_write_error_test 00:08:21.978 ************************************ 00:08:21.978 17:29:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1125 -- # raid_io_error_test raid1 2 write 00:08:21.978 17:29:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid1 00:08:21.978 17:29:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=2 00:08:21.978 17:29:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:08:21.978 17:29:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:08:21.978 17:29:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:21.978 17:29:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:08:21.978 17:29:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:08:21.978 17:29:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:21.978 17:29:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:08:21.978 17:29:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:08:21.978 17:29:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:21.978 17:29:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:08:21.978 17:29:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:08:21.978 17:29:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:08:21.978 17:29:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:08:21.978 17:29:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:08:21.978 17:29:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:08:21.978 17:29:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:08:21.978 17:29:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid1 '!=' raid1 ']' 00:08:21.978 17:29:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@804 -- # strip_size=0 00:08:21.978 17:29:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:08:21.978 17:29:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.OXIRugIDUD 00:08:21.978 17:29:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=74717 00:08:21.978 17:29:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:08:21.978 17:29:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 74717 00:08:21.978 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:21.978 17:29:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@831 -- # '[' -z 74717 ']' 00:08:21.978 17:29:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:21.978 17:29:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:08:21.978 17:29:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:21.978 17:29:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:08:21.978 17:29:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:22.239 [2024-11-27 17:29:53.223698] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:08:22.239 [2024-11-27 17:29:53.223852] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid74717 ] 00:08:22.239 [2024-11-27 17:29:53.350201] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:22.239 [2024-11-27 17:29:53.416583] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:08:22.498 [2024-11-27 17:29:53.492681] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:22.498 [2024-11-27 17:29:53.492725] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:23.068 17:29:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:08:23.068 17:29:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@864 -- # return 0 00:08:23.068 17:29:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:08:23.068 17:29:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:08:23.068 17:29:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:23.068 17:29:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:23.068 BaseBdev1_malloc 00:08:23.068 17:29:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:23.068 17:29:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:08:23.068 17:29:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:23.068 17:29:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:23.068 true 00:08:23.068 17:29:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:23.068 17:29:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:08:23.068 17:29:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:23.068 17:29:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:23.068 [2024-11-27 17:29:54.094080] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:08:23.068 [2024-11-27 17:29:54.094155] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:23.068 [2024-11-27 17:29:54.094180] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:08:23.068 [2024-11-27 17:29:54.094189] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:23.068 [2024-11-27 17:29:54.096620] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:23.068 [2024-11-27 17:29:54.096694] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:08:23.068 BaseBdev1 00:08:23.069 17:29:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:23.069 17:29:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:08:23.069 17:29:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:08:23.069 17:29:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:23.069 17:29:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:23.069 BaseBdev2_malloc 00:08:23.069 17:29:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:23.069 17:29:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:08:23.069 17:29:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:23.069 17:29:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:23.069 true 00:08:23.069 17:29:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:23.069 17:29:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:08:23.069 17:29:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:23.069 17:29:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:23.069 [2024-11-27 17:29:54.154409] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:08:23.069 [2024-11-27 17:29:54.154525] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:23.069 [2024-11-27 17:29:54.154558] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:08:23.069 [2024-11-27 17:29:54.154570] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:23.069 [2024-11-27 17:29:54.157417] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:23.069 [2024-11-27 17:29:54.157458] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:08:23.069 BaseBdev2 00:08:23.069 17:29:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:23.069 17:29:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 -s 00:08:23.069 17:29:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:23.069 17:29:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:23.069 [2024-11-27 17:29:54.166452] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:23.069 [2024-11-27 17:29:54.168544] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:23.069 [2024-11-27 17:29:54.168739] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:08:23.069 [2024-11-27 17:29:54.168752] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:08:23.069 [2024-11-27 17:29:54.169025] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:08:23.069 [2024-11-27 17:29:54.169175] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:08:23.069 [2024-11-27 17:29:54.169194] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:08:23.069 [2024-11-27 17:29:54.169330] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:23.069 17:29:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:23.069 17:29:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:08:23.069 17:29:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:23.069 17:29:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:23.069 17:29:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:23.069 17:29:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:23.069 17:29:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:23.069 17:29:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:23.069 17:29:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:23.069 17:29:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:23.069 17:29:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:23.069 17:29:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:23.069 17:29:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:23.069 17:29:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:23.069 17:29:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:23.069 17:29:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:23.069 17:29:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:23.069 "name": "raid_bdev1", 00:08:23.069 "uuid": "775da72e-f230-4953-900e-e49f224c7485", 00:08:23.069 "strip_size_kb": 0, 00:08:23.069 "state": "online", 00:08:23.069 "raid_level": "raid1", 00:08:23.069 "superblock": true, 00:08:23.069 "num_base_bdevs": 2, 00:08:23.069 "num_base_bdevs_discovered": 2, 00:08:23.069 "num_base_bdevs_operational": 2, 00:08:23.069 "base_bdevs_list": [ 00:08:23.069 { 00:08:23.069 "name": "BaseBdev1", 00:08:23.069 "uuid": "f4040a35-3ebe-56f1-9b30-5761b7a23df5", 00:08:23.069 "is_configured": true, 00:08:23.069 "data_offset": 2048, 00:08:23.069 "data_size": 63488 00:08:23.069 }, 00:08:23.069 { 00:08:23.069 "name": "BaseBdev2", 00:08:23.069 "uuid": "fb6138e9-4106-5b63-a948-8a957094a335", 00:08:23.069 "is_configured": true, 00:08:23.069 "data_offset": 2048, 00:08:23.069 "data_size": 63488 00:08:23.069 } 00:08:23.069 ] 00:08:23.069 }' 00:08:23.069 17:29:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:23.069 17:29:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:23.638 17:29:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:08:23.638 17:29:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:08:23.638 [2024-11-27 17:29:54.737922] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:08:24.582 17:29:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:08:24.582 17:29:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:24.582 17:29:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:24.582 [2024-11-27 17:29:55.664375] bdev_raid.c:2272:_raid_bdev_fail_base_bdev: *NOTICE*: Failing base bdev in slot 0 ('BaseBdev1') of raid bdev 'raid_bdev1' 00:08:24.582 [2024-11-27 17:29:55.664535] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:08:24.582 [2024-11-27 17:29:55.664828] bdev_raid.c:1970:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 0 raid_ch: 0x60d000002530 00:08:24.582 17:29:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:24.582 17:29:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:08:24.582 17:29:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid1 = \r\a\i\d\1 ]] 00:08:24.582 17:29:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ write = \w\r\i\t\e ]] 00:08:24.582 17:29:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@833 -- # expected_num_base_bdevs=1 00:08:24.582 17:29:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:08:24.582 17:29:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:24.582 17:29:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:24.582 17:29:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:08:24.582 17:29:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:08:24.582 17:29:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:08:24.582 17:29:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:24.582 17:29:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:24.582 17:29:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:24.582 17:29:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:24.582 17:29:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:24.582 17:29:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:24.582 17:29:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:24.582 17:29:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:24.582 17:29:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:24.582 17:29:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:24.582 "name": "raid_bdev1", 00:08:24.582 "uuid": "775da72e-f230-4953-900e-e49f224c7485", 00:08:24.582 "strip_size_kb": 0, 00:08:24.582 "state": "online", 00:08:24.582 "raid_level": "raid1", 00:08:24.582 "superblock": true, 00:08:24.582 "num_base_bdevs": 2, 00:08:24.582 "num_base_bdevs_discovered": 1, 00:08:24.582 "num_base_bdevs_operational": 1, 00:08:24.582 "base_bdevs_list": [ 00:08:24.582 { 00:08:24.582 "name": null, 00:08:24.582 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:24.582 "is_configured": false, 00:08:24.582 "data_offset": 0, 00:08:24.582 "data_size": 63488 00:08:24.582 }, 00:08:24.582 { 00:08:24.582 "name": "BaseBdev2", 00:08:24.582 "uuid": "fb6138e9-4106-5b63-a948-8a957094a335", 00:08:24.582 "is_configured": true, 00:08:24.582 "data_offset": 2048, 00:08:24.582 "data_size": 63488 00:08:24.582 } 00:08:24.582 ] 00:08:24.582 }' 00:08:24.582 17:29:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:24.582 17:29:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:25.158 17:29:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:08:25.158 17:29:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:25.158 17:29:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:25.158 [2024-11-27 17:29:56.137348] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:08:25.158 [2024-11-27 17:29:56.137463] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:25.158 [2024-11-27 17:29:56.139974] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:25.158 [2024-11-27 17:29:56.140030] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:25.158 [2024-11-27 17:29:56.140088] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:25.158 [2024-11-27 17:29:56.140100] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:08:25.158 { 00:08:25.158 "results": [ 00:08:25.158 { 00:08:25.158 "job": "raid_bdev1", 00:08:25.158 "core_mask": "0x1", 00:08:25.158 "workload": "randrw", 00:08:25.158 "percentage": 50, 00:08:25.158 "status": "finished", 00:08:25.158 "queue_depth": 1, 00:08:25.158 "io_size": 131072, 00:08:25.158 "runtime": 1.399995, 00:08:25.158 "iops": 20223.6436558702, 00:08:25.158 "mibps": 2527.955456983775, 00:08:25.158 "io_failed": 0, 00:08:25.158 "io_timeout": 0, 00:08:25.158 "avg_latency_us": 46.970196695486216, 00:08:25.158 "min_latency_us": 21.128384279475984, 00:08:25.158 "max_latency_us": 1359.3711790393013 00:08:25.158 } 00:08:25.158 ], 00:08:25.158 "core_count": 1 00:08:25.158 } 00:08:25.158 17:29:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:25.158 17:29:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 74717 00:08:25.158 17:29:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@950 -- # '[' -z 74717 ']' 00:08:25.158 17:29:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # kill -0 74717 00:08:25.158 17:29:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@955 -- # uname 00:08:25.158 17:29:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:08:25.158 17:29:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 74717 00:08:25.158 17:29:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:08:25.158 17:29:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:08:25.158 killing process with pid 74717 00:08:25.158 17:29:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 74717' 00:08:25.158 17:29:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@969 -- # kill 74717 00:08:25.158 [2024-11-27 17:29:56.187518] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:25.158 17:29:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@974 -- # wait 74717 00:08:25.158 [2024-11-27 17:29:56.216147] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:25.419 17:29:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.OXIRugIDUD 00:08:25.419 17:29:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:08:25.419 17:29:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:08:25.419 17:29:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.00 00:08:25.419 17:29:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid1 00:08:25.419 17:29:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:08:25.419 17:29:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@199 -- # return 0 00:08:25.419 17:29:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@847 -- # [[ 0.00 = \0\.\0\0 ]] 00:08:25.419 00:08:25.419 real 0m3.469s 00:08:25.419 user 0m4.286s 00:08:25.419 sys 0m0.635s 00:08:25.419 ************************************ 00:08:25.419 END TEST raid_write_error_test 00:08:25.419 ************************************ 00:08:25.419 17:29:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:08:25.419 17:29:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:25.679 17:29:56 bdev_raid -- bdev/bdev_raid.sh@966 -- # for n in {2..4} 00:08:25.679 17:29:56 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:08:25.679 17:29:56 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test raid0 3 false 00:08:25.679 17:29:56 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:08:25.679 17:29:56 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:08:25.679 17:29:56 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:25.679 ************************************ 00:08:25.679 START TEST raid_state_function_test 00:08:25.679 ************************************ 00:08:25.679 17:29:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1125 -- # raid_state_function_test raid0 3 false 00:08:25.679 17:29:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=raid0 00:08:25.679 17:29:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=3 00:08:25.679 17:29:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:08:25.679 17:29:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:08:25.679 17:29:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:08:25.679 17:29:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:25.679 17:29:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:08:25.679 17:29:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:25.680 17:29:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:25.680 17:29:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:08:25.680 17:29:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:25.680 17:29:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:25.680 17:29:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:08:25.680 17:29:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:25.680 17:29:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:25.680 17:29:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:08:25.680 17:29:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:08:25.680 17:29:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:08:25.680 17:29:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:08:25.680 17:29:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:08:25.680 17:29:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:08:25.680 17:29:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' raid0 '!=' raid1 ']' 00:08:25.680 17:29:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:08:25.680 17:29:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:08:25.680 17:29:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:08:25.680 Process raid pid: 74850 00:08:25.680 17:29:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:08:25.680 17:29:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=74850 00:08:25.680 17:29:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:08:25.680 17:29:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 74850' 00:08:25.680 17:29:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 74850 00:08:25.680 17:29:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@831 -- # '[' -z 74850 ']' 00:08:25.680 17:29:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:25.680 17:29:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:08:25.680 17:29:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:25.680 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:25.680 17:29:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:08:25.680 17:29:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:25.680 [2024-11-27 17:29:56.760456] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:08:25.680 [2024-11-27 17:29:56.760567] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:08:25.940 [2024-11-27 17:29:56.886903] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:25.940 [2024-11-27 17:29:56.954074] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:08:25.940 [2024-11-27 17:29:57.029685] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:25.940 [2024-11-27 17:29:57.029725] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:26.511 17:29:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:08:26.511 17:29:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@864 -- # return 0 00:08:26.511 17:29:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:26.511 17:29:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:26.511 17:29:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:26.511 [2024-11-27 17:29:57.584648] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:26.511 [2024-11-27 17:29:57.584767] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:26.511 [2024-11-27 17:29:57.584801] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:26.511 [2024-11-27 17:29:57.584812] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:26.511 [2024-11-27 17:29:57.584818] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:08:26.511 [2024-11-27 17:29:57.584830] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:08:26.511 17:29:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:26.511 17:29:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:26.511 17:29:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:26.511 17:29:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:26.511 17:29:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:26.511 17:29:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:26.511 17:29:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:26.511 17:29:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:26.511 17:29:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:26.511 17:29:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:26.511 17:29:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:26.511 17:29:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:26.511 17:29:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:26.511 17:29:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:26.511 17:29:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:26.511 17:29:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:26.511 17:29:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:26.511 "name": "Existed_Raid", 00:08:26.511 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:26.511 "strip_size_kb": 64, 00:08:26.511 "state": "configuring", 00:08:26.511 "raid_level": "raid0", 00:08:26.511 "superblock": false, 00:08:26.511 "num_base_bdevs": 3, 00:08:26.511 "num_base_bdevs_discovered": 0, 00:08:26.511 "num_base_bdevs_operational": 3, 00:08:26.511 "base_bdevs_list": [ 00:08:26.511 { 00:08:26.511 "name": "BaseBdev1", 00:08:26.511 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:26.511 "is_configured": false, 00:08:26.511 "data_offset": 0, 00:08:26.511 "data_size": 0 00:08:26.511 }, 00:08:26.511 { 00:08:26.511 "name": "BaseBdev2", 00:08:26.511 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:26.511 "is_configured": false, 00:08:26.511 "data_offset": 0, 00:08:26.511 "data_size": 0 00:08:26.511 }, 00:08:26.511 { 00:08:26.511 "name": "BaseBdev3", 00:08:26.511 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:26.511 "is_configured": false, 00:08:26.511 "data_offset": 0, 00:08:26.511 "data_size": 0 00:08:26.511 } 00:08:26.511 ] 00:08:26.511 }' 00:08:26.511 17:29:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:26.511 17:29:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:27.079 17:29:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:27.079 17:29:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:27.079 17:29:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:27.079 [2024-11-27 17:29:58.047716] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:27.079 [2024-11-27 17:29:58.047803] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:08:27.079 17:29:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:27.079 17:29:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:27.079 17:29:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:27.079 17:29:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:27.079 [2024-11-27 17:29:58.059723] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:27.079 [2024-11-27 17:29:58.059765] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:27.079 [2024-11-27 17:29:58.059774] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:27.079 [2024-11-27 17:29:58.059799] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:27.079 [2024-11-27 17:29:58.059805] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:08:27.079 [2024-11-27 17:29:58.059814] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:08:27.079 17:29:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:27.079 17:29:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:08:27.079 17:29:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:27.079 17:29:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:27.079 [2024-11-27 17:29:58.086585] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:27.079 BaseBdev1 00:08:27.079 17:29:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:27.079 17:29:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:08:27.079 17:29:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:08:27.079 17:29:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:08:27.079 17:29:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:08:27.079 17:29:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:08:27.079 17:29:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:08:27.079 17:29:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:08:27.079 17:29:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:27.079 17:29:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:27.079 17:29:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:27.079 17:29:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:08:27.079 17:29:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:27.079 17:29:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:27.079 [ 00:08:27.079 { 00:08:27.079 "name": "BaseBdev1", 00:08:27.079 "aliases": [ 00:08:27.079 "161b3d93-c1f0-4d49-bb09-2f6568471749" 00:08:27.079 ], 00:08:27.079 "product_name": "Malloc disk", 00:08:27.079 "block_size": 512, 00:08:27.079 "num_blocks": 65536, 00:08:27.079 "uuid": "161b3d93-c1f0-4d49-bb09-2f6568471749", 00:08:27.079 "assigned_rate_limits": { 00:08:27.079 "rw_ios_per_sec": 0, 00:08:27.079 "rw_mbytes_per_sec": 0, 00:08:27.079 "r_mbytes_per_sec": 0, 00:08:27.079 "w_mbytes_per_sec": 0 00:08:27.079 }, 00:08:27.079 "claimed": true, 00:08:27.079 "claim_type": "exclusive_write", 00:08:27.079 "zoned": false, 00:08:27.079 "supported_io_types": { 00:08:27.080 "read": true, 00:08:27.080 "write": true, 00:08:27.080 "unmap": true, 00:08:27.080 "flush": true, 00:08:27.080 "reset": true, 00:08:27.080 "nvme_admin": false, 00:08:27.080 "nvme_io": false, 00:08:27.080 "nvme_io_md": false, 00:08:27.080 "write_zeroes": true, 00:08:27.080 "zcopy": true, 00:08:27.080 "get_zone_info": false, 00:08:27.080 "zone_management": false, 00:08:27.080 "zone_append": false, 00:08:27.080 "compare": false, 00:08:27.080 "compare_and_write": false, 00:08:27.080 "abort": true, 00:08:27.080 "seek_hole": false, 00:08:27.080 "seek_data": false, 00:08:27.080 "copy": true, 00:08:27.080 "nvme_iov_md": false 00:08:27.080 }, 00:08:27.080 "memory_domains": [ 00:08:27.080 { 00:08:27.080 "dma_device_id": "system", 00:08:27.080 "dma_device_type": 1 00:08:27.080 }, 00:08:27.080 { 00:08:27.080 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:27.080 "dma_device_type": 2 00:08:27.080 } 00:08:27.080 ], 00:08:27.080 "driver_specific": {} 00:08:27.080 } 00:08:27.080 ] 00:08:27.080 17:29:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:27.080 17:29:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:08:27.080 17:29:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:27.080 17:29:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:27.080 17:29:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:27.080 17:29:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:27.080 17:29:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:27.080 17:29:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:27.080 17:29:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:27.080 17:29:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:27.080 17:29:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:27.080 17:29:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:27.080 17:29:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:27.080 17:29:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:27.080 17:29:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:27.080 17:29:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:27.080 17:29:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:27.080 17:29:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:27.080 "name": "Existed_Raid", 00:08:27.080 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:27.080 "strip_size_kb": 64, 00:08:27.080 "state": "configuring", 00:08:27.080 "raid_level": "raid0", 00:08:27.080 "superblock": false, 00:08:27.080 "num_base_bdevs": 3, 00:08:27.080 "num_base_bdevs_discovered": 1, 00:08:27.080 "num_base_bdevs_operational": 3, 00:08:27.080 "base_bdevs_list": [ 00:08:27.080 { 00:08:27.080 "name": "BaseBdev1", 00:08:27.080 "uuid": "161b3d93-c1f0-4d49-bb09-2f6568471749", 00:08:27.080 "is_configured": true, 00:08:27.080 "data_offset": 0, 00:08:27.080 "data_size": 65536 00:08:27.080 }, 00:08:27.080 { 00:08:27.080 "name": "BaseBdev2", 00:08:27.080 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:27.080 "is_configured": false, 00:08:27.080 "data_offset": 0, 00:08:27.080 "data_size": 0 00:08:27.080 }, 00:08:27.080 { 00:08:27.080 "name": "BaseBdev3", 00:08:27.080 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:27.080 "is_configured": false, 00:08:27.080 "data_offset": 0, 00:08:27.080 "data_size": 0 00:08:27.080 } 00:08:27.080 ] 00:08:27.080 }' 00:08:27.080 17:29:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:27.080 17:29:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:27.650 17:29:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:27.650 17:29:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:27.650 17:29:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:27.650 [2024-11-27 17:29:58.569777] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:27.650 [2024-11-27 17:29:58.569824] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:08:27.650 17:29:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:27.650 17:29:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:27.650 17:29:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:27.650 17:29:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:27.650 [2024-11-27 17:29:58.581809] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:27.650 [2024-11-27 17:29:58.583992] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:27.650 [2024-11-27 17:29:58.584082] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:27.650 [2024-11-27 17:29:58.584110] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:08:27.650 [2024-11-27 17:29:58.584133] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:08:27.650 17:29:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:27.650 17:29:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:08:27.650 17:29:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:27.650 17:29:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:27.650 17:29:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:27.650 17:29:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:27.650 17:29:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:27.650 17:29:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:27.650 17:29:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:27.650 17:29:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:27.650 17:29:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:27.650 17:29:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:27.650 17:29:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:27.650 17:29:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:27.650 17:29:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:27.650 17:29:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:27.650 17:29:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:27.650 17:29:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:27.650 17:29:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:27.650 "name": "Existed_Raid", 00:08:27.650 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:27.650 "strip_size_kb": 64, 00:08:27.650 "state": "configuring", 00:08:27.650 "raid_level": "raid0", 00:08:27.650 "superblock": false, 00:08:27.650 "num_base_bdevs": 3, 00:08:27.650 "num_base_bdevs_discovered": 1, 00:08:27.650 "num_base_bdevs_operational": 3, 00:08:27.650 "base_bdevs_list": [ 00:08:27.650 { 00:08:27.650 "name": "BaseBdev1", 00:08:27.650 "uuid": "161b3d93-c1f0-4d49-bb09-2f6568471749", 00:08:27.650 "is_configured": true, 00:08:27.650 "data_offset": 0, 00:08:27.650 "data_size": 65536 00:08:27.650 }, 00:08:27.650 { 00:08:27.650 "name": "BaseBdev2", 00:08:27.650 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:27.650 "is_configured": false, 00:08:27.650 "data_offset": 0, 00:08:27.650 "data_size": 0 00:08:27.650 }, 00:08:27.650 { 00:08:27.650 "name": "BaseBdev3", 00:08:27.650 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:27.650 "is_configured": false, 00:08:27.650 "data_offset": 0, 00:08:27.650 "data_size": 0 00:08:27.650 } 00:08:27.650 ] 00:08:27.650 }' 00:08:27.650 17:29:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:27.650 17:29:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:27.911 17:29:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:08:27.911 17:29:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:27.911 17:29:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:27.911 [2024-11-27 17:29:59.083726] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:27.911 BaseBdev2 00:08:27.911 17:29:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:27.911 17:29:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:08:27.911 17:29:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:08:27.911 17:29:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:08:27.911 17:29:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:08:27.911 17:29:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:08:27.911 17:29:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:08:27.911 17:29:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:08:27.911 17:29:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:27.911 17:29:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:28.172 17:29:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:28.172 17:29:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:08:28.172 17:29:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:28.172 17:29:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:28.172 [ 00:08:28.172 { 00:08:28.172 "name": "BaseBdev2", 00:08:28.172 "aliases": [ 00:08:28.172 "cb63dfcd-9a45-40ab-a91b-352649ba9fb0" 00:08:28.172 ], 00:08:28.172 "product_name": "Malloc disk", 00:08:28.172 "block_size": 512, 00:08:28.172 "num_blocks": 65536, 00:08:28.172 "uuid": "cb63dfcd-9a45-40ab-a91b-352649ba9fb0", 00:08:28.172 "assigned_rate_limits": { 00:08:28.172 "rw_ios_per_sec": 0, 00:08:28.172 "rw_mbytes_per_sec": 0, 00:08:28.172 "r_mbytes_per_sec": 0, 00:08:28.172 "w_mbytes_per_sec": 0 00:08:28.172 }, 00:08:28.172 "claimed": true, 00:08:28.172 "claim_type": "exclusive_write", 00:08:28.172 "zoned": false, 00:08:28.172 "supported_io_types": { 00:08:28.172 "read": true, 00:08:28.172 "write": true, 00:08:28.172 "unmap": true, 00:08:28.172 "flush": true, 00:08:28.172 "reset": true, 00:08:28.172 "nvme_admin": false, 00:08:28.172 "nvme_io": false, 00:08:28.172 "nvme_io_md": false, 00:08:28.172 "write_zeroes": true, 00:08:28.172 "zcopy": true, 00:08:28.172 "get_zone_info": false, 00:08:28.172 "zone_management": false, 00:08:28.172 "zone_append": false, 00:08:28.172 "compare": false, 00:08:28.172 "compare_and_write": false, 00:08:28.172 "abort": true, 00:08:28.172 "seek_hole": false, 00:08:28.172 "seek_data": false, 00:08:28.172 "copy": true, 00:08:28.172 "nvme_iov_md": false 00:08:28.172 }, 00:08:28.172 "memory_domains": [ 00:08:28.172 { 00:08:28.172 "dma_device_id": "system", 00:08:28.172 "dma_device_type": 1 00:08:28.172 }, 00:08:28.172 { 00:08:28.172 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:28.172 "dma_device_type": 2 00:08:28.172 } 00:08:28.172 ], 00:08:28.172 "driver_specific": {} 00:08:28.172 } 00:08:28.172 ] 00:08:28.172 17:29:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:28.172 17:29:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:08:28.172 17:29:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:08:28.172 17:29:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:28.172 17:29:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:28.172 17:29:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:28.172 17:29:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:28.172 17:29:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:28.172 17:29:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:28.172 17:29:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:28.172 17:29:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:28.172 17:29:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:28.172 17:29:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:28.172 17:29:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:28.172 17:29:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:28.172 17:29:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:28.172 17:29:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:28.172 17:29:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:28.172 17:29:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:28.172 17:29:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:28.172 "name": "Existed_Raid", 00:08:28.172 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:28.172 "strip_size_kb": 64, 00:08:28.172 "state": "configuring", 00:08:28.172 "raid_level": "raid0", 00:08:28.172 "superblock": false, 00:08:28.172 "num_base_bdevs": 3, 00:08:28.172 "num_base_bdevs_discovered": 2, 00:08:28.172 "num_base_bdevs_operational": 3, 00:08:28.172 "base_bdevs_list": [ 00:08:28.172 { 00:08:28.172 "name": "BaseBdev1", 00:08:28.172 "uuid": "161b3d93-c1f0-4d49-bb09-2f6568471749", 00:08:28.172 "is_configured": true, 00:08:28.172 "data_offset": 0, 00:08:28.172 "data_size": 65536 00:08:28.172 }, 00:08:28.172 { 00:08:28.172 "name": "BaseBdev2", 00:08:28.172 "uuid": "cb63dfcd-9a45-40ab-a91b-352649ba9fb0", 00:08:28.172 "is_configured": true, 00:08:28.172 "data_offset": 0, 00:08:28.172 "data_size": 65536 00:08:28.172 }, 00:08:28.172 { 00:08:28.172 "name": "BaseBdev3", 00:08:28.172 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:28.172 "is_configured": false, 00:08:28.172 "data_offset": 0, 00:08:28.172 "data_size": 0 00:08:28.172 } 00:08:28.172 ] 00:08:28.172 }' 00:08:28.172 17:29:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:28.172 17:29:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:28.433 17:29:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:08:28.433 17:29:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:28.433 17:29:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:28.433 [2024-11-27 17:29:59.591695] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:28.433 [2024-11-27 17:29:59.591738] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:08:28.433 [2024-11-27 17:29:59.591755] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 196608, blocklen 512 00:08:28.433 [2024-11-27 17:29:59.592091] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:08:28.433 [2024-11-27 17:29:59.592279] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:08:28.433 [2024-11-27 17:29:59.592291] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:08:28.433 [2024-11-27 17:29:59.592524] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:28.433 BaseBdev3 00:08:28.433 17:29:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:28.433 17:29:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:08:28.433 17:29:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:08:28.433 17:29:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:08:28.433 17:29:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:08:28.433 17:29:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:08:28.433 17:29:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:08:28.433 17:29:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:08:28.433 17:29:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:28.433 17:29:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:28.433 17:29:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:28.433 17:29:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:08:28.433 17:29:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:28.433 17:29:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:28.433 [ 00:08:28.433 { 00:08:28.433 "name": "BaseBdev3", 00:08:28.433 "aliases": [ 00:08:28.433 "e8115b5f-4db3-45c7-bc22-5aa658bb05bc" 00:08:28.433 ], 00:08:28.433 "product_name": "Malloc disk", 00:08:28.433 "block_size": 512, 00:08:28.433 "num_blocks": 65536, 00:08:28.433 "uuid": "e8115b5f-4db3-45c7-bc22-5aa658bb05bc", 00:08:28.433 "assigned_rate_limits": { 00:08:28.433 "rw_ios_per_sec": 0, 00:08:28.433 "rw_mbytes_per_sec": 0, 00:08:28.433 "r_mbytes_per_sec": 0, 00:08:28.433 "w_mbytes_per_sec": 0 00:08:28.433 }, 00:08:28.433 "claimed": true, 00:08:28.433 "claim_type": "exclusive_write", 00:08:28.433 "zoned": false, 00:08:28.433 "supported_io_types": { 00:08:28.433 "read": true, 00:08:28.433 "write": true, 00:08:28.694 "unmap": true, 00:08:28.694 "flush": true, 00:08:28.694 "reset": true, 00:08:28.694 "nvme_admin": false, 00:08:28.694 "nvme_io": false, 00:08:28.694 "nvme_io_md": false, 00:08:28.694 "write_zeroes": true, 00:08:28.694 "zcopy": true, 00:08:28.694 "get_zone_info": false, 00:08:28.694 "zone_management": false, 00:08:28.694 "zone_append": false, 00:08:28.694 "compare": false, 00:08:28.694 "compare_and_write": false, 00:08:28.694 "abort": true, 00:08:28.694 "seek_hole": false, 00:08:28.694 "seek_data": false, 00:08:28.694 "copy": true, 00:08:28.694 "nvme_iov_md": false 00:08:28.694 }, 00:08:28.694 "memory_domains": [ 00:08:28.694 { 00:08:28.694 "dma_device_id": "system", 00:08:28.694 "dma_device_type": 1 00:08:28.694 }, 00:08:28.694 { 00:08:28.694 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:28.694 "dma_device_type": 2 00:08:28.694 } 00:08:28.694 ], 00:08:28.694 "driver_specific": {} 00:08:28.694 } 00:08:28.694 ] 00:08:28.694 17:29:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:28.694 17:29:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:08:28.694 17:29:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:08:28.694 17:29:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:28.694 17:29:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid0 64 3 00:08:28.694 17:29:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:28.694 17:29:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:28.694 17:29:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:28.694 17:29:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:28.694 17:29:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:28.694 17:29:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:28.694 17:29:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:28.694 17:29:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:28.694 17:29:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:28.694 17:29:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:28.694 17:29:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:28.694 17:29:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:28.694 17:29:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:28.694 17:29:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:28.694 17:29:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:28.694 "name": "Existed_Raid", 00:08:28.694 "uuid": "240b899c-0249-4ffd-9139-3edd9446ec0d", 00:08:28.694 "strip_size_kb": 64, 00:08:28.694 "state": "online", 00:08:28.694 "raid_level": "raid0", 00:08:28.694 "superblock": false, 00:08:28.694 "num_base_bdevs": 3, 00:08:28.694 "num_base_bdevs_discovered": 3, 00:08:28.694 "num_base_bdevs_operational": 3, 00:08:28.694 "base_bdevs_list": [ 00:08:28.694 { 00:08:28.694 "name": "BaseBdev1", 00:08:28.694 "uuid": "161b3d93-c1f0-4d49-bb09-2f6568471749", 00:08:28.694 "is_configured": true, 00:08:28.694 "data_offset": 0, 00:08:28.694 "data_size": 65536 00:08:28.694 }, 00:08:28.694 { 00:08:28.694 "name": "BaseBdev2", 00:08:28.694 "uuid": "cb63dfcd-9a45-40ab-a91b-352649ba9fb0", 00:08:28.694 "is_configured": true, 00:08:28.694 "data_offset": 0, 00:08:28.694 "data_size": 65536 00:08:28.694 }, 00:08:28.694 { 00:08:28.694 "name": "BaseBdev3", 00:08:28.694 "uuid": "e8115b5f-4db3-45c7-bc22-5aa658bb05bc", 00:08:28.694 "is_configured": true, 00:08:28.694 "data_offset": 0, 00:08:28.694 "data_size": 65536 00:08:28.694 } 00:08:28.694 ] 00:08:28.694 }' 00:08:28.694 17:29:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:28.694 17:29:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:28.954 17:30:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:08:28.954 17:30:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:08:28.954 17:30:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:28.954 17:30:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:28.954 17:30:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:08:28.954 17:30:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:28.954 17:30:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:28.954 17:30:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:08:28.954 17:30:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:28.954 17:30:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:28.954 [2024-11-27 17:30:00.071195] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:28.954 17:30:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:28.954 17:30:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:28.954 "name": "Existed_Raid", 00:08:28.954 "aliases": [ 00:08:28.954 "240b899c-0249-4ffd-9139-3edd9446ec0d" 00:08:28.954 ], 00:08:28.954 "product_name": "Raid Volume", 00:08:28.954 "block_size": 512, 00:08:28.954 "num_blocks": 196608, 00:08:28.954 "uuid": "240b899c-0249-4ffd-9139-3edd9446ec0d", 00:08:28.954 "assigned_rate_limits": { 00:08:28.954 "rw_ios_per_sec": 0, 00:08:28.954 "rw_mbytes_per_sec": 0, 00:08:28.954 "r_mbytes_per_sec": 0, 00:08:28.954 "w_mbytes_per_sec": 0 00:08:28.954 }, 00:08:28.954 "claimed": false, 00:08:28.954 "zoned": false, 00:08:28.954 "supported_io_types": { 00:08:28.954 "read": true, 00:08:28.954 "write": true, 00:08:28.954 "unmap": true, 00:08:28.954 "flush": true, 00:08:28.954 "reset": true, 00:08:28.954 "nvme_admin": false, 00:08:28.954 "nvme_io": false, 00:08:28.954 "nvme_io_md": false, 00:08:28.954 "write_zeroes": true, 00:08:28.954 "zcopy": false, 00:08:28.954 "get_zone_info": false, 00:08:28.954 "zone_management": false, 00:08:28.954 "zone_append": false, 00:08:28.954 "compare": false, 00:08:28.954 "compare_and_write": false, 00:08:28.954 "abort": false, 00:08:28.954 "seek_hole": false, 00:08:28.954 "seek_data": false, 00:08:28.954 "copy": false, 00:08:28.954 "nvme_iov_md": false 00:08:28.954 }, 00:08:28.954 "memory_domains": [ 00:08:28.954 { 00:08:28.954 "dma_device_id": "system", 00:08:28.954 "dma_device_type": 1 00:08:28.954 }, 00:08:28.954 { 00:08:28.954 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:28.954 "dma_device_type": 2 00:08:28.954 }, 00:08:28.954 { 00:08:28.954 "dma_device_id": "system", 00:08:28.954 "dma_device_type": 1 00:08:28.954 }, 00:08:28.954 { 00:08:28.954 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:28.954 "dma_device_type": 2 00:08:28.954 }, 00:08:28.954 { 00:08:28.954 "dma_device_id": "system", 00:08:28.954 "dma_device_type": 1 00:08:28.954 }, 00:08:28.954 { 00:08:28.954 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:28.954 "dma_device_type": 2 00:08:28.954 } 00:08:28.954 ], 00:08:28.954 "driver_specific": { 00:08:28.954 "raid": { 00:08:28.954 "uuid": "240b899c-0249-4ffd-9139-3edd9446ec0d", 00:08:28.954 "strip_size_kb": 64, 00:08:28.954 "state": "online", 00:08:28.954 "raid_level": "raid0", 00:08:28.954 "superblock": false, 00:08:28.954 "num_base_bdevs": 3, 00:08:28.954 "num_base_bdevs_discovered": 3, 00:08:28.954 "num_base_bdevs_operational": 3, 00:08:28.954 "base_bdevs_list": [ 00:08:28.954 { 00:08:28.954 "name": "BaseBdev1", 00:08:28.954 "uuid": "161b3d93-c1f0-4d49-bb09-2f6568471749", 00:08:28.954 "is_configured": true, 00:08:28.954 "data_offset": 0, 00:08:28.954 "data_size": 65536 00:08:28.954 }, 00:08:28.954 { 00:08:28.954 "name": "BaseBdev2", 00:08:28.954 "uuid": "cb63dfcd-9a45-40ab-a91b-352649ba9fb0", 00:08:28.954 "is_configured": true, 00:08:28.954 "data_offset": 0, 00:08:28.954 "data_size": 65536 00:08:28.954 }, 00:08:28.954 { 00:08:28.954 "name": "BaseBdev3", 00:08:28.954 "uuid": "e8115b5f-4db3-45c7-bc22-5aa658bb05bc", 00:08:28.955 "is_configured": true, 00:08:28.955 "data_offset": 0, 00:08:28.955 "data_size": 65536 00:08:28.955 } 00:08:28.955 ] 00:08:28.955 } 00:08:28.955 } 00:08:28.955 }' 00:08:28.955 17:30:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:29.215 17:30:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:08:29.215 BaseBdev2 00:08:29.215 BaseBdev3' 00:08:29.215 17:30:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:29.215 17:30:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:29.215 17:30:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:29.215 17:30:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:29.215 17:30:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:08:29.215 17:30:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:29.215 17:30:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:29.215 17:30:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:29.215 17:30:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:29.215 17:30:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:29.215 17:30:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:29.215 17:30:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:08:29.215 17:30:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:29.215 17:30:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:29.215 17:30:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:29.215 17:30:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:29.215 17:30:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:29.215 17:30:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:29.215 17:30:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:29.215 17:30:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:29.215 17:30:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:08:29.215 17:30:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:29.215 17:30:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:29.215 17:30:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:29.215 17:30:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:29.215 17:30:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:29.215 17:30:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:08:29.215 17:30:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:29.215 17:30:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:29.215 [2024-11-27 17:30:00.318562] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:08:29.215 [2024-11-27 17:30:00.318628] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:29.215 [2024-11-27 17:30:00.318734] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:29.215 17:30:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:29.215 17:30:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:08:29.215 17:30:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy raid0 00:08:29.215 17:30:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:08:29.215 17:30:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # return 1 00:08:29.215 17:30:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:08:29.215 17:30:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline raid0 64 2 00:08:29.215 17:30:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:29.215 17:30:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:08:29.215 17:30:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:29.215 17:30:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:29.215 17:30:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:29.215 17:30:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:29.215 17:30:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:29.215 17:30:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:29.215 17:30:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:29.215 17:30:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:29.215 17:30:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:29.215 17:30:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:29.215 17:30:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:29.215 17:30:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:29.215 17:30:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:29.215 "name": "Existed_Raid", 00:08:29.215 "uuid": "240b899c-0249-4ffd-9139-3edd9446ec0d", 00:08:29.215 "strip_size_kb": 64, 00:08:29.215 "state": "offline", 00:08:29.215 "raid_level": "raid0", 00:08:29.215 "superblock": false, 00:08:29.215 "num_base_bdevs": 3, 00:08:29.215 "num_base_bdevs_discovered": 2, 00:08:29.215 "num_base_bdevs_operational": 2, 00:08:29.215 "base_bdevs_list": [ 00:08:29.215 { 00:08:29.215 "name": null, 00:08:29.215 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:29.215 "is_configured": false, 00:08:29.215 "data_offset": 0, 00:08:29.215 "data_size": 65536 00:08:29.215 }, 00:08:29.215 { 00:08:29.215 "name": "BaseBdev2", 00:08:29.215 "uuid": "cb63dfcd-9a45-40ab-a91b-352649ba9fb0", 00:08:29.215 "is_configured": true, 00:08:29.215 "data_offset": 0, 00:08:29.215 "data_size": 65536 00:08:29.215 }, 00:08:29.215 { 00:08:29.215 "name": "BaseBdev3", 00:08:29.215 "uuid": "e8115b5f-4db3-45c7-bc22-5aa658bb05bc", 00:08:29.215 "is_configured": true, 00:08:29.215 "data_offset": 0, 00:08:29.215 "data_size": 65536 00:08:29.215 } 00:08:29.215 ] 00:08:29.215 }' 00:08:29.215 17:30:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:29.215 17:30:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:29.785 17:30:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:08:29.785 17:30:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:29.785 17:30:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:08:29.785 17:30:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:29.785 17:30:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:29.785 17:30:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:29.785 17:30:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:29.785 17:30:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:08:29.785 17:30:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:08:29.785 17:30:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:08:29.785 17:30:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:29.785 17:30:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:29.785 [2024-11-27 17:30:00.782231] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:08:29.785 17:30:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:29.785 17:30:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:08:29.785 17:30:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:29.785 17:30:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:29.785 17:30:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:08:29.785 17:30:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:29.785 17:30:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:29.785 17:30:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:29.785 17:30:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:08:29.785 17:30:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:08:29.785 17:30:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:08:29.785 17:30:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:29.785 17:30:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:29.785 [2024-11-27 17:30:00.858590] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:08:29.785 [2024-11-27 17:30:00.858686] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:08:29.785 17:30:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:29.785 17:30:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:08:29.785 17:30:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:29.785 17:30:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:29.785 17:30:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:08:29.785 17:30:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:29.785 17:30:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:29.785 17:30:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:29.785 17:30:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:08:29.785 17:30:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:08:29.785 17:30:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 3 -gt 2 ']' 00:08:29.785 17:30:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:08:29.785 17:30:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:08:29.785 17:30:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:08:29.785 17:30:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:29.785 17:30:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:29.785 BaseBdev2 00:08:29.785 17:30:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:29.785 17:30:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:08:29.785 17:30:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:08:29.785 17:30:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:08:29.785 17:30:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:08:29.785 17:30:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:08:29.785 17:30:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:08:29.785 17:30:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:08:29.785 17:30:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:29.785 17:30:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:29.785 17:30:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:29.785 17:30:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:08:29.786 17:30:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:29.786 17:30:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:30.046 [ 00:08:30.046 { 00:08:30.046 "name": "BaseBdev2", 00:08:30.046 "aliases": [ 00:08:30.046 "4dc3811a-9503-4ea8-ac1b-4db2dbeebae0" 00:08:30.046 ], 00:08:30.046 "product_name": "Malloc disk", 00:08:30.046 "block_size": 512, 00:08:30.046 "num_blocks": 65536, 00:08:30.046 "uuid": "4dc3811a-9503-4ea8-ac1b-4db2dbeebae0", 00:08:30.046 "assigned_rate_limits": { 00:08:30.046 "rw_ios_per_sec": 0, 00:08:30.046 "rw_mbytes_per_sec": 0, 00:08:30.046 "r_mbytes_per_sec": 0, 00:08:30.046 "w_mbytes_per_sec": 0 00:08:30.046 }, 00:08:30.046 "claimed": false, 00:08:30.046 "zoned": false, 00:08:30.046 "supported_io_types": { 00:08:30.046 "read": true, 00:08:30.046 "write": true, 00:08:30.046 "unmap": true, 00:08:30.046 "flush": true, 00:08:30.046 "reset": true, 00:08:30.046 "nvme_admin": false, 00:08:30.046 "nvme_io": false, 00:08:30.046 "nvme_io_md": false, 00:08:30.046 "write_zeroes": true, 00:08:30.046 "zcopy": true, 00:08:30.046 "get_zone_info": false, 00:08:30.046 "zone_management": false, 00:08:30.046 "zone_append": false, 00:08:30.046 "compare": false, 00:08:30.046 "compare_and_write": false, 00:08:30.046 "abort": true, 00:08:30.046 "seek_hole": false, 00:08:30.046 "seek_data": false, 00:08:30.046 "copy": true, 00:08:30.046 "nvme_iov_md": false 00:08:30.046 }, 00:08:30.046 "memory_domains": [ 00:08:30.046 { 00:08:30.046 "dma_device_id": "system", 00:08:30.046 "dma_device_type": 1 00:08:30.046 }, 00:08:30.046 { 00:08:30.046 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:30.046 "dma_device_type": 2 00:08:30.046 } 00:08:30.046 ], 00:08:30.046 "driver_specific": {} 00:08:30.046 } 00:08:30.046 ] 00:08:30.046 17:30:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:30.046 17:30:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:08:30.046 17:30:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:08:30.046 17:30:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:08:30.046 17:30:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:08:30.046 17:30:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:30.046 17:30:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:30.046 BaseBdev3 00:08:30.046 17:30:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:30.046 17:30:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:08:30.046 17:30:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:08:30.046 17:30:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:08:30.046 17:30:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:08:30.046 17:30:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:08:30.046 17:30:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:08:30.046 17:30:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:08:30.046 17:30:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:30.046 17:30:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:30.046 17:30:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:30.046 17:30:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:08:30.046 17:30:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:30.046 17:30:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:30.046 [ 00:08:30.046 { 00:08:30.046 "name": "BaseBdev3", 00:08:30.046 "aliases": [ 00:08:30.046 "e7fdff07-1cd0-4387-aa05-006b907df034" 00:08:30.046 ], 00:08:30.046 "product_name": "Malloc disk", 00:08:30.046 "block_size": 512, 00:08:30.046 "num_blocks": 65536, 00:08:30.046 "uuid": "e7fdff07-1cd0-4387-aa05-006b907df034", 00:08:30.046 "assigned_rate_limits": { 00:08:30.046 "rw_ios_per_sec": 0, 00:08:30.046 "rw_mbytes_per_sec": 0, 00:08:30.046 "r_mbytes_per_sec": 0, 00:08:30.046 "w_mbytes_per_sec": 0 00:08:30.046 }, 00:08:30.046 "claimed": false, 00:08:30.046 "zoned": false, 00:08:30.046 "supported_io_types": { 00:08:30.046 "read": true, 00:08:30.046 "write": true, 00:08:30.046 "unmap": true, 00:08:30.046 "flush": true, 00:08:30.046 "reset": true, 00:08:30.046 "nvme_admin": false, 00:08:30.046 "nvme_io": false, 00:08:30.046 "nvme_io_md": false, 00:08:30.046 "write_zeroes": true, 00:08:30.046 "zcopy": true, 00:08:30.046 "get_zone_info": false, 00:08:30.046 "zone_management": false, 00:08:30.046 "zone_append": false, 00:08:30.046 "compare": false, 00:08:30.046 "compare_and_write": false, 00:08:30.046 "abort": true, 00:08:30.046 "seek_hole": false, 00:08:30.046 "seek_data": false, 00:08:30.046 "copy": true, 00:08:30.046 "nvme_iov_md": false 00:08:30.046 }, 00:08:30.046 "memory_domains": [ 00:08:30.046 { 00:08:30.046 "dma_device_id": "system", 00:08:30.046 "dma_device_type": 1 00:08:30.046 }, 00:08:30.046 { 00:08:30.046 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:30.046 "dma_device_type": 2 00:08:30.046 } 00:08:30.046 ], 00:08:30.046 "driver_specific": {} 00:08:30.046 } 00:08:30.046 ] 00:08:30.046 17:30:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:30.046 17:30:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:08:30.046 17:30:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:08:30.046 17:30:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:08:30.046 17:30:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:30.046 17:30:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:30.046 17:30:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:30.046 [2024-11-27 17:30:01.056951] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:30.046 [2024-11-27 17:30:01.057038] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:30.046 [2024-11-27 17:30:01.057098] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:30.046 [2024-11-27 17:30:01.059186] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:30.046 17:30:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:30.046 17:30:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:30.046 17:30:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:30.046 17:30:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:30.046 17:30:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:30.046 17:30:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:30.046 17:30:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:30.046 17:30:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:30.046 17:30:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:30.046 17:30:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:30.046 17:30:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:30.046 17:30:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:30.046 17:30:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:30.046 17:30:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:30.046 17:30:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:30.046 17:30:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:30.046 17:30:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:30.046 "name": "Existed_Raid", 00:08:30.046 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:30.046 "strip_size_kb": 64, 00:08:30.046 "state": "configuring", 00:08:30.046 "raid_level": "raid0", 00:08:30.046 "superblock": false, 00:08:30.046 "num_base_bdevs": 3, 00:08:30.046 "num_base_bdevs_discovered": 2, 00:08:30.046 "num_base_bdevs_operational": 3, 00:08:30.046 "base_bdevs_list": [ 00:08:30.047 { 00:08:30.047 "name": "BaseBdev1", 00:08:30.047 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:30.047 "is_configured": false, 00:08:30.047 "data_offset": 0, 00:08:30.047 "data_size": 0 00:08:30.047 }, 00:08:30.047 { 00:08:30.047 "name": "BaseBdev2", 00:08:30.047 "uuid": "4dc3811a-9503-4ea8-ac1b-4db2dbeebae0", 00:08:30.047 "is_configured": true, 00:08:30.047 "data_offset": 0, 00:08:30.047 "data_size": 65536 00:08:30.047 }, 00:08:30.047 { 00:08:30.047 "name": "BaseBdev3", 00:08:30.047 "uuid": "e7fdff07-1cd0-4387-aa05-006b907df034", 00:08:30.047 "is_configured": true, 00:08:30.047 "data_offset": 0, 00:08:30.047 "data_size": 65536 00:08:30.047 } 00:08:30.047 ] 00:08:30.047 }' 00:08:30.047 17:30:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:30.047 17:30:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:30.306 17:30:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:08:30.306 17:30:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:30.306 17:30:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:30.306 [2024-11-27 17:30:01.456263] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:08:30.306 17:30:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:30.306 17:30:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:30.306 17:30:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:30.306 17:30:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:30.306 17:30:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:30.306 17:30:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:30.306 17:30:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:30.306 17:30:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:30.306 17:30:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:30.306 17:30:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:30.306 17:30:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:30.306 17:30:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:30.306 17:30:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:30.306 17:30:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:30.306 17:30:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:30.306 17:30:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:30.565 17:30:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:30.565 "name": "Existed_Raid", 00:08:30.565 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:30.565 "strip_size_kb": 64, 00:08:30.565 "state": "configuring", 00:08:30.565 "raid_level": "raid0", 00:08:30.565 "superblock": false, 00:08:30.565 "num_base_bdevs": 3, 00:08:30.565 "num_base_bdevs_discovered": 1, 00:08:30.565 "num_base_bdevs_operational": 3, 00:08:30.565 "base_bdevs_list": [ 00:08:30.565 { 00:08:30.565 "name": "BaseBdev1", 00:08:30.565 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:30.565 "is_configured": false, 00:08:30.565 "data_offset": 0, 00:08:30.565 "data_size": 0 00:08:30.565 }, 00:08:30.565 { 00:08:30.565 "name": null, 00:08:30.565 "uuid": "4dc3811a-9503-4ea8-ac1b-4db2dbeebae0", 00:08:30.565 "is_configured": false, 00:08:30.565 "data_offset": 0, 00:08:30.565 "data_size": 65536 00:08:30.565 }, 00:08:30.565 { 00:08:30.565 "name": "BaseBdev3", 00:08:30.565 "uuid": "e7fdff07-1cd0-4387-aa05-006b907df034", 00:08:30.565 "is_configured": true, 00:08:30.565 "data_offset": 0, 00:08:30.565 "data_size": 65536 00:08:30.565 } 00:08:30.565 ] 00:08:30.565 }' 00:08:30.565 17:30:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:30.565 17:30:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:30.826 17:30:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:08:30.826 17:30:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:30.826 17:30:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:30.826 17:30:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:30.826 17:30:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:30.826 17:30:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:08:30.826 17:30:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:08:30.826 17:30:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:30.826 17:30:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:30.826 BaseBdev1 00:08:30.826 [2024-11-27 17:30:01.976058] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:30.826 17:30:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:30.826 17:30:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:08:30.826 17:30:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:08:30.826 17:30:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:08:30.826 17:30:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:08:30.826 17:30:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:08:30.826 17:30:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:08:30.826 17:30:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:08:30.826 17:30:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:30.826 17:30:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:30.826 17:30:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:30.826 17:30:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:08:30.826 17:30:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:30.826 17:30:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:30.826 [ 00:08:30.826 { 00:08:30.826 "name": "BaseBdev1", 00:08:30.826 "aliases": [ 00:08:30.826 "d415eb05-bb68-405d-a232-a690988bf527" 00:08:30.826 ], 00:08:30.826 "product_name": "Malloc disk", 00:08:30.826 "block_size": 512, 00:08:30.826 "num_blocks": 65536, 00:08:30.826 "uuid": "d415eb05-bb68-405d-a232-a690988bf527", 00:08:30.826 "assigned_rate_limits": { 00:08:30.826 "rw_ios_per_sec": 0, 00:08:30.826 "rw_mbytes_per_sec": 0, 00:08:30.826 "r_mbytes_per_sec": 0, 00:08:30.826 "w_mbytes_per_sec": 0 00:08:30.826 }, 00:08:30.826 "claimed": true, 00:08:30.826 "claim_type": "exclusive_write", 00:08:30.826 "zoned": false, 00:08:30.826 "supported_io_types": { 00:08:30.826 "read": true, 00:08:30.826 "write": true, 00:08:30.826 "unmap": true, 00:08:30.826 "flush": true, 00:08:30.826 "reset": true, 00:08:30.826 "nvme_admin": false, 00:08:30.826 "nvme_io": false, 00:08:30.826 "nvme_io_md": false, 00:08:30.826 "write_zeroes": true, 00:08:30.826 "zcopy": true, 00:08:30.826 "get_zone_info": false, 00:08:30.826 "zone_management": false, 00:08:30.826 "zone_append": false, 00:08:30.826 "compare": false, 00:08:30.826 "compare_and_write": false, 00:08:30.826 "abort": true, 00:08:30.826 "seek_hole": false, 00:08:30.826 "seek_data": false, 00:08:30.826 "copy": true, 00:08:30.826 "nvme_iov_md": false 00:08:30.826 }, 00:08:30.826 "memory_domains": [ 00:08:30.826 { 00:08:30.826 "dma_device_id": "system", 00:08:30.826 "dma_device_type": 1 00:08:30.826 }, 00:08:30.826 { 00:08:30.826 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:30.826 "dma_device_type": 2 00:08:30.826 } 00:08:30.826 ], 00:08:30.826 "driver_specific": {} 00:08:30.826 } 00:08:30.826 ] 00:08:30.826 17:30:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:30.826 17:30:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:08:30.826 17:30:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:30.826 17:30:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:30.826 17:30:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:30.826 17:30:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:30.826 17:30:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:30.826 17:30:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:31.086 17:30:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:31.086 17:30:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:31.086 17:30:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:31.086 17:30:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:31.086 17:30:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:31.086 17:30:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:31.086 17:30:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:31.086 17:30:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:31.086 17:30:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:31.086 17:30:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:31.086 "name": "Existed_Raid", 00:08:31.086 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:31.086 "strip_size_kb": 64, 00:08:31.086 "state": "configuring", 00:08:31.086 "raid_level": "raid0", 00:08:31.086 "superblock": false, 00:08:31.086 "num_base_bdevs": 3, 00:08:31.086 "num_base_bdevs_discovered": 2, 00:08:31.086 "num_base_bdevs_operational": 3, 00:08:31.086 "base_bdevs_list": [ 00:08:31.086 { 00:08:31.086 "name": "BaseBdev1", 00:08:31.086 "uuid": "d415eb05-bb68-405d-a232-a690988bf527", 00:08:31.086 "is_configured": true, 00:08:31.086 "data_offset": 0, 00:08:31.086 "data_size": 65536 00:08:31.086 }, 00:08:31.086 { 00:08:31.086 "name": null, 00:08:31.086 "uuid": "4dc3811a-9503-4ea8-ac1b-4db2dbeebae0", 00:08:31.086 "is_configured": false, 00:08:31.086 "data_offset": 0, 00:08:31.086 "data_size": 65536 00:08:31.086 }, 00:08:31.087 { 00:08:31.087 "name": "BaseBdev3", 00:08:31.087 "uuid": "e7fdff07-1cd0-4387-aa05-006b907df034", 00:08:31.087 "is_configured": true, 00:08:31.087 "data_offset": 0, 00:08:31.087 "data_size": 65536 00:08:31.087 } 00:08:31.087 ] 00:08:31.087 }' 00:08:31.087 17:30:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:31.087 17:30:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:31.346 17:30:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:31.346 17:30:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:08:31.347 17:30:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:31.347 17:30:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:31.347 17:30:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:31.347 17:30:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:08:31.347 17:30:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:08:31.347 17:30:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:31.347 17:30:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:31.347 [2024-11-27 17:30:02.475226] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:08:31.347 17:30:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:31.347 17:30:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:31.347 17:30:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:31.347 17:30:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:31.347 17:30:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:31.347 17:30:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:31.347 17:30:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:31.347 17:30:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:31.347 17:30:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:31.347 17:30:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:31.347 17:30:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:31.347 17:30:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:31.347 17:30:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:31.347 17:30:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:31.347 17:30:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:31.347 17:30:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:31.347 17:30:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:31.347 "name": "Existed_Raid", 00:08:31.347 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:31.347 "strip_size_kb": 64, 00:08:31.347 "state": "configuring", 00:08:31.347 "raid_level": "raid0", 00:08:31.347 "superblock": false, 00:08:31.347 "num_base_bdevs": 3, 00:08:31.347 "num_base_bdevs_discovered": 1, 00:08:31.347 "num_base_bdevs_operational": 3, 00:08:31.347 "base_bdevs_list": [ 00:08:31.347 { 00:08:31.347 "name": "BaseBdev1", 00:08:31.347 "uuid": "d415eb05-bb68-405d-a232-a690988bf527", 00:08:31.347 "is_configured": true, 00:08:31.347 "data_offset": 0, 00:08:31.347 "data_size": 65536 00:08:31.347 }, 00:08:31.347 { 00:08:31.347 "name": null, 00:08:31.347 "uuid": "4dc3811a-9503-4ea8-ac1b-4db2dbeebae0", 00:08:31.347 "is_configured": false, 00:08:31.347 "data_offset": 0, 00:08:31.347 "data_size": 65536 00:08:31.347 }, 00:08:31.347 { 00:08:31.347 "name": null, 00:08:31.347 "uuid": "e7fdff07-1cd0-4387-aa05-006b907df034", 00:08:31.347 "is_configured": false, 00:08:31.347 "data_offset": 0, 00:08:31.347 "data_size": 65536 00:08:31.347 } 00:08:31.347 ] 00:08:31.347 }' 00:08:31.347 17:30:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:31.347 17:30:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:31.916 17:30:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:08:31.916 17:30:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:31.916 17:30:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:31.916 17:30:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:31.916 17:30:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:31.916 17:30:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:08:31.916 17:30:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:08:31.916 17:30:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:31.916 17:30:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:31.916 [2024-11-27 17:30:02.930420] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:31.916 17:30:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:31.916 17:30:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:31.916 17:30:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:31.916 17:30:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:31.916 17:30:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:31.916 17:30:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:31.916 17:30:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:31.916 17:30:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:31.916 17:30:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:31.916 17:30:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:31.916 17:30:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:31.916 17:30:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:31.916 17:30:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:31.916 17:30:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:31.916 17:30:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:31.916 17:30:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:31.916 17:30:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:31.916 "name": "Existed_Raid", 00:08:31.916 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:31.916 "strip_size_kb": 64, 00:08:31.916 "state": "configuring", 00:08:31.916 "raid_level": "raid0", 00:08:31.916 "superblock": false, 00:08:31.916 "num_base_bdevs": 3, 00:08:31.916 "num_base_bdevs_discovered": 2, 00:08:31.916 "num_base_bdevs_operational": 3, 00:08:31.916 "base_bdevs_list": [ 00:08:31.916 { 00:08:31.916 "name": "BaseBdev1", 00:08:31.916 "uuid": "d415eb05-bb68-405d-a232-a690988bf527", 00:08:31.916 "is_configured": true, 00:08:31.916 "data_offset": 0, 00:08:31.916 "data_size": 65536 00:08:31.916 }, 00:08:31.916 { 00:08:31.916 "name": null, 00:08:31.916 "uuid": "4dc3811a-9503-4ea8-ac1b-4db2dbeebae0", 00:08:31.916 "is_configured": false, 00:08:31.916 "data_offset": 0, 00:08:31.916 "data_size": 65536 00:08:31.916 }, 00:08:31.916 { 00:08:31.916 "name": "BaseBdev3", 00:08:31.916 "uuid": "e7fdff07-1cd0-4387-aa05-006b907df034", 00:08:31.916 "is_configured": true, 00:08:31.916 "data_offset": 0, 00:08:31.916 "data_size": 65536 00:08:31.916 } 00:08:31.916 ] 00:08:31.916 }' 00:08:31.916 17:30:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:31.916 17:30:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:32.486 17:30:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:32.486 17:30:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:32.486 17:30:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:32.486 17:30:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:08:32.486 17:30:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:32.486 17:30:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:08:32.486 17:30:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:08:32.486 17:30:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:32.486 17:30:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:32.486 [2024-11-27 17:30:03.417612] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:08:32.486 17:30:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:32.486 17:30:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:32.486 17:30:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:32.486 17:30:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:32.486 17:30:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:32.486 17:30:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:32.486 17:30:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:32.486 17:30:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:32.486 17:30:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:32.486 17:30:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:32.486 17:30:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:32.486 17:30:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:32.486 17:30:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:32.486 17:30:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:32.486 17:30:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:32.486 17:30:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:32.486 17:30:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:32.486 "name": "Existed_Raid", 00:08:32.486 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:32.486 "strip_size_kb": 64, 00:08:32.486 "state": "configuring", 00:08:32.486 "raid_level": "raid0", 00:08:32.486 "superblock": false, 00:08:32.486 "num_base_bdevs": 3, 00:08:32.486 "num_base_bdevs_discovered": 1, 00:08:32.486 "num_base_bdevs_operational": 3, 00:08:32.486 "base_bdevs_list": [ 00:08:32.486 { 00:08:32.486 "name": null, 00:08:32.486 "uuid": "d415eb05-bb68-405d-a232-a690988bf527", 00:08:32.486 "is_configured": false, 00:08:32.486 "data_offset": 0, 00:08:32.486 "data_size": 65536 00:08:32.486 }, 00:08:32.486 { 00:08:32.486 "name": null, 00:08:32.486 "uuid": "4dc3811a-9503-4ea8-ac1b-4db2dbeebae0", 00:08:32.486 "is_configured": false, 00:08:32.486 "data_offset": 0, 00:08:32.486 "data_size": 65536 00:08:32.486 }, 00:08:32.486 { 00:08:32.486 "name": "BaseBdev3", 00:08:32.486 "uuid": "e7fdff07-1cd0-4387-aa05-006b907df034", 00:08:32.486 "is_configured": true, 00:08:32.486 "data_offset": 0, 00:08:32.486 "data_size": 65536 00:08:32.486 } 00:08:32.486 ] 00:08:32.486 }' 00:08:32.486 17:30:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:32.486 17:30:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:32.746 17:30:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:32.746 17:30:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:08:32.746 17:30:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:32.746 17:30:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:32.746 17:30:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:32.746 17:30:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:08:32.746 17:30:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:08:32.746 17:30:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:32.746 17:30:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:32.746 [2024-11-27 17:30:03.912289] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:32.746 17:30:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:32.746 17:30:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:32.746 17:30:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:32.746 17:30:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:32.746 17:30:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:32.746 17:30:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:32.746 17:30:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:32.746 17:30:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:32.746 17:30:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:32.746 17:30:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:32.746 17:30:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:32.746 17:30:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:32.746 17:30:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:32.746 17:30:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:32.746 17:30:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:33.006 17:30:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:33.006 17:30:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:33.006 "name": "Existed_Raid", 00:08:33.006 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:33.006 "strip_size_kb": 64, 00:08:33.006 "state": "configuring", 00:08:33.006 "raid_level": "raid0", 00:08:33.006 "superblock": false, 00:08:33.006 "num_base_bdevs": 3, 00:08:33.006 "num_base_bdevs_discovered": 2, 00:08:33.006 "num_base_bdevs_operational": 3, 00:08:33.006 "base_bdevs_list": [ 00:08:33.006 { 00:08:33.006 "name": null, 00:08:33.006 "uuid": "d415eb05-bb68-405d-a232-a690988bf527", 00:08:33.006 "is_configured": false, 00:08:33.006 "data_offset": 0, 00:08:33.006 "data_size": 65536 00:08:33.006 }, 00:08:33.006 { 00:08:33.006 "name": "BaseBdev2", 00:08:33.006 "uuid": "4dc3811a-9503-4ea8-ac1b-4db2dbeebae0", 00:08:33.006 "is_configured": true, 00:08:33.006 "data_offset": 0, 00:08:33.006 "data_size": 65536 00:08:33.006 }, 00:08:33.006 { 00:08:33.006 "name": "BaseBdev3", 00:08:33.006 "uuid": "e7fdff07-1cd0-4387-aa05-006b907df034", 00:08:33.006 "is_configured": true, 00:08:33.006 "data_offset": 0, 00:08:33.006 "data_size": 65536 00:08:33.006 } 00:08:33.006 ] 00:08:33.006 }' 00:08:33.006 17:30:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:33.006 17:30:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:33.266 17:30:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:33.266 17:30:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:33.266 17:30:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:33.266 17:30:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:08:33.266 17:30:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:33.266 17:30:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:08:33.266 17:30:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:33.266 17:30:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:33.266 17:30:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:33.266 17:30:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:08:33.266 17:30:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:33.526 17:30:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u d415eb05-bb68-405d-a232-a690988bf527 00:08:33.526 17:30:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:33.526 17:30:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:33.526 [2024-11-27 17:30:04.495998] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:08:33.526 [2024-11-27 17:30:04.496040] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:08:33.526 [2024-11-27 17:30:04.496050] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 196608, blocklen 512 00:08:33.526 [2024-11-27 17:30:04.496348] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002870 00:08:33.526 [2024-11-27 17:30:04.496505] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:08:33.526 [2024-11-27 17:30:04.496520] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001c80 00:08:33.526 [2024-11-27 17:30:04.496719] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:33.526 NewBaseBdev 00:08:33.526 17:30:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:33.526 17:30:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:08:33.526 17:30:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=NewBaseBdev 00:08:33.526 17:30:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:08:33.526 17:30:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:08:33.526 17:30:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:08:33.526 17:30:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:08:33.526 17:30:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:08:33.526 17:30:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:33.526 17:30:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:33.526 17:30:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:33.526 17:30:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:08:33.526 17:30:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:33.526 17:30:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:33.526 [ 00:08:33.526 { 00:08:33.526 "name": "NewBaseBdev", 00:08:33.526 "aliases": [ 00:08:33.526 "d415eb05-bb68-405d-a232-a690988bf527" 00:08:33.526 ], 00:08:33.526 "product_name": "Malloc disk", 00:08:33.526 "block_size": 512, 00:08:33.526 "num_blocks": 65536, 00:08:33.526 "uuid": "d415eb05-bb68-405d-a232-a690988bf527", 00:08:33.526 "assigned_rate_limits": { 00:08:33.526 "rw_ios_per_sec": 0, 00:08:33.526 "rw_mbytes_per_sec": 0, 00:08:33.526 "r_mbytes_per_sec": 0, 00:08:33.526 "w_mbytes_per_sec": 0 00:08:33.526 }, 00:08:33.526 "claimed": true, 00:08:33.526 "claim_type": "exclusive_write", 00:08:33.526 "zoned": false, 00:08:33.526 "supported_io_types": { 00:08:33.526 "read": true, 00:08:33.526 "write": true, 00:08:33.526 "unmap": true, 00:08:33.526 "flush": true, 00:08:33.526 "reset": true, 00:08:33.526 "nvme_admin": false, 00:08:33.526 "nvme_io": false, 00:08:33.526 "nvme_io_md": false, 00:08:33.526 "write_zeroes": true, 00:08:33.526 "zcopy": true, 00:08:33.526 "get_zone_info": false, 00:08:33.526 "zone_management": false, 00:08:33.526 "zone_append": false, 00:08:33.526 "compare": false, 00:08:33.526 "compare_and_write": false, 00:08:33.526 "abort": true, 00:08:33.526 "seek_hole": false, 00:08:33.526 "seek_data": false, 00:08:33.526 "copy": true, 00:08:33.526 "nvme_iov_md": false 00:08:33.526 }, 00:08:33.526 "memory_domains": [ 00:08:33.526 { 00:08:33.526 "dma_device_id": "system", 00:08:33.526 "dma_device_type": 1 00:08:33.526 }, 00:08:33.526 { 00:08:33.526 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:33.526 "dma_device_type": 2 00:08:33.526 } 00:08:33.526 ], 00:08:33.526 "driver_specific": {} 00:08:33.526 } 00:08:33.526 ] 00:08:33.526 17:30:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:33.526 17:30:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:08:33.526 17:30:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid0 64 3 00:08:33.526 17:30:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:33.526 17:30:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:33.526 17:30:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:33.526 17:30:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:33.526 17:30:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:33.526 17:30:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:33.526 17:30:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:33.526 17:30:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:33.527 17:30:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:33.527 17:30:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:33.527 17:30:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:33.527 17:30:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:33.527 17:30:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:33.527 17:30:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:33.527 17:30:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:33.527 "name": "Existed_Raid", 00:08:33.527 "uuid": "2dcc2c89-4c8d-41ba-9cfe-d1335f83a4c9", 00:08:33.527 "strip_size_kb": 64, 00:08:33.527 "state": "online", 00:08:33.527 "raid_level": "raid0", 00:08:33.527 "superblock": false, 00:08:33.527 "num_base_bdevs": 3, 00:08:33.527 "num_base_bdevs_discovered": 3, 00:08:33.527 "num_base_bdevs_operational": 3, 00:08:33.527 "base_bdevs_list": [ 00:08:33.527 { 00:08:33.527 "name": "NewBaseBdev", 00:08:33.527 "uuid": "d415eb05-bb68-405d-a232-a690988bf527", 00:08:33.527 "is_configured": true, 00:08:33.527 "data_offset": 0, 00:08:33.527 "data_size": 65536 00:08:33.527 }, 00:08:33.527 { 00:08:33.527 "name": "BaseBdev2", 00:08:33.527 "uuid": "4dc3811a-9503-4ea8-ac1b-4db2dbeebae0", 00:08:33.527 "is_configured": true, 00:08:33.527 "data_offset": 0, 00:08:33.527 "data_size": 65536 00:08:33.527 }, 00:08:33.527 { 00:08:33.527 "name": "BaseBdev3", 00:08:33.527 "uuid": "e7fdff07-1cd0-4387-aa05-006b907df034", 00:08:33.527 "is_configured": true, 00:08:33.527 "data_offset": 0, 00:08:33.527 "data_size": 65536 00:08:33.527 } 00:08:33.527 ] 00:08:33.527 }' 00:08:33.527 17:30:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:33.527 17:30:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:33.787 17:30:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:08:33.787 17:30:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:08:33.787 17:30:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:33.787 17:30:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:33.787 17:30:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:08:33.787 17:30:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:33.787 17:30:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:08:33.787 17:30:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:33.787 17:30:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:33.787 17:30:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:33.787 [2024-11-27 17:30:04.955542] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:33.787 17:30:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:34.046 17:30:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:34.046 "name": "Existed_Raid", 00:08:34.046 "aliases": [ 00:08:34.046 "2dcc2c89-4c8d-41ba-9cfe-d1335f83a4c9" 00:08:34.046 ], 00:08:34.046 "product_name": "Raid Volume", 00:08:34.046 "block_size": 512, 00:08:34.046 "num_blocks": 196608, 00:08:34.046 "uuid": "2dcc2c89-4c8d-41ba-9cfe-d1335f83a4c9", 00:08:34.046 "assigned_rate_limits": { 00:08:34.047 "rw_ios_per_sec": 0, 00:08:34.047 "rw_mbytes_per_sec": 0, 00:08:34.047 "r_mbytes_per_sec": 0, 00:08:34.047 "w_mbytes_per_sec": 0 00:08:34.047 }, 00:08:34.047 "claimed": false, 00:08:34.047 "zoned": false, 00:08:34.047 "supported_io_types": { 00:08:34.047 "read": true, 00:08:34.047 "write": true, 00:08:34.047 "unmap": true, 00:08:34.047 "flush": true, 00:08:34.047 "reset": true, 00:08:34.047 "nvme_admin": false, 00:08:34.047 "nvme_io": false, 00:08:34.047 "nvme_io_md": false, 00:08:34.047 "write_zeroes": true, 00:08:34.047 "zcopy": false, 00:08:34.047 "get_zone_info": false, 00:08:34.047 "zone_management": false, 00:08:34.047 "zone_append": false, 00:08:34.047 "compare": false, 00:08:34.047 "compare_and_write": false, 00:08:34.047 "abort": false, 00:08:34.047 "seek_hole": false, 00:08:34.047 "seek_data": false, 00:08:34.047 "copy": false, 00:08:34.047 "nvme_iov_md": false 00:08:34.047 }, 00:08:34.047 "memory_domains": [ 00:08:34.047 { 00:08:34.047 "dma_device_id": "system", 00:08:34.047 "dma_device_type": 1 00:08:34.047 }, 00:08:34.047 { 00:08:34.047 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:34.047 "dma_device_type": 2 00:08:34.047 }, 00:08:34.047 { 00:08:34.047 "dma_device_id": "system", 00:08:34.047 "dma_device_type": 1 00:08:34.047 }, 00:08:34.047 { 00:08:34.047 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:34.047 "dma_device_type": 2 00:08:34.047 }, 00:08:34.047 { 00:08:34.047 "dma_device_id": "system", 00:08:34.047 "dma_device_type": 1 00:08:34.047 }, 00:08:34.047 { 00:08:34.047 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:34.047 "dma_device_type": 2 00:08:34.047 } 00:08:34.047 ], 00:08:34.047 "driver_specific": { 00:08:34.047 "raid": { 00:08:34.047 "uuid": "2dcc2c89-4c8d-41ba-9cfe-d1335f83a4c9", 00:08:34.047 "strip_size_kb": 64, 00:08:34.047 "state": "online", 00:08:34.047 "raid_level": "raid0", 00:08:34.047 "superblock": false, 00:08:34.047 "num_base_bdevs": 3, 00:08:34.047 "num_base_bdevs_discovered": 3, 00:08:34.047 "num_base_bdevs_operational": 3, 00:08:34.047 "base_bdevs_list": [ 00:08:34.047 { 00:08:34.047 "name": "NewBaseBdev", 00:08:34.047 "uuid": "d415eb05-bb68-405d-a232-a690988bf527", 00:08:34.047 "is_configured": true, 00:08:34.047 "data_offset": 0, 00:08:34.047 "data_size": 65536 00:08:34.047 }, 00:08:34.047 { 00:08:34.047 "name": "BaseBdev2", 00:08:34.047 "uuid": "4dc3811a-9503-4ea8-ac1b-4db2dbeebae0", 00:08:34.047 "is_configured": true, 00:08:34.047 "data_offset": 0, 00:08:34.047 "data_size": 65536 00:08:34.047 }, 00:08:34.047 { 00:08:34.047 "name": "BaseBdev3", 00:08:34.047 "uuid": "e7fdff07-1cd0-4387-aa05-006b907df034", 00:08:34.047 "is_configured": true, 00:08:34.047 "data_offset": 0, 00:08:34.047 "data_size": 65536 00:08:34.047 } 00:08:34.047 ] 00:08:34.047 } 00:08:34.047 } 00:08:34.047 }' 00:08:34.047 17:30:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:34.047 17:30:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:08:34.047 BaseBdev2 00:08:34.047 BaseBdev3' 00:08:34.047 17:30:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:34.047 17:30:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:34.047 17:30:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:34.047 17:30:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:34.047 17:30:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:08:34.047 17:30:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:34.047 17:30:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:34.047 17:30:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:34.047 17:30:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:34.047 17:30:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:34.047 17:30:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:34.047 17:30:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:08:34.047 17:30:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:34.047 17:30:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:34.047 17:30:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:34.047 17:30:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:34.047 17:30:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:34.047 17:30:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:34.047 17:30:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:34.047 17:30:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:08:34.047 17:30:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:34.047 17:30:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:34.047 17:30:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:34.047 17:30:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:34.047 17:30:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:34.047 17:30:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:34.047 17:30:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:34.047 17:30:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:34.047 17:30:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:34.047 [2024-11-27 17:30:05.190851] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:34.047 [2024-11-27 17:30:05.190916] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:34.047 [2024-11-27 17:30:05.191009] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:34.047 [2024-11-27 17:30:05.191093] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:34.047 [2024-11-27 17:30:05.191153] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name Existed_Raid, state offline 00:08:34.047 17:30:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:34.047 17:30:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 74850 00:08:34.047 17:30:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@950 -- # '[' -z 74850 ']' 00:08:34.048 17:30:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # kill -0 74850 00:08:34.048 17:30:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@955 -- # uname 00:08:34.048 17:30:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:08:34.048 17:30:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 74850 00:08:34.308 killing process with pid 74850 00:08:34.308 17:30:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:08:34.308 17:30:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:08:34.308 17:30:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 74850' 00:08:34.308 17:30:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@969 -- # kill 74850 00:08:34.308 [2024-11-27 17:30:05.238865] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:34.308 17:30:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@974 -- # wait 74850 00:08:34.308 [2024-11-27 17:30:05.297752] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:34.568 ************************************ 00:08:34.568 END TEST raid_state_function_test 00:08:34.568 ************************************ 00:08:34.568 17:30:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:08:34.568 00:08:34.568 real 0m8.994s 00:08:34.568 user 0m15.078s 00:08:34.568 sys 0m1.853s 00:08:34.568 17:30:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:08:34.568 17:30:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:34.568 17:30:05 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test raid0 3 true 00:08:34.568 17:30:05 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:08:34.568 17:30:05 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:08:34.568 17:30:05 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:34.568 ************************************ 00:08:34.568 START TEST raid_state_function_test_sb 00:08:34.568 ************************************ 00:08:34.568 17:30:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1125 -- # raid_state_function_test raid0 3 true 00:08:34.568 17:30:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=raid0 00:08:34.568 17:30:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=3 00:08:34.568 17:30:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:08:34.568 17:30:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:08:34.568 17:30:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:08:34.568 17:30:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:34.568 17:30:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:08:34.568 17:30:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:34.568 17:30:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:34.568 17:30:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:08:34.568 17:30:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:34.568 17:30:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:34.568 17:30:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:08:34.568 17:30:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:34.568 17:30:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:34.568 17:30:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:08:34.568 17:30:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:08:34.568 17:30:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:08:34.568 17:30:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:08:34.568 17:30:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:08:34.568 17:30:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:08:34.568 17:30:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' raid0 '!=' raid1 ']' 00:08:34.568 17:30:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:08:34.568 17:30:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:08:34.568 17:30:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:08:34.568 17:30:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:08:34.568 17:30:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=75460 00:08:34.568 17:30:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:08:34.568 17:30:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 75460' 00:08:34.568 Process raid pid: 75460 00:08:34.568 17:30:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 75460 00:08:34.568 17:30:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@831 -- # '[' -z 75460 ']' 00:08:34.568 17:30:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:34.568 17:30:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@836 -- # local max_retries=100 00:08:34.568 17:30:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:34.568 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:34.568 17:30:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # xtrace_disable 00:08:34.568 17:30:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:34.828 [2024-11-27 17:30:05.831569] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:08:34.828 [2024-11-27 17:30:05.831788] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:08:34.828 [2024-11-27 17:30:05.978023] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:35.088 [2024-11-27 17:30:06.046518] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:08:35.088 [2024-11-27 17:30:06.122327] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:35.088 [2024-11-27 17:30:06.122367] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:35.658 17:30:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:08:35.658 17:30:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@864 -- # return 0 00:08:35.658 17:30:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:35.658 17:30:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:35.658 17:30:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:35.658 [2024-11-27 17:30:06.661270] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:35.658 [2024-11-27 17:30:06.661387] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:35.658 [2024-11-27 17:30:06.661432] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:35.658 [2024-11-27 17:30:06.661457] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:35.658 [2024-11-27 17:30:06.661475] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:08:35.658 [2024-11-27 17:30:06.661498] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:08:35.658 17:30:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:35.658 17:30:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:35.658 17:30:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:35.658 17:30:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:35.658 17:30:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:35.658 17:30:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:35.658 17:30:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:35.658 17:30:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:35.658 17:30:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:35.658 17:30:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:35.658 17:30:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:35.658 17:30:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:35.658 17:30:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:35.658 17:30:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:35.658 17:30:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:35.658 17:30:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:35.658 17:30:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:35.658 "name": "Existed_Raid", 00:08:35.658 "uuid": "7abfe33f-746d-4b44-acd7-1e811fed6e0e", 00:08:35.658 "strip_size_kb": 64, 00:08:35.658 "state": "configuring", 00:08:35.658 "raid_level": "raid0", 00:08:35.658 "superblock": true, 00:08:35.658 "num_base_bdevs": 3, 00:08:35.658 "num_base_bdevs_discovered": 0, 00:08:35.658 "num_base_bdevs_operational": 3, 00:08:35.658 "base_bdevs_list": [ 00:08:35.658 { 00:08:35.658 "name": "BaseBdev1", 00:08:35.658 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:35.658 "is_configured": false, 00:08:35.658 "data_offset": 0, 00:08:35.658 "data_size": 0 00:08:35.658 }, 00:08:35.658 { 00:08:35.658 "name": "BaseBdev2", 00:08:35.658 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:35.658 "is_configured": false, 00:08:35.658 "data_offset": 0, 00:08:35.658 "data_size": 0 00:08:35.658 }, 00:08:35.658 { 00:08:35.658 "name": "BaseBdev3", 00:08:35.658 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:35.658 "is_configured": false, 00:08:35.658 "data_offset": 0, 00:08:35.658 "data_size": 0 00:08:35.658 } 00:08:35.658 ] 00:08:35.658 }' 00:08:35.658 17:30:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:35.658 17:30:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:35.917 17:30:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:35.917 17:30:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:35.917 17:30:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:35.917 [2024-11-27 17:30:07.104341] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:35.917 [2024-11-27 17:30:07.104429] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:08:36.186 17:30:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:36.186 17:30:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:36.186 17:30:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:36.186 17:30:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:36.186 [2024-11-27 17:30:07.116348] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:36.186 [2024-11-27 17:30:07.116426] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:36.186 [2024-11-27 17:30:07.116452] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:36.186 [2024-11-27 17:30:07.116490] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:36.186 [2024-11-27 17:30:07.116498] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:08:36.186 [2024-11-27 17:30:07.116507] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:08:36.186 17:30:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:36.186 17:30:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:08:36.186 17:30:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:36.186 17:30:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:36.186 [2024-11-27 17:30:07.143188] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:36.186 BaseBdev1 00:08:36.186 17:30:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:36.186 17:30:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:08:36.186 17:30:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:08:36.186 17:30:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:08:36.186 17:30:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:08:36.186 17:30:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:08:36.186 17:30:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:08:36.186 17:30:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:08:36.186 17:30:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:36.186 17:30:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:36.186 17:30:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:36.186 17:30:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:08:36.186 17:30:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:36.186 17:30:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:36.186 [ 00:08:36.186 { 00:08:36.186 "name": "BaseBdev1", 00:08:36.186 "aliases": [ 00:08:36.186 "05b9b550-d8cd-4d33-b4e8-3589926c160f" 00:08:36.186 ], 00:08:36.186 "product_name": "Malloc disk", 00:08:36.186 "block_size": 512, 00:08:36.186 "num_blocks": 65536, 00:08:36.186 "uuid": "05b9b550-d8cd-4d33-b4e8-3589926c160f", 00:08:36.186 "assigned_rate_limits": { 00:08:36.186 "rw_ios_per_sec": 0, 00:08:36.186 "rw_mbytes_per_sec": 0, 00:08:36.186 "r_mbytes_per_sec": 0, 00:08:36.186 "w_mbytes_per_sec": 0 00:08:36.186 }, 00:08:36.186 "claimed": true, 00:08:36.186 "claim_type": "exclusive_write", 00:08:36.186 "zoned": false, 00:08:36.186 "supported_io_types": { 00:08:36.186 "read": true, 00:08:36.186 "write": true, 00:08:36.186 "unmap": true, 00:08:36.186 "flush": true, 00:08:36.186 "reset": true, 00:08:36.186 "nvme_admin": false, 00:08:36.186 "nvme_io": false, 00:08:36.186 "nvme_io_md": false, 00:08:36.186 "write_zeroes": true, 00:08:36.186 "zcopy": true, 00:08:36.186 "get_zone_info": false, 00:08:36.186 "zone_management": false, 00:08:36.186 "zone_append": false, 00:08:36.186 "compare": false, 00:08:36.186 "compare_and_write": false, 00:08:36.186 "abort": true, 00:08:36.186 "seek_hole": false, 00:08:36.186 "seek_data": false, 00:08:36.186 "copy": true, 00:08:36.186 "nvme_iov_md": false 00:08:36.186 }, 00:08:36.186 "memory_domains": [ 00:08:36.186 { 00:08:36.186 "dma_device_id": "system", 00:08:36.186 "dma_device_type": 1 00:08:36.186 }, 00:08:36.186 { 00:08:36.186 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:36.186 "dma_device_type": 2 00:08:36.186 } 00:08:36.186 ], 00:08:36.186 "driver_specific": {} 00:08:36.186 } 00:08:36.186 ] 00:08:36.186 17:30:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:36.186 17:30:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:08:36.186 17:30:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:36.186 17:30:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:36.186 17:30:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:36.186 17:30:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:36.186 17:30:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:36.186 17:30:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:36.186 17:30:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:36.186 17:30:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:36.187 17:30:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:36.187 17:30:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:36.187 17:30:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:36.187 17:30:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:36.187 17:30:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:36.187 17:30:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:36.187 17:30:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:36.187 17:30:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:36.187 "name": "Existed_Raid", 00:08:36.187 "uuid": "233b4353-d060-43a4-ae96-aa628bdbaf7a", 00:08:36.187 "strip_size_kb": 64, 00:08:36.187 "state": "configuring", 00:08:36.187 "raid_level": "raid0", 00:08:36.187 "superblock": true, 00:08:36.187 "num_base_bdevs": 3, 00:08:36.187 "num_base_bdevs_discovered": 1, 00:08:36.187 "num_base_bdevs_operational": 3, 00:08:36.187 "base_bdevs_list": [ 00:08:36.187 { 00:08:36.187 "name": "BaseBdev1", 00:08:36.187 "uuid": "05b9b550-d8cd-4d33-b4e8-3589926c160f", 00:08:36.187 "is_configured": true, 00:08:36.187 "data_offset": 2048, 00:08:36.187 "data_size": 63488 00:08:36.187 }, 00:08:36.187 { 00:08:36.187 "name": "BaseBdev2", 00:08:36.187 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:36.187 "is_configured": false, 00:08:36.187 "data_offset": 0, 00:08:36.187 "data_size": 0 00:08:36.187 }, 00:08:36.187 { 00:08:36.187 "name": "BaseBdev3", 00:08:36.187 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:36.187 "is_configured": false, 00:08:36.187 "data_offset": 0, 00:08:36.187 "data_size": 0 00:08:36.187 } 00:08:36.187 ] 00:08:36.187 }' 00:08:36.187 17:30:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:36.187 17:30:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:36.464 17:30:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:36.464 17:30:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:36.464 17:30:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:36.464 [2024-11-27 17:30:07.586421] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:36.464 [2024-11-27 17:30:07.586517] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:08:36.464 17:30:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:36.464 17:30:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:36.464 17:30:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:36.464 17:30:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:36.464 [2024-11-27 17:30:07.598471] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:36.464 [2024-11-27 17:30:07.600608] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:36.464 [2024-11-27 17:30:07.600684] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:36.464 [2024-11-27 17:30:07.600730] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:08:36.464 [2024-11-27 17:30:07.600755] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:08:36.464 17:30:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:36.464 17:30:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:08:36.464 17:30:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:36.464 17:30:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:36.464 17:30:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:36.464 17:30:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:36.464 17:30:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:36.464 17:30:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:36.464 17:30:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:36.464 17:30:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:36.464 17:30:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:36.464 17:30:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:36.464 17:30:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:36.464 17:30:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:36.464 17:30:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:36.464 17:30:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:36.464 17:30:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:36.464 17:30:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:36.724 17:30:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:36.724 "name": "Existed_Raid", 00:08:36.724 "uuid": "5a4a306e-e43c-4d0e-a4a1-e59bc3aa08cd", 00:08:36.724 "strip_size_kb": 64, 00:08:36.724 "state": "configuring", 00:08:36.724 "raid_level": "raid0", 00:08:36.724 "superblock": true, 00:08:36.724 "num_base_bdevs": 3, 00:08:36.724 "num_base_bdevs_discovered": 1, 00:08:36.724 "num_base_bdevs_operational": 3, 00:08:36.724 "base_bdevs_list": [ 00:08:36.724 { 00:08:36.724 "name": "BaseBdev1", 00:08:36.724 "uuid": "05b9b550-d8cd-4d33-b4e8-3589926c160f", 00:08:36.724 "is_configured": true, 00:08:36.724 "data_offset": 2048, 00:08:36.724 "data_size": 63488 00:08:36.724 }, 00:08:36.724 { 00:08:36.724 "name": "BaseBdev2", 00:08:36.724 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:36.724 "is_configured": false, 00:08:36.724 "data_offset": 0, 00:08:36.724 "data_size": 0 00:08:36.724 }, 00:08:36.724 { 00:08:36.724 "name": "BaseBdev3", 00:08:36.724 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:36.724 "is_configured": false, 00:08:36.724 "data_offset": 0, 00:08:36.724 "data_size": 0 00:08:36.724 } 00:08:36.724 ] 00:08:36.724 }' 00:08:36.724 17:30:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:36.724 17:30:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:36.984 17:30:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:08:36.984 17:30:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:36.984 17:30:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:36.984 [2024-11-27 17:30:08.097289] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:36.984 BaseBdev2 00:08:36.984 17:30:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:36.984 17:30:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:08:36.984 17:30:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:08:36.984 17:30:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:08:36.984 17:30:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:08:36.984 17:30:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:08:36.984 17:30:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:08:36.984 17:30:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:08:36.984 17:30:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:36.984 17:30:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:36.984 17:30:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:36.984 17:30:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:08:36.984 17:30:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:36.984 17:30:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:36.984 [ 00:08:36.984 { 00:08:36.984 "name": "BaseBdev2", 00:08:36.984 "aliases": [ 00:08:36.984 "0e6be602-5f0e-4a96-9ef3-122dbf599544" 00:08:36.984 ], 00:08:36.984 "product_name": "Malloc disk", 00:08:36.984 "block_size": 512, 00:08:36.984 "num_blocks": 65536, 00:08:36.984 "uuid": "0e6be602-5f0e-4a96-9ef3-122dbf599544", 00:08:36.984 "assigned_rate_limits": { 00:08:36.984 "rw_ios_per_sec": 0, 00:08:36.984 "rw_mbytes_per_sec": 0, 00:08:36.984 "r_mbytes_per_sec": 0, 00:08:36.984 "w_mbytes_per_sec": 0 00:08:36.984 }, 00:08:36.984 "claimed": true, 00:08:36.984 "claim_type": "exclusive_write", 00:08:36.984 "zoned": false, 00:08:36.984 "supported_io_types": { 00:08:36.984 "read": true, 00:08:36.984 "write": true, 00:08:36.984 "unmap": true, 00:08:36.984 "flush": true, 00:08:36.984 "reset": true, 00:08:36.984 "nvme_admin": false, 00:08:36.984 "nvme_io": false, 00:08:36.984 "nvme_io_md": false, 00:08:36.984 "write_zeroes": true, 00:08:36.984 "zcopy": true, 00:08:36.984 "get_zone_info": false, 00:08:36.984 "zone_management": false, 00:08:36.984 "zone_append": false, 00:08:36.984 "compare": false, 00:08:36.984 "compare_and_write": false, 00:08:36.984 "abort": true, 00:08:36.984 "seek_hole": false, 00:08:36.984 "seek_data": false, 00:08:36.984 "copy": true, 00:08:36.984 "nvme_iov_md": false 00:08:36.984 }, 00:08:36.984 "memory_domains": [ 00:08:36.984 { 00:08:36.984 "dma_device_id": "system", 00:08:36.984 "dma_device_type": 1 00:08:36.984 }, 00:08:36.984 { 00:08:36.984 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:36.984 "dma_device_type": 2 00:08:36.984 } 00:08:36.984 ], 00:08:36.984 "driver_specific": {} 00:08:36.984 } 00:08:36.984 ] 00:08:36.984 17:30:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:36.984 17:30:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:08:36.984 17:30:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:08:36.984 17:30:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:36.984 17:30:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:36.984 17:30:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:36.984 17:30:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:36.984 17:30:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:36.984 17:30:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:36.984 17:30:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:36.984 17:30:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:36.984 17:30:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:36.984 17:30:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:36.984 17:30:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:36.984 17:30:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:36.984 17:30:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:36.984 17:30:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:36.984 17:30:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:36.984 17:30:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:37.244 17:30:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:37.244 "name": "Existed_Raid", 00:08:37.244 "uuid": "5a4a306e-e43c-4d0e-a4a1-e59bc3aa08cd", 00:08:37.244 "strip_size_kb": 64, 00:08:37.244 "state": "configuring", 00:08:37.244 "raid_level": "raid0", 00:08:37.244 "superblock": true, 00:08:37.244 "num_base_bdevs": 3, 00:08:37.244 "num_base_bdevs_discovered": 2, 00:08:37.244 "num_base_bdevs_operational": 3, 00:08:37.244 "base_bdevs_list": [ 00:08:37.244 { 00:08:37.244 "name": "BaseBdev1", 00:08:37.244 "uuid": "05b9b550-d8cd-4d33-b4e8-3589926c160f", 00:08:37.244 "is_configured": true, 00:08:37.244 "data_offset": 2048, 00:08:37.244 "data_size": 63488 00:08:37.244 }, 00:08:37.244 { 00:08:37.244 "name": "BaseBdev2", 00:08:37.244 "uuid": "0e6be602-5f0e-4a96-9ef3-122dbf599544", 00:08:37.244 "is_configured": true, 00:08:37.244 "data_offset": 2048, 00:08:37.244 "data_size": 63488 00:08:37.244 }, 00:08:37.244 { 00:08:37.244 "name": "BaseBdev3", 00:08:37.244 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:37.244 "is_configured": false, 00:08:37.244 "data_offset": 0, 00:08:37.244 "data_size": 0 00:08:37.244 } 00:08:37.244 ] 00:08:37.244 }' 00:08:37.244 17:30:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:37.244 17:30:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:37.505 17:30:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:08:37.505 17:30:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:37.505 17:30:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:37.505 [2024-11-27 17:30:08.621071] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:37.505 [2024-11-27 17:30:08.621419] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:08:37.505 [2024-11-27 17:30:08.621488] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:08:37.505 [2024-11-27 17:30:08.621826] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:08:37.505 BaseBdev3 00:08:37.505 [2024-11-27 17:30:08.622007] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:08:37.505 [2024-11-27 17:30:08.622019] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:08:37.505 [2024-11-27 17:30:08.622169] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:37.505 17:30:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:37.505 17:30:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:08:37.505 17:30:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:08:37.505 17:30:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:08:37.505 17:30:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:08:37.505 17:30:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:08:37.505 17:30:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:08:37.505 17:30:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:08:37.505 17:30:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:37.505 17:30:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:37.505 17:30:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:37.505 17:30:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:08:37.505 17:30:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:37.505 17:30:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:37.505 [ 00:08:37.505 { 00:08:37.505 "name": "BaseBdev3", 00:08:37.505 "aliases": [ 00:08:37.505 "c5985947-53b2-48b6-8d04-4aafb8bcf351" 00:08:37.505 ], 00:08:37.505 "product_name": "Malloc disk", 00:08:37.505 "block_size": 512, 00:08:37.505 "num_blocks": 65536, 00:08:37.505 "uuid": "c5985947-53b2-48b6-8d04-4aafb8bcf351", 00:08:37.505 "assigned_rate_limits": { 00:08:37.505 "rw_ios_per_sec": 0, 00:08:37.505 "rw_mbytes_per_sec": 0, 00:08:37.505 "r_mbytes_per_sec": 0, 00:08:37.505 "w_mbytes_per_sec": 0 00:08:37.505 }, 00:08:37.505 "claimed": true, 00:08:37.505 "claim_type": "exclusive_write", 00:08:37.505 "zoned": false, 00:08:37.505 "supported_io_types": { 00:08:37.505 "read": true, 00:08:37.505 "write": true, 00:08:37.505 "unmap": true, 00:08:37.505 "flush": true, 00:08:37.505 "reset": true, 00:08:37.505 "nvme_admin": false, 00:08:37.505 "nvme_io": false, 00:08:37.505 "nvme_io_md": false, 00:08:37.505 "write_zeroes": true, 00:08:37.505 "zcopy": true, 00:08:37.505 "get_zone_info": false, 00:08:37.505 "zone_management": false, 00:08:37.505 "zone_append": false, 00:08:37.505 "compare": false, 00:08:37.505 "compare_and_write": false, 00:08:37.505 "abort": true, 00:08:37.505 "seek_hole": false, 00:08:37.505 "seek_data": false, 00:08:37.505 "copy": true, 00:08:37.505 "nvme_iov_md": false 00:08:37.505 }, 00:08:37.505 "memory_domains": [ 00:08:37.505 { 00:08:37.505 "dma_device_id": "system", 00:08:37.505 "dma_device_type": 1 00:08:37.505 }, 00:08:37.505 { 00:08:37.505 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:37.505 "dma_device_type": 2 00:08:37.505 } 00:08:37.505 ], 00:08:37.505 "driver_specific": {} 00:08:37.505 } 00:08:37.505 ] 00:08:37.505 17:30:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:37.505 17:30:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:08:37.505 17:30:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:08:37.505 17:30:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:37.505 17:30:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid0 64 3 00:08:37.505 17:30:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:37.505 17:30:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:37.505 17:30:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:37.505 17:30:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:37.505 17:30:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:37.505 17:30:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:37.505 17:30:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:37.505 17:30:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:37.505 17:30:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:37.505 17:30:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:37.505 17:30:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:37.505 17:30:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:37.505 17:30:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:37.505 17:30:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:37.765 17:30:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:37.765 "name": "Existed_Raid", 00:08:37.765 "uuid": "5a4a306e-e43c-4d0e-a4a1-e59bc3aa08cd", 00:08:37.765 "strip_size_kb": 64, 00:08:37.765 "state": "online", 00:08:37.765 "raid_level": "raid0", 00:08:37.765 "superblock": true, 00:08:37.765 "num_base_bdevs": 3, 00:08:37.765 "num_base_bdevs_discovered": 3, 00:08:37.765 "num_base_bdevs_operational": 3, 00:08:37.765 "base_bdevs_list": [ 00:08:37.765 { 00:08:37.765 "name": "BaseBdev1", 00:08:37.765 "uuid": "05b9b550-d8cd-4d33-b4e8-3589926c160f", 00:08:37.765 "is_configured": true, 00:08:37.765 "data_offset": 2048, 00:08:37.765 "data_size": 63488 00:08:37.765 }, 00:08:37.765 { 00:08:37.765 "name": "BaseBdev2", 00:08:37.765 "uuid": "0e6be602-5f0e-4a96-9ef3-122dbf599544", 00:08:37.765 "is_configured": true, 00:08:37.765 "data_offset": 2048, 00:08:37.765 "data_size": 63488 00:08:37.765 }, 00:08:37.765 { 00:08:37.765 "name": "BaseBdev3", 00:08:37.765 "uuid": "c5985947-53b2-48b6-8d04-4aafb8bcf351", 00:08:37.765 "is_configured": true, 00:08:37.765 "data_offset": 2048, 00:08:37.765 "data_size": 63488 00:08:37.765 } 00:08:37.765 ] 00:08:37.765 }' 00:08:37.765 17:30:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:37.765 17:30:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:38.025 17:30:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:08:38.026 17:30:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:08:38.026 17:30:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:38.026 17:30:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:38.026 17:30:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:08:38.026 17:30:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:38.026 17:30:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:38.026 17:30:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:08:38.026 17:30:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:38.026 17:30:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:38.026 [2024-11-27 17:30:09.056636] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:38.026 17:30:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:38.026 17:30:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:38.026 "name": "Existed_Raid", 00:08:38.026 "aliases": [ 00:08:38.026 "5a4a306e-e43c-4d0e-a4a1-e59bc3aa08cd" 00:08:38.026 ], 00:08:38.026 "product_name": "Raid Volume", 00:08:38.026 "block_size": 512, 00:08:38.026 "num_blocks": 190464, 00:08:38.026 "uuid": "5a4a306e-e43c-4d0e-a4a1-e59bc3aa08cd", 00:08:38.026 "assigned_rate_limits": { 00:08:38.026 "rw_ios_per_sec": 0, 00:08:38.026 "rw_mbytes_per_sec": 0, 00:08:38.026 "r_mbytes_per_sec": 0, 00:08:38.026 "w_mbytes_per_sec": 0 00:08:38.026 }, 00:08:38.026 "claimed": false, 00:08:38.026 "zoned": false, 00:08:38.026 "supported_io_types": { 00:08:38.026 "read": true, 00:08:38.026 "write": true, 00:08:38.026 "unmap": true, 00:08:38.026 "flush": true, 00:08:38.026 "reset": true, 00:08:38.026 "nvme_admin": false, 00:08:38.026 "nvme_io": false, 00:08:38.026 "nvme_io_md": false, 00:08:38.026 "write_zeroes": true, 00:08:38.026 "zcopy": false, 00:08:38.026 "get_zone_info": false, 00:08:38.026 "zone_management": false, 00:08:38.026 "zone_append": false, 00:08:38.026 "compare": false, 00:08:38.026 "compare_and_write": false, 00:08:38.026 "abort": false, 00:08:38.026 "seek_hole": false, 00:08:38.026 "seek_data": false, 00:08:38.026 "copy": false, 00:08:38.026 "nvme_iov_md": false 00:08:38.026 }, 00:08:38.026 "memory_domains": [ 00:08:38.026 { 00:08:38.026 "dma_device_id": "system", 00:08:38.026 "dma_device_type": 1 00:08:38.026 }, 00:08:38.026 { 00:08:38.026 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:38.026 "dma_device_type": 2 00:08:38.026 }, 00:08:38.026 { 00:08:38.026 "dma_device_id": "system", 00:08:38.026 "dma_device_type": 1 00:08:38.026 }, 00:08:38.026 { 00:08:38.026 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:38.026 "dma_device_type": 2 00:08:38.026 }, 00:08:38.026 { 00:08:38.026 "dma_device_id": "system", 00:08:38.026 "dma_device_type": 1 00:08:38.026 }, 00:08:38.026 { 00:08:38.026 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:38.026 "dma_device_type": 2 00:08:38.026 } 00:08:38.026 ], 00:08:38.026 "driver_specific": { 00:08:38.026 "raid": { 00:08:38.026 "uuid": "5a4a306e-e43c-4d0e-a4a1-e59bc3aa08cd", 00:08:38.026 "strip_size_kb": 64, 00:08:38.026 "state": "online", 00:08:38.026 "raid_level": "raid0", 00:08:38.026 "superblock": true, 00:08:38.026 "num_base_bdevs": 3, 00:08:38.026 "num_base_bdevs_discovered": 3, 00:08:38.026 "num_base_bdevs_operational": 3, 00:08:38.026 "base_bdevs_list": [ 00:08:38.026 { 00:08:38.026 "name": "BaseBdev1", 00:08:38.026 "uuid": "05b9b550-d8cd-4d33-b4e8-3589926c160f", 00:08:38.026 "is_configured": true, 00:08:38.026 "data_offset": 2048, 00:08:38.026 "data_size": 63488 00:08:38.026 }, 00:08:38.026 { 00:08:38.026 "name": "BaseBdev2", 00:08:38.026 "uuid": "0e6be602-5f0e-4a96-9ef3-122dbf599544", 00:08:38.026 "is_configured": true, 00:08:38.026 "data_offset": 2048, 00:08:38.026 "data_size": 63488 00:08:38.026 }, 00:08:38.026 { 00:08:38.026 "name": "BaseBdev3", 00:08:38.026 "uuid": "c5985947-53b2-48b6-8d04-4aafb8bcf351", 00:08:38.026 "is_configured": true, 00:08:38.026 "data_offset": 2048, 00:08:38.026 "data_size": 63488 00:08:38.026 } 00:08:38.026 ] 00:08:38.026 } 00:08:38.026 } 00:08:38.026 }' 00:08:38.026 17:30:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:38.026 17:30:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:08:38.026 BaseBdev2 00:08:38.026 BaseBdev3' 00:08:38.026 17:30:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:38.026 17:30:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:38.026 17:30:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:38.026 17:30:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:08:38.026 17:30:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:38.026 17:30:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:38.026 17:30:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:38.026 17:30:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:38.286 17:30:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:38.286 17:30:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:38.286 17:30:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:38.286 17:30:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:08:38.286 17:30:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:38.286 17:30:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:38.286 17:30:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:38.286 17:30:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:38.286 17:30:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:38.286 17:30:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:38.286 17:30:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:38.286 17:30:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:08:38.286 17:30:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:38.286 17:30:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:38.286 17:30:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:38.286 17:30:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:38.286 17:30:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:38.286 17:30:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:38.286 17:30:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:08:38.286 17:30:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:38.286 17:30:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:38.286 [2024-11-27 17:30:09.335922] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:08:38.286 [2024-11-27 17:30:09.335989] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:38.286 [2024-11-27 17:30:09.336085] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:38.286 17:30:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:38.286 17:30:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:08:38.286 17:30:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy raid0 00:08:38.286 17:30:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:08:38.286 17:30:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # return 1 00:08:38.286 17:30:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:08:38.286 17:30:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline raid0 64 2 00:08:38.286 17:30:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:38.286 17:30:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:08:38.286 17:30:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:38.286 17:30:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:38.286 17:30:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:38.286 17:30:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:38.286 17:30:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:38.286 17:30:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:38.286 17:30:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:38.286 17:30:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:38.286 17:30:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:38.286 17:30:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:38.286 17:30:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:38.286 17:30:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:38.286 17:30:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:38.286 "name": "Existed_Raid", 00:08:38.286 "uuid": "5a4a306e-e43c-4d0e-a4a1-e59bc3aa08cd", 00:08:38.286 "strip_size_kb": 64, 00:08:38.286 "state": "offline", 00:08:38.286 "raid_level": "raid0", 00:08:38.286 "superblock": true, 00:08:38.286 "num_base_bdevs": 3, 00:08:38.286 "num_base_bdevs_discovered": 2, 00:08:38.286 "num_base_bdevs_operational": 2, 00:08:38.286 "base_bdevs_list": [ 00:08:38.286 { 00:08:38.286 "name": null, 00:08:38.286 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:38.286 "is_configured": false, 00:08:38.286 "data_offset": 0, 00:08:38.286 "data_size": 63488 00:08:38.286 }, 00:08:38.286 { 00:08:38.286 "name": "BaseBdev2", 00:08:38.286 "uuid": "0e6be602-5f0e-4a96-9ef3-122dbf599544", 00:08:38.286 "is_configured": true, 00:08:38.286 "data_offset": 2048, 00:08:38.286 "data_size": 63488 00:08:38.286 }, 00:08:38.286 { 00:08:38.286 "name": "BaseBdev3", 00:08:38.286 "uuid": "c5985947-53b2-48b6-8d04-4aafb8bcf351", 00:08:38.286 "is_configured": true, 00:08:38.286 "data_offset": 2048, 00:08:38.286 "data_size": 63488 00:08:38.286 } 00:08:38.286 ] 00:08:38.286 }' 00:08:38.286 17:30:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:38.286 17:30:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:38.856 17:30:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:08:38.856 17:30:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:38.856 17:30:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:08:38.856 17:30:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:38.856 17:30:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:38.856 17:30:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:38.856 17:30:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:38.856 17:30:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:08:38.856 17:30:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:08:38.856 17:30:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:08:38.856 17:30:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:38.856 17:30:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:38.856 [2024-11-27 17:30:09.895786] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:08:38.856 17:30:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:38.856 17:30:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:08:38.856 17:30:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:38.856 17:30:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:38.856 17:30:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:38.856 17:30:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:38.856 17:30:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:08:38.856 17:30:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:38.856 17:30:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:08:38.856 17:30:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:08:38.856 17:30:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:08:38.856 17:30:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:38.856 17:30:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:38.856 [2024-11-27 17:30:09.972252] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:08:38.856 [2024-11-27 17:30:09.972349] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:08:38.856 17:30:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:38.856 17:30:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:08:38.856 17:30:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:38.856 17:30:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:08:38.856 17:30:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:38.856 17:30:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:38.856 17:30:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:38.856 17:30:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:38.856 17:30:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:08:38.856 17:30:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:08:38.857 17:30:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 3 -gt 2 ']' 00:08:38.857 17:30:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:08:38.857 17:30:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:08:38.857 17:30:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:08:38.857 17:30:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:38.857 17:30:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:38.857 BaseBdev2 00:08:38.857 17:30:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:38.857 17:30:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:08:38.857 17:30:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:08:38.857 17:30:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:08:39.118 17:30:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:08:39.118 17:30:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:08:39.118 17:30:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:08:39.118 17:30:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:08:39.118 17:30:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:39.118 17:30:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:39.118 17:30:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:39.118 17:30:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:08:39.118 17:30:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:39.118 17:30:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:39.118 [ 00:08:39.118 { 00:08:39.118 "name": "BaseBdev2", 00:08:39.118 "aliases": [ 00:08:39.118 "ee3db6df-cfb5-4911-b1b3-15c56606c20e" 00:08:39.118 ], 00:08:39.118 "product_name": "Malloc disk", 00:08:39.118 "block_size": 512, 00:08:39.118 "num_blocks": 65536, 00:08:39.118 "uuid": "ee3db6df-cfb5-4911-b1b3-15c56606c20e", 00:08:39.118 "assigned_rate_limits": { 00:08:39.118 "rw_ios_per_sec": 0, 00:08:39.118 "rw_mbytes_per_sec": 0, 00:08:39.118 "r_mbytes_per_sec": 0, 00:08:39.118 "w_mbytes_per_sec": 0 00:08:39.118 }, 00:08:39.118 "claimed": false, 00:08:39.118 "zoned": false, 00:08:39.118 "supported_io_types": { 00:08:39.118 "read": true, 00:08:39.118 "write": true, 00:08:39.118 "unmap": true, 00:08:39.118 "flush": true, 00:08:39.118 "reset": true, 00:08:39.118 "nvme_admin": false, 00:08:39.118 "nvme_io": false, 00:08:39.118 "nvme_io_md": false, 00:08:39.118 "write_zeroes": true, 00:08:39.118 "zcopy": true, 00:08:39.118 "get_zone_info": false, 00:08:39.118 "zone_management": false, 00:08:39.118 "zone_append": false, 00:08:39.118 "compare": false, 00:08:39.118 "compare_and_write": false, 00:08:39.118 "abort": true, 00:08:39.118 "seek_hole": false, 00:08:39.118 "seek_data": false, 00:08:39.118 "copy": true, 00:08:39.118 "nvme_iov_md": false 00:08:39.118 }, 00:08:39.118 "memory_domains": [ 00:08:39.118 { 00:08:39.118 "dma_device_id": "system", 00:08:39.118 "dma_device_type": 1 00:08:39.118 }, 00:08:39.118 { 00:08:39.118 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:39.118 "dma_device_type": 2 00:08:39.118 } 00:08:39.118 ], 00:08:39.118 "driver_specific": {} 00:08:39.118 } 00:08:39.118 ] 00:08:39.118 17:30:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:39.118 17:30:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:08:39.118 17:30:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:08:39.118 17:30:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:08:39.118 17:30:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:08:39.118 17:30:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:39.118 17:30:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:39.118 BaseBdev3 00:08:39.118 17:30:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:39.118 17:30:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:08:39.118 17:30:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:08:39.118 17:30:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:08:39.118 17:30:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:08:39.118 17:30:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:08:39.118 17:30:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:08:39.118 17:30:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:08:39.118 17:30:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:39.118 17:30:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:39.118 17:30:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:39.118 17:30:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:08:39.118 17:30:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:39.118 17:30:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:39.118 [ 00:08:39.118 { 00:08:39.118 "name": "BaseBdev3", 00:08:39.118 "aliases": [ 00:08:39.118 "8347771e-9a14-43f8-8637-10bccef51256" 00:08:39.118 ], 00:08:39.118 "product_name": "Malloc disk", 00:08:39.118 "block_size": 512, 00:08:39.118 "num_blocks": 65536, 00:08:39.118 "uuid": "8347771e-9a14-43f8-8637-10bccef51256", 00:08:39.118 "assigned_rate_limits": { 00:08:39.118 "rw_ios_per_sec": 0, 00:08:39.118 "rw_mbytes_per_sec": 0, 00:08:39.118 "r_mbytes_per_sec": 0, 00:08:39.118 "w_mbytes_per_sec": 0 00:08:39.118 }, 00:08:39.118 "claimed": false, 00:08:39.118 "zoned": false, 00:08:39.118 "supported_io_types": { 00:08:39.118 "read": true, 00:08:39.118 "write": true, 00:08:39.118 "unmap": true, 00:08:39.118 "flush": true, 00:08:39.118 "reset": true, 00:08:39.118 "nvme_admin": false, 00:08:39.118 "nvme_io": false, 00:08:39.118 "nvme_io_md": false, 00:08:39.118 "write_zeroes": true, 00:08:39.118 "zcopy": true, 00:08:39.118 "get_zone_info": false, 00:08:39.118 "zone_management": false, 00:08:39.118 "zone_append": false, 00:08:39.118 "compare": false, 00:08:39.118 "compare_and_write": false, 00:08:39.118 "abort": true, 00:08:39.118 "seek_hole": false, 00:08:39.118 "seek_data": false, 00:08:39.118 "copy": true, 00:08:39.118 "nvme_iov_md": false 00:08:39.118 }, 00:08:39.118 "memory_domains": [ 00:08:39.118 { 00:08:39.118 "dma_device_id": "system", 00:08:39.118 "dma_device_type": 1 00:08:39.118 }, 00:08:39.118 { 00:08:39.118 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:39.118 "dma_device_type": 2 00:08:39.118 } 00:08:39.118 ], 00:08:39.118 "driver_specific": {} 00:08:39.118 } 00:08:39.118 ] 00:08:39.118 17:30:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:39.118 17:30:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:08:39.118 17:30:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:08:39.118 17:30:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:08:39.118 17:30:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:39.118 17:30:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:39.118 17:30:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:39.118 [2024-11-27 17:30:10.150063] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:39.118 [2024-11-27 17:30:10.150192] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:39.118 [2024-11-27 17:30:10.150252] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:39.118 [2024-11-27 17:30:10.152405] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:39.118 17:30:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:39.118 17:30:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:39.118 17:30:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:39.118 17:30:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:39.118 17:30:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:39.118 17:30:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:39.118 17:30:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:39.118 17:30:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:39.118 17:30:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:39.118 17:30:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:39.118 17:30:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:39.118 17:30:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:39.118 17:30:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:39.118 17:30:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:39.118 17:30:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:39.118 17:30:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:39.118 17:30:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:39.118 "name": "Existed_Raid", 00:08:39.118 "uuid": "1a56a3f9-18e1-4735-aec1-c86a4787f310", 00:08:39.118 "strip_size_kb": 64, 00:08:39.118 "state": "configuring", 00:08:39.118 "raid_level": "raid0", 00:08:39.118 "superblock": true, 00:08:39.118 "num_base_bdevs": 3, 00:08:39.118 "num_base_bdevs_discovered": 2, 00:08:39.118 "num_base_bdevs_operational": 3, 00:08:39.118 "base_bdevs_list": [ 00:08:39.119 { 00:08:39.119 "name": "BaseBdev1", 00:08:39.119 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:39.119 "is_configured": false, 00:08:39.119 "data_offset": 0, 00:08:39.119 "data_size": 0 00:08:39.119 }, 00:08:39.119 { 00:08:39.119 "name": "BaseBdev2", 00:08:39.119 "uuid": "ee3db6df-cfb5-4911-b1b3-15c56606c20e", 00:08:39.119 "is_configured": true, 00:08:39.119 "data_offset": 2048, 00:08:39.119 "data_size": 63488 00:08:39.119 }, 00:08:39.119 { 00:08:39.119 "name": "BaseBdev3", 00:08:39.119 "uuid": "8347771e-9a14-43f8-8637-10bccef51256", 00:08:39.119 "is_configured": true, 00:08:39.119 "data_offset": 2048, 00:08:39.119 "data_size": 63488 00:08:39.119 } 00:08:39.119 ] 00:08:39.119 }' 00:08:39.119 17:30:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:39.119 17:30:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:39.690 17:30:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:08:39.690 17:30:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:39.690 17:30:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:39.690 [2024-11-27 17:30:10.609233] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:08:39.690 17:30:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:39.690 17:30:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:39.690 17:30:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:39.690 17:30:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:39.690 17:30:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:39.690 17:30:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:39.690 17:30:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:39.690 17:30:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:39.690 17:30:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:39.690 17:30:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:39.690 17:30:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:39.690 17:30:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:39.690 17:30:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:39.690 17:30:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:39.690 17:30:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:39.690 17:30:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:39.690 17:30:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:39.690 "name": "Existed_Raid", 00:08:39.690 "uuid": "1a56a3f9-18e1-4735-aec1-c86a4787f310", 00:08:39.690 "strip_size_kb": 64, 00:08:39.690 "state": "configuring", 00:08:39.690 "raid_level": "raid0", 00:08:39.690 "superblock": true, 00:08:39.690 "num_base_bdevs": 3, 00:08:39.690 "num_base_bdevs_discovered": 1, 00:08:39.690 "num_base_bdevs_operational": 3, 00:08:39.690 "base_bdevs_list": [ 00:08:39.690 { 00:08:39.690 "name": "BaseBdev1", 00:08:39.690 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:39.690 "is_configured": false, 00:08:39.690 "data_offset": 0, 00:08:39.690 "data_size": 0 00:08:39.690 }, 00:08:39.690 { 00:08:39.690 "name": null, 00:08:39.690 "uuid": "ee3db6df-cfb5-4911-b1b3-15c56606c20e", 00:08:39.690 "is_configured": false, 00:08:39.690 "data_offset": 0, 00:08:39.690 "data_size": 63488 00:08:39.690 }, 00:08:39.690 { 00:08:39.690 "name": "BaseBdev3", 00:08:39.690 "uuid": "8347771e-9a14-43f8-8637-10bccef51256", 00:08:39.690 "is_configured": true, 00:08:39.691 "data_offset": 2048, 00:08:39.691 "data_size": 63488 00:08:39.691 } 00:08:39.691 ] 00:08:39.691 }' 00:08:39.691 17:30:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:39.691 17:30:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:39.951 17:30:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:39.951 17:30:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:39.951 17:30:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:39.951 17:30:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:08:39.951 17:30:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:39.951 17:30:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:08:39.951 17:30:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:08:39.951 17:30:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:39.951 17:30:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:39.951 [2024-11-27 17:30:11.097212] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:39.951 BaseBdev1 00:08:39.951 17:30:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:39.951 17:30:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:08:39.951 17:30:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:08:39.951 17:30:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:08:39.951 17:30:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:08:39.951 17:30:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:08:39.951 17:30:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:08:39.951 17:30:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:08:39.951 17:30:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:39.951 17:30:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:39.951 17:30:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:39.951 17:30:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:08:39.951 17:30:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:39.951 17:30:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:39.952 [ 00:08:39.952 { 00:08:39.952 "name": "BaseBdev1", 00:08:39.952 "aliases": [ 00:08:39.952 "7714e17b-9643-4eb7-b3a1-fafbf963c815" 00:08:39.952 ], 00:08:39.952 "product_name": "Malloc disk", 00:08:39.952 "block_size": 512, 00:08:39.952 "num_blocks": 65536, 00:08:39.952 "uuid": "7714e17b-9643-4eb7-b3a1-fafbf963c815", 00:08:39.952 "assigned_rate_limits": { 00:08:39.952 "rw_ios_per_sec": 0, 00:08:39.952 "rw_mbytes_per_sec": 0, 00:08:39.952 "r_mbytes_per_sec": 0, 00:08:39.952 "w_mbytes_per_sec": 0 00:08:39.952 }, 00:08:39.952 "claimed": true, 00:08:39.952 "claim_type": "exclusive_write", 00:08:39.952 "zoned": false, 00:08:39.952 "supported_io_types": { 00:08:39.952 "read": true, 00:08:39.952 "write": true, 00:08:39.952 "unmap": true, 00:08:39.952 "flush": true, 00:08:39.952 "reset": true, 00:08:39.952 "nvme_admin": false, 00:08:39.952 "nvme_io": false, 00:08:39.952 "nvme_io_md": false, 00:08:39.952 "write_zeroes": true, 00:08:39.952 "zcopy": true, 00:08:39.952 "get_zone_info": false, 00:08:39.952 "zone_management": false, 00:08:39.952 "zone_append": false, 00:08:39.952 "compare": false, 00:08:39.952 "compare_and_write": false, 00:08:39.952 "abort": true, 00:08:39.952 "seek_hole": false, 00:08:39.952 "seek_data": false, 00:08:39.952 "copy": true, 00:08:39.952 "nvme_iov_md": false 00:08:39.952 }, 00:08:39.952 "memory_domains": [ 00:08:39.952 { 00:08:39.952 "dma_device_id": "system", 00:08:39.952 "dma_device_type": 1 00:08:39.952 }, 00:08:39.952 { 00:08:39.952 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:39.952 "dma_device_type": 2 00:08:39.952 } 00:08:39.952 ], 00:08:39.952 "driver_specific": {} 00:08:39.952 } 00:08:39.952 ] 00:08:39.952 17:30:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:39.952 17:30:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:08:39.952 17:30:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:39.952 17:30:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:39.952 17:30:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:39.952 17:30:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:39.952 17:30:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:39.952 17:30:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:39.952 17:30:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:39.952 17:30:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:39.952 17:30:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:39.952 17:30:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:39.952 17:30:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:39.952 17:30:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:39.952 17:30:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:40.213 17:30:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:40.213 17:30:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:40.213 17:30:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:40.213 "name": "Existed_Raid", 00:08:40.213 "uuid": "1a56a3f9-18e1-4735-aec1-c86a4787f310", 00:08:40.213 "strip_size_kb": 64, 00:08:40.213 "state": "configuring", 00:08:40.213 "raid_level": "raid0", 00:08:40.213 "superblock": true, 00:08:40.213 "num_base_bdevs": 3, 00:08:40.213 "num_base_bdevs_discovered": 2, 00:08:40.213 "num_base_bdevs_operational": 3, 00:08:40.213 "base_bdevs_list": [ 00:08:40.213 { 00:08:40.213 "name": "BaseBdev1", 00:08:40.213 "uuid": "7714e17b-9643-4eb7-b3a1-fafbf963c815", 00:08:40.213 "is_configured": true, 00:08:40.213 "data_offset": 2048, 00:08:40.213 "data_size": 63488 00:08:40.213 }, 00:08:40.213 { 00:08:40.213 "name": null, 00:08:40.213 "uuid": "ee3db6df-cfb5-4911-b1b3-15c56606c20e", 00:08:40.213 "is_configured": false, 00:08:40.213 "data_offset": 0, 00:08:40.213 "data_size": 63488 00:08:40.213 }, 00:08:40.213 { 00:08:40.213 "name": "BaseBdev3", 00:08:40.213 "uuid": "8347771e-9a14-43f8-8637-10bccef51256", 00:08:40.213 "is_configured": true, 00:08:40.213 "data_offset": 2048, 00:08:40.213 "data_size": 63488 00:08:40.213 } 00:08:40.213 ] 00:08:40.213 }' 00:08:40.213 17:30:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:40.213 17:30:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:40.473 17:30:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:40.473 17:30:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:08:40.473 17:30:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:40.473 17:30:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:40.473 17:30:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:40.473 17:30:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:08:40.473 17:30:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:08:40.473 17:30:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:40.473 17:30:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:40.473 [2024-11-27 17:30:11.640280] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:08:40.473 17:30:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:40.473 17:30:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:40.473 17:30:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:40.473 17:30:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:40.473 17:30:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:40.473 17:30:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:40.473 17:30:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:40.473 17:30:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:40.473 17:30:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:40.473 17:30:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:40.473 17:30:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:40.473 17:30:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:40.473 17:30:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:40.473 17:30:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:40.473 17:30:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:40.733 17:30:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:40.733 17:30:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:40.733 "name": "Existed_Raid", 00:08:40.733 "uuid": "1a56a3f9-18e1-4735-aec1-c86a4787f310", 00:08:40.733 "strip_size_kb": 64, 00:08:40.733 "state": "configuring", 00:08:40.733 "raid_level": "raid0", 00:08:40.733 "superblock": true, 00:08:40.733 "num_base_bdevs": 3, 00:08:40.733 "num_base_bdevs_discovered": 1, 00:08:40.733 "num_base_bdevs_operational": 3, 00:08:40.733 "base_bdevs_list": [ 00:08:40.733 { 00:08:40.733 "name": "BaseBdev1", 00:08:40.733 "uuid": "7714e17b-9643-4eb7-b3a1-fafbf963c815", 00:08:40.733 "is_configured": true, 00:08:40.733 "data_offset": 2048, 00:08:40.733 "data_size": 63488 00:08:40.733 }, 00:08:40.733 { 00:08:40.733 "name": null, 00:08:40.733 "uuid": "ee3db6df-cfb5-4911-b1b3-15c56606c20e", 00:08:40.733 "is_configured": false, 00:08:40.733 "data_offset": 0, 00:08:40.733 "data_size": 63488 00:08:40.733 }, 00:08:40.733 { 00:08:40.733 "name": null, 00:08:40.733 "uuid": "8347771e-9a14-43f8-8637-10bccef51256", 00:08:40.733 "is_configured": false, 00:08:40.733 "data_offset": 0, 00:08:40.733 "data_size": 63488 00:08:40.733 } 00:08:40.733 ] 00:08:40.733 }' 00:08:40.733 17:30:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:40.733 17:30:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:40.993 17:30:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:40.993 17:30:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:08:40.993 17:30:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:40.993 17:30:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:40.993 17:30:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:40.993 17:30:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:08:40.993 17:30:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:08:40.993 17:30:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:40.993 17:30:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:40.993 [2024-11-27 17:30:12.143446] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:40.993 17:30:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:40.993 17:30:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:40.993 17:30:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:40.993 17:30:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:40.993 17:30:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:40.993 17:30:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:40.993 17:30:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:40.993 17:30:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:40.993 17:30:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:40.993 17:30:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:40.993 17:30:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:40.993 17:30:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:40.993 17:30:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:40.993 17:30:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:40.993 17:30:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:40.993 17:30:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:41.253 17:30:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:41.253 "name": "Existed_Raid", 00:08:41.253 "uuid": "1a56a3f9-18e1-4735-aec1-c86a4787f310", 00:08:41.253 "strip_size_kb": 64, 00:08:41.253 "state": "configuring", 00:08:41.253 "raid_level": "raid0", 00:08:41.253 "superblock": true, 00:08:41.253 "num_base_bdevs": 3, 00:08:41.253 "num_base_bdevs_discovered": 2, 00:08:41.253 "num_base_bdevs_operational": 3, 00:08:41.253 "base_bdevs_list": [ 00:08:41.253 { 00:08:41.253 "name": "BaseBdev1", 00:08:41.253 "uuid": "7714e17b-9643-4eb7-b3a1-fafbf963c815", 00:08:41.253 "is_configured": true, 00:08:41.253 "data_offset": 2048, 00:08:41.253 "data_size": 63488 00:08:41.253 }, 00:08:41.253 { 00:08:41.253 "name": null, 00:08:41.253 "uuid": "ee3db6df-cfb5-4911-b1b3-15c56606c20e", 00:08:41.253 "is_configured": false, 00:08:41.253 "data_offset": 0, 00:08:41.253 "data_size": 63488 00:08:41.253 }, 00:08:41.253 { 00:08:41.253 "name": "BaseBdev3", 00:08:41.253 "uuid": "8347771e-9a14-43f8-8637-10bccef51256", 00:08:41.253 "is_configured": true, 00:08:41.253 "data_offset": 2048, 00:08:41.253 "data_size": 63488 00:08:41.253 } 00:08:41.253 ] 00:08:41.253 }' 00:08:41.253 17:30:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:41.253 17:30:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:41.514 17:30:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:41.514 17:30:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:08:41.514 17:30:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:41.514 17:30:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:41.514 17:30:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:41.514 17:30:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:08:41.514 17:30:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:08:41.514 17:30:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:41.514 17:30:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:41.514 [2024-11-27 17:30:12.598751] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:08:41.514 17:30:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:41.514 17:30:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:41.514 17:30:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:41.514 17:30:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:41.514 17:30:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:41.514 17:30:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:41.514 17:30:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:41.514 17:30:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:41.514 17:30:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:41.514 17:30:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:41.514 17:30:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:41.514 17:30:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:41.514 17:30:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:41.514 17:30:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:41.514 17:30:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:41.514 17:30:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:41.514 17:30:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:41.514 "name": "Existed_Raid", 00:08:41.514 "uuid": "1a56a3f9-18e1-4735-aec1-c86a4787f310", 00:08:41.514 "strip_size_kb": 64, 00:08:41.514 "state": "configuring", 00:08:41.514 "raid_level": "raid0", 00:08:41.514 "superblock": true, 00:08:41.514 "num_base_bdevs": 3, 00:08:41.514 "num_base_bdevs_discovered": 1, 00:08:41.514 "num_base_bdevs_operational": 3, 00:08:41.514 "base_bdevs_list": [ 00:08:41.514 { 00:08:41.514 "name": null, 00:08:41.514 "uuid": "7714e17b-9643-4eb7-b3a1-fafbf963c815", 00:08:41.514 "is_configured": false, 00:08:41.514 "data_offset": 0, 00:08:41.514 "data_size": 63488 00:08:41.514 }, 00:08:41.514 { 00:08:41.514 "name": null, 00:08:41.514 "uuid": "ee3db6df-cfb5-4911-b1b3-15c56606c20e", 00:08:41.514 "is_configured": false, 00:08:41.514 "data_offset": 0, 00:08:41.514 "data_size": 63488 00:08:41.514 }, 00:08:41.514 { 00:08:41.514 "name": "BaseBdev3", 00:08:41.514 "uuid": "8347771e-9a14-43f8-8637-10bccef51256", 00:08:41.514 "is_configured": true, 00:08:41.514 "data_offset": 2048, 00:08:41.514 "data_size": 63488 00:08:41.514 } 00:08:41.514 ] 00:08:41.514 }' 00:08:41.514 17:30:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:41.514 17:30:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:42.084 17:30:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:42.085 17:30:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:42.085 17:30:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:08:42.085 17:30:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:42.085 17:30:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:42.085 17:30:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:08:42.085 17:30:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:08:42.085 17:30:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:42.085 17:30:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:42.085 [2024-11-27 17:30:13.113768] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:42.085 17:30:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:42.085 17:30:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 3 00:08:42.085 17:30:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:42.085 17:30:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:42.085 17:30:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:42.085 17:30:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:42.085 17:30:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:42.085 17:30:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:42.085 17:30:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:42.085 17:30:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:42.085 17:30:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:42.085 17:30:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:42.085 17:30:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:42.085 17:30:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:42.085 17:30:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:42.085 17:30:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:42.085 17:30:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:42.085 "name": "Existed_Raid", 00:08:42.085 "uuid": "1a56a3f9-18e1-4735-aec1-c86a4787f310", 00:08:42.085 "strip_size_kb": 64, 00:08:42.085 "state": "configuring", 00:08:42.085 "raid_level": "raid0", 00:08:42.085 "superblock": true, 00:08:42.085 "num_base_bdevs": 3, 00:08:42.085 "num_base_bdevs_discovered": 2, 00:08:42.085 "num_base_bdevs_operational": 3, 00:08:42.085 "base_bdevs_list": [ 00:08:42.085 { 00:08:42.085 "name": null, 00:08:42.085 "uuid": "7714e17b-9643-4eb7-b3a1-fafbf963c815", 00:08:42.085 "is_configured": false, 00:08:42.085 "data_offset": 0, 00:08:42.085 "data_size": 63488 00:08:42.085 }, 00:08:42.085 { 00:08:42.085 "name": "BaseBdev2", 00:08:42.085 "uuid": "ee3db6df-cfb5-4911-b1b3-15c56606c20e", 00:08:42.085 "is_configured": true, 00:08:42.085 "data_offset": 2048, 00:08:42.085 "data_size": 63488 00:08:42.085 }, 00:08:42.085 { 00:08:42.085 "name": "BaseBdev3", 00:08:42.085 "uuid": "8347771e-9a14-43f8-8637-10bccef51256", 00:08:42.085 "is_configured": true, 00:08:42.085 "data_offset": 2048, 00:08:42.085 "data_size": 63488 00:08:42.085 } 00:08:42.085 ] 00:08:42.085 }' 00:08:42.085 17:30:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:42.085 17:30:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:42.656 17:30:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:42.656 17:30:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:42.656 17:30:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:42.656 17:30:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:08:42.656 17:30:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:42.656 17:30:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:08:42.656 17:30:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:42.656 17:30:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:08:42.656 17:30:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:42.656 17:30:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:42.656 17:30:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:42.656 17:30:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 7714e17b-9643-4eb7-b3a1-fafbf963c815 00:08:42.656 17:30:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:42.656 17:30:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:42.656 [2024-11-27 17:30:13.661645] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:08:42.656 [2024-11-27 17:30:13.661927] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:08:42.656 [2024-11-27 17:30:13.661982] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:08:42.656 NewBaseBdev 00:08:42.656 [2024-11-27 17:30:13.662275] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002870 00:08:42.656 [2024-11-27 17:30:13.662419] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:08:42.656 [2024-11-27 17:30:13.662473] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001c80 00:08:42.656 [2024-11-27 17:30:13.662637] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:42.656 17:30:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:42.656 17:30:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:08:42.656 17:30:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=NewBaseBdev 00:08:42.656 17:30:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:08:42.656 17:30:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:08:42.656 17:30:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:08:42.656 17:30:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:08:42.656 17:30:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:08:42.656 17:30:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:42.656 17:30:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:42.656 17:30:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:42.656 17:30:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:08:42.656 17:30:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:42.656 17:30:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:42.656 [ 00:08:42.656 { 00:08:42.656 "name": "NewBaseBdev", 00:08:42.656 "aliases": [ 00:08:42.656 "7714e17b-9643-4eb7-b3a1-fafbf963c815" 00:08:42.656 ], 00:08:42.656 "product_name": "Malloc disk", 00:08:42.656 "block_size": 512, 00:08:42.656 "num_blocks": 65536, 00:08:42.656 "uuid": "7714e17b-9643-4eb7-b3a1-fafbf963c815", 00:08:42.656 "assigned_rate_limits": { 00:08:42.656 "rw_ios_per_sec": 0, 00:08:42.656 "rw_mbytes_per_sec": 0, 00:08:42.656 "r_mbytes_per_sec": 0, 00:08:42.656 "w_mbytes_per_sec": 0 00:08:42.656 }, 00:08:42.656 "claimed": true, 00:08:42.656 "claim_type": "exclusive_write", 00:08:42.656 "zoned": false, 00:08:42.656 "supported_io_types": { 00:08:42.656 "read": true, 00:08:42.656 "write": true, 00:08:42.656 "unmap": true, 00:08:42.656 "flush": true, 00:08:42.656 "reset": true, 00:08:42.656 "nvme_admin": false, 00:08:42.656 "nvme_io": false, 00:08:42.656 "nvme_io_md": false, 00:08:42.656 "write_zeroes": true, 00:08:42.656 "zcopy": true, 00:08:42.656 "get_zone_info": false, 00:08:42.656 "zone_management": false, 00:08:42.656 "zone_append": false, 00:08:42.656 "compare": false, 00:08:42.656 "compare_and_write": false, 00:08:42.656 "abort": true, 00:08:42.656 "seek_hole": false, 00:08:42.656 "seek_data": false, 00:08:42.656 "copy": true, 00:08:42.656 "nvme_iov_md": false 00:08:42.656 }, 00:08:42.656 "memory_domains": [ 00:08:42.656 { 00:08:42.656 "dma_device_id": "system", 00:08:42.656 "dma_device_type": 1 00:08:42.656 }, 00:08:42.656 { 00:08:42.656 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:42.656 "dma_device_type": 2 00:08:42.656 } 00:08:42.656 ], 00:08:42.656 "driver_specific": {} 00:08:42.656 } 00:08:42.656 ] 00:08:42.656 17:30:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:42.656 17:30:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:08:42.656 17:30:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid0 64 3 00:08:42.656 17:30:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:42.656 17:30:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:42.656 17:30:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:42.656 17:30:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:42.656 17:30:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:42.656 17:30:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:42.656 17:30:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:42.656 17:30:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:42.656 17:30:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:42.656 17:30:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:42.656 17:30:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:42.656 17:30:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:42.656 17:30:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:42.656 17:30:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:42.657 17:30:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:42.657 "name": "Existed_Raid", 00:08:42.657 "uuid": "1a56a3f9-18e1-4735-aec1-c86a4787f310", 00:08:42.657 "strip_size_kb": 64, 00:08:42.657 "state": "online", 00:08:42.657 "raid_level": "raid0", 00:08:42.657 "superblock": true, 00:08:42.657 "num_base_bdevs": 3, 00:08:42.657 "num_base_bdevs_discovered": 3, 00:08:42.657 "num_base_bdevs_operational": 3, 00:08:42.657 "base_bdevs_list": [ 00:08:42.657 { 00:08:42.657 "name": "NewBaseBdev", 00:08:42.657 "uuid": "7714e17b-9643-4eb7-b3a1-fafbf963c815", 00:08:42.657 "is_configured": true, 00:08:42.657 "data_offset": 2048, 00:08:42.657 "data_size": 63488 00:08:42.657 }, 00:08:42.657 { 00:08:42.657 "name": "BaseBdev2", 00:08:42.657 "uuid": "ee3db6df-cfb5-4911-b1b3-15c56606c20e", 00:08:42.657 "is_configured": true, 00:08:42.657 "data_offset": 2048, 00:08:42.657 "data_size": 63488 00:08:42.657 }, 00:08:42.657 { 00:08:42.657 "name": "BaseBdev3", 00:08:42.657 "uuid": "8347771e-9a14-43f8-8637-10bccef51256", 00:08:42.657 "is_configured": true, 00:08:42.657 "data_offset": 2048, 00:08:42.657 "data_size": 63488 00:08:42.657 } 00:08:42.657 ] 00:08:42.657 }' 00:08:42.657 17:30:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:42.657 17:30:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:43.227 17:30:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:08:43.227 17:30:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:08:43.227 17:30:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:43.227 17:30:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:43.227 17:30:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:08:43.227 17:30:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:43.227 17:30:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:08:43.227 17:30:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:43.227 17:30:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:43.227 17:30:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:43.227 [2024-11-27 17:30:14.121169] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:43.227 17:30:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:43.227 17:30:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:43.227 "name": "Existed_Raid", 00:08:43.227 "aliases": [ 00:08:43.227 "1a56a3f9-18e1-4735-aec1-c86a4787f310" 00:08:43.227 ], 00:08:43.227 "product_name": "Raid Volume", 00:08:43.227 "block_size": 512, 00:08:43.227 "num_blocks": 190464, 00:08:43.227 "uuid": "1a56a3f9-18e1-4735-aec1-c86a4787f310", 00:08:43.227 "assigned_rate_limits": { 00:08:43.227 "rw_ios_per_sec": 0, 00:08:43.227 "rw_mbytes_per_sec": 0, 00:08:43.227 "r_mbytes_per_sec": 0, 00:08:43.227 "w_mbytes_per_sec": 0 00:08:43.227 }, 00:08:43.227 "claimed": false, 00:08:43.227 "zoned": false, 00:08:43.227 "supported_io_types": { 00:08:43.227 "read": true, 00:08:43.227 "write": true, 00:08:43.227 "unmap": true, 00:08:43.227 "flush": true, 00:08:43.227 "reset": true, 00:08:43.227 "nvme_admin": false, 00:08:43.227 "nvme_io": false, 00:08:43.228 "nvme_io_md": false, 00:08:43.228 "write_zeroes": true, 00:08:43.228 "zcopy": false, 00:08:43.228 "get_zone_info": false, 00:08:43.228 "zone_management": false, 00:08:43.228 "zone_append": false, 00:08:43.228 "compare": false, 00:08:43.228 "compare_and_write": false, 00:08:43.228 "abort": false, 00:08:43.228 "seek_hole": false, 00:08:43.228 "seek_data": false, 00:08:43.228 "copy": false, 00:08:43.228 "nvme_iov_md": false 00:08:43.228 }, 00:08:43.228 "memory_domains": [ 00:08:43.228 { 00:08:43.228 "dma_device_id": "system", 00:08:43.228 "dma_device_type": 1 00:08:43.228 }, 00:08:43.228 { 00:08:43.228 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:43.228 "dma_device_type": 2 00:08:43.228 }, 00:08:43.228 { 00:08:43.228 "dma_device_id": "system", 00:08:43.228 "dma_device_type": 1 00:08:43.228 }, 00:08:43.228 { 00:08:43.228 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:43.228 "dma_device_type": 2 00:08:43.228 }, 00:08:43.228 { 00:08:43.228 "dma_device_id": "system", 00:08:43.228 "dma_device_type": 1 00:08:43.228 }, 00:08:43.228 { 00:08:43.228 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:43.228 "dma_device_type": 2 00:08:43.228 } 00:08:43.228 ], 00:08:43.228 "driver_specific": { 00:08:43.228 "raid": { 00:08:43.228 "uuid": "1a56a3f9-18e1-4735-aec1-c86a4787f310", 00:08:43.228 "strip_size_kb": 64, 00:08:43.228 "state": "online", 00:08:43.228 "raid_level": "raid0", 00:08:43.228 "superblock": true, 00:08:43.228 "num_base_bdevs": 3, 00:08:43.228 "num_base_bdevs_discovered": 3, 00:08:43.228 "num_base_bdevs_operational": 3, 00:08:43.228 "base_bdevs_list": [ 00:08:43.228 { 00:08:43.228 "name": "NewBaseBdev", 00:08:43.228 "uuid": "7714e17b-9643-4eb7-b3a1-fafbf963c815", 00:08:43.228 "is_configured": true, 00:08:43.228 "data_offset": 2048, 00:08:43.228 "data_size": 63488 00:08:43.228 }, 00:08:43.228 { 00:08:43.228 "name": "BaseBdev2", 00:08:43.228 "uuid": "ee3db6df-cfb5-4911-b1b3-15c56606c20e", 00:08:43.228 "is_configured": true, 00:08:43.228 "data_offset": 2048, 00:08:43.228 "data_size": 63488 00:08:43.228 }, 00:08:43.228 { 00:08:43.228 "name": "BaseBdev3", 00:08:43.228 "uuid": "8347771e-9a14-43f8-8637-10bccef51256", 00:08:43.228 "is_configured": true, 00:08:43.228 "data_offset": 2048, 00:08:43.228 "data_size": 63488 00:08:43.228 } 00:08:43.228 ] 00:08:43.228 } 00:08:43.228 } 00:08:43.228 }' 00:08:43.228 17:30:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:43.228 17:30:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:08:43.228 BaseBdev2 00:08:43.228 BaseBdev3' 00:08:43.228 17:30:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:43.228 17:30:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:43.228 17:30:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:43.228 17:30:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:43.228 17:30:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:08:43.228 17:30:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:43.228 17:30:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:43.228 17:30:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:43.228 17:30:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:43.228 17:30:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:43.228 17:30:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:43.228 17:30:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:08:43.228 17:30:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:43.228 17:30:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:43.228 17:30:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:43.228 17:30:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:43.228 17:30:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:43.228 17:30:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:43.228 17:30:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:43.228 17:30:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:08:43.228 17:30:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:43.228 17:30:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:43.228 17:30:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:43.228 17:30:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:43.228 17:30:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:43.228 17:30:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:43.228 17:30:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:43.228 17:30:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:43.228 17:30:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:43.489 [2024-11-27 17:30:14.420366] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:43.489 [2024-11-27 17:30:14.420431] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:43.489 [2024-11-27 17:30:14.420544] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:43.489 [2024-11-27 17:30:14.420626] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:43.489 [2024-11-27 17:30:14.420675] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name Existed_Raid, state offline 00:08:43.489 17:30:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:43.489 17:30:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 75460 00:08:43.489 17:30:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@950 -- # '[' -z 75460 ']' 00:08:43.489 17:30:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # kill -0 75460 00:08:43.489 17:30:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@955 -- # uname 00:08:43.489 17:30:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:08:43.489 17:30:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 75460 00:08:43.489 killing process with pid 75460 00:08:43.489 17:30:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:08:43.489 17:30:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:08:43.489 17:30:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@968 -- # echo 'killing process with pid 75460' 00:08:43.489 17:30:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@969 -- # kill 75460 00:08:43.489 [2024-11-27 17:30:14.468822] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:43.489 17:30:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@974 -- # wait 75460 00:08:43.489 [2024-11-27 17:30:14.526711] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:43.749 17:30:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:08:43.749 00:08:43.749 real 0m9.165s 00:08:43.749 user 0m15.327s 00:08:43.749 sys 0m1.932s 00:08:43.749 ************************************ 00:08:43.749 END TEST raid_state_function_test_sb 00:08:43.749 ************************************ 00:08:43.749 17:30:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1126 -- # xtrace_disable 00:08:43.749 17:30:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:08:44.010 17:30:14 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test raid0 3 00:08:44.010 17:30:14 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:08:44.010 17:30:14 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:08:44.010 17:30:14 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:44.010 ************************************ 00:08:44.010 START TEST raid_superblock_test 00:08:44.010 ************************************ 00:08:44.010 17:30:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1125 -- # raid_superblock_test raid0 3 00:08:44.010 17:30:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=raid0 00:08:44.010 17:30:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=3 00:08:44.010 17:30:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:08:44.010 17:30:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:08:44.010 17:30:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:08:44.010 17:30:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:08:44.010 17:30:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:08:44.010 17:30:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:08:44.010 17:30:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:08:44.010 17:30:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:08:44.010 17:30:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:08:44.010 17:30:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:08:44.010 17:30:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:08:44.010 17:30:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' raid0 '!=' raid1 ']' 00:08:44.010 17:30:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@405 -- # strip_size=64 00:08:44.010 17:30:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@406 -- # strip_size_create_arg='-z 64' 00:08:44.010 17:30:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=76064 00:08:44.010 17:30:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:08:44.010 17:30:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 76064 00:08:44.010 17:30:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@831 -- # '[' -z 76064 ']' 00:08:44.010 17:30:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:44.010 17:30:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:08:44.010 17:30:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:44.010 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:44.010 17:30:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:08:44.010 17:30:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:44.010 [2024-11-27 17:30:15.056625] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:08:44.010 [2024-11-27 17:30:15.056748] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid76064 ] 00:08:44.270 [2024-11-27 17:30:15.203036] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:44.270 [2024-11-27 17:30:15.272331] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:08:44.270 [2024-11-27 17:30:15.349303] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:44.270 [2024-11-27 17:30:15.349354] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:44.841 17:30:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:08:44.841 17:30:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@864 -- # return 0 00:08:44.841 17:30:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:08:44.841 17:30:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:08:44.841 17:30:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:08:44.841 17:30:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:08:44.841 17:30:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:08:44.841 17:30:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:08:44.841 17:30:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:08:44.841 17:30:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:08:44.841 17:30:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:08:44.841 17:30:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:44.841 17:30:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:44.841 malloc1 00:08:44.841 17:30:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:44.841 17:30:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:08:44.841 17:30:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:44.841 17:30:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:44.841 [2024-11-27 17:30:15.911552] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:08:44.841 [2024-11-27 17:30:15.911699] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:44.841 [2024-11-27 17:30:15.911753] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:08:44.841 [2024-11-27 17:30:15.911793] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:44.841 [2024-11-27 17:30:15.914243] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:44.841 [2024-11-27 17:30:15.914317] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:08:44.841 pt1 00:08:44.841 17:30:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:44.841 17:30:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:08:44.841 17:30:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:08:44.841 17:30:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:08:44.841 17:30:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:08:44.841 17:30:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:08:44.841 17:30:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:08:44.841 17:30:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:08:44.841 17:30:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:08:44.841 17:30:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:08:44.841 17:30:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:44.841 17:30:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:44.841 malloc2 00:08:44.841 17:30:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:44.841 17:30:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:08:44.841 17:30:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:44.841 17:30:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:44.841 [2024-11-27 17:30:15.961732] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:08:44.841 [2024-11-27 17:30:15.961891] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:44.841 [2024-11-27 17:30:15.961952] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:08:44.841 [2024-11-27 17:30:15.962011] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:44.841 [2024-11-27 17:30:15.965959] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:44.841 [2024-11-27 17:30:15.966072] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:08:44.841 pt2 00:08:44.841 17:30:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:44.841 17:30:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:08:44.841 17:30:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:08:44.841 17:30:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc3 00:08:44.841 17:30:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt3 00:08:44.841 17:30:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000003 00:08:44.841 17:30:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:08:44.841 17:30:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:08:44.841 17:30:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:08:44.841 17:30:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc3 00:08:44.841 17:30:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:44.841 17:30:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:44.841 malloc3 00:08:44.841 17:30:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:44.841 17:30:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:08:44.841 17:30:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:44.841 17:30:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:44.841 [2024-11-27 17:30:15.996915] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:08:44.841 [2024-11-27 17:30:15.997011] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:44.841 [2024-11-27 17:30:15.997064] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:08:44.841 [2024-11-27 17:30:15.997094] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:44.841 [2024-11-27 17:30:15.999521] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:44.841 [2024-11-27 17:30:15.999593] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:08:44.841 pt3 00:08:44.841 17:30:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:44.841 17:30:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:08:44.841 17:30:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:08:44.841 17:30:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''pt1 pt2 pt3'\''' -n raid_bdev1 -s 00:08:44.841 17:30:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:44.841 17:30:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:44.841 [2024-11-27 17:30:16.008993] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:08:44.841 [2024-11-27 17:30:16.011213] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:08:44.841 [2024-11-27 17:30:16.011305] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:08:44.841 [2024-11-27 17:30:16.011491] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:08:44.841 [2024-11-27 17:30:16.011539] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:08:44.841 [2024-11-27 17:30:16.011861] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:08:44.841 [2024-11-27 17:30:16.012038] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:08:44.842 [2024-11-27 17:30:16.012085] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:08:44.842 [2024-11-27 17:30:16.012254] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:44.842 17:30:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:44.842 17:30:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 3 00:08:44.842 17:30:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:44.842 17:30:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:44.842 17:30:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:44.842 17:30:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:44.842 17:30:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:44.842 17:30:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:44.842 17:30:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:44.842 17:30:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:44.842 17:30:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:44.842 17:30:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:44.842 17:30:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:44.842 17:30:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:44.842 17:30:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:45.102 17:30:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:45.102 17:30:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:45.102 "name": "raid_bdev1", 00:08:45.102 "uuid": "dcbceef0-9cfd-4641-9474-4d79244524a7", 00:08:45.102 "strip_size_kb": 64, 00:08:45.102 "state": "online", 00:08:45.102 "raid_level": "raid0", 00:08:45.102 "superblock": true, 00:08:45.102 "num_base_bdevs": 3, 00:08:45.102 "num_base_bdevs_discovered": 3, 00:08:45.102 "num_base_bdevs_operational": 3, 00:08:45.102 "base_bdevs_list": [ 00:08:45.102 { 00:08:45.102 "name": "pt1", 00:08:45.102 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:45.102 "is_configured": true, 00:08:45.102 "data_offset": 2048, 00:08:45.102 "data_size": 63488 00:08:45.102 }, 00:08:45.102 { 00:08:45.102 "name": "pt2", 00:08:45.102 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:45.102 "is_configured": true, 00:08:45.102 "data_offset": 2048, 00:08:45.102 "data_size": 63488 00:08:45.102 }, 00:08:45.102 { 00:08:45.102 "name": "pt3", 00:08:45.102 "uuid": "00000000-0000-0000-0000-000000000003", 00:08:45.102 "is_configured": true, 00:08:45.102 "data_offset": 2048, 00:08:45.102 "data_size": 63488 00:08:45.102 } 00:08:45.102 ] 00:08:45.102 }' 00:08:45.102 17:30:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:45.102 17:30:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:45.362 17:30:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:08:45.362 17:30:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:08:45.362 17:30:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:45.362 17:30:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:45.362 17:30:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:08:45.362 17:30:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:45.362 17:30:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:08:45.362 17:30:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:45.362 17:30:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:45.362 17:30:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:45.362 [2024-11-27 17:30:16.452541] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:45.362 17:30:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:45.362 17:30:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:45.362 "name": "raid_bdev1", 00:08:45.362 "aliases": [ 00:08:45.362 "dcbceef0-9cfd-4641-9474-4d79244524a7" 00:08:45.362 ], 00:08:45.362 "product_name": "Raid Volume", 00:08:45.362 "block_size": 512, 00:08:45.362 "num_blocks": 190464, 00:08:45.362 "uuid": "dcbceef0-9cfd-4641-9474-4d79244524a7", 00:08:45.362 "assigned_rate_limits": { 00:08:45.362 "rw_ios_per_sec": 0, 00:08:45.362 "rw_mbytes_per_sec": 0, 00:08:45.362 "r_mbytes_per_sec": 0, 00:08:45.362 "w_mbytes_per_sec": 0 00:08:45.362 }, 00:08:45.362 "claimed": false, 00:08:45.362 "zoned": false, 00:08:45.362 "supported_io_types": { 00:08:45.362 "read": true, 00:08:45.362 "write": true, 00:08:45.362 "unmap": true, 00:08:45.362 "flush": true, 00:08:45.362 "reset": true, 00:08:45.362 "nvme_admin": false, 00:08:45.362 "nvme_io": false, 00:08:45.362 "nvme_io_md": false, 00:08:45.362 "write_zeroes": true, 00:08:45.362 "zcopy": false, 00:08:45.362 "get_zone_info": false, 00:08:45.362 "zone_management": false, 00:08:45.362 "zone_append": false, 00:08:45.362 "compare": false, 00:08:45.362 "compare_and_write": false, 00:08:45.362 "abort": false, 00:08:45.362 "seek_hole": false, 00:08:45.362 "seek_data": false, 00:08:45.362 "copy": false, 00:08:45.362 "nvme_iov_md": false 00:08:45.362 }, 00:08:45.362 "memory_domains": [ 00:08:45.362 { 00:08:45.362 "dma_device_id": "system", 00:08:45.362 "dma_device_type": 1 00:08:45.362 }, 00:08:45.362 { 00:08:45.362 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:45.362 "dma_device_type": 2 00:08:45.362 }, 00:08:45.362 { 00:08:45.362 "dma_device_id": "system", 00:08:45.362 "dma_device_type": 1 00:08:45.362 }, 00:08:45.362 { 00:08:45.362 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:45.362 "dma_device_type": 2 00:08:45.362 }, 00:08:45.362 { 00:08:45.362 "dma_device_id": "system", 00:08:45.362 "dma_device_type": 1 00:08:45.362 }, 00:08:45.362 { 00:08:45.362 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:45.362 "dma_device_type": 2 00:08:45.362 } 00:08:45.362 ], 00:08:45.362 "driver_specific": { 00:08:45.362 "raid": { 00:08:45.362 "uuid": "dcbceef0-9cfd-4641-9474-4d79244524a7", 00:08:45.362 "strip_size_kb": 64, 00:08:45.362 "state": "online", 00:08:45.362 "raid_level": "raid0", 00:08:45.362 "superblock": true, 00:08:45.362 "num_base_bdevs": 3, 00:08:45.362 "num_base_bdevs_discovered": 3, 00:08:45.362 "num_base_bdevs_operational": 3, 00:08:45.362 "base_bdevs_list": [ 00:08:45.362 { 00:08:45.362 "name": "pt1", 00:08:45.362 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:45.362 "is_configured": true, 00:08:45.362 "data_offset": 2048, 00:08:45.362 "data_size": 63488 00:08:45.362 }, 00:08:45.362 { 00:08:45.362 "name": "pt2", 00:08:45.362 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:45.362 "is_configured": true, 00:08:45.362 "data_offset": 2048, 00:08:45.362 "data_size": 63488 00:08:45.362 }, 00:08:45.362 { 00:08:45.362 "name": "pt3", 00:08:45.362 "uuid": "00000000-0000-0000-0000-000000000003", 00:08:45.362 "is_configured": true, 00:08:45.362 "data_offset": 2048, 00:08:45.362 "data_size": 63488 00:08:45.362 } 00:08:45.362 ] 00:08:45.363 } 00:08:45.363 } 00:08:45.363 }' 00:08:45.363 17:30:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:45.363 17:30:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:08:45.363 pt2 00:08:45.363 pt3' 00:08:45.363 17:30:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:45.623 17:30:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:45.623 17:30:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:45.623 17:30:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:08:45.623 17:30:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:45.623 17:30:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:45.623 17:30:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:45.623 17:30:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:45.623 17:30:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:45.623 17:30:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:45.623 17:30:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:45.624 17:30:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:45.624 17:30:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:08:45.624 17:30:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:45.624 17:30:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:45.624 17:30:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:45.624 17:30:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:45.624 17:30:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:45.624 17:30:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:45.624 17:30:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:45.624 17:30:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:08:45.624 17:30:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:45.624 17:30:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:45.624 17:30:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:45.624 17:30:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:45.624 17:30:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:45.624 17:30:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:08:45.624 17:30:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:08:45.624 17:30:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:45.624 17:30:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:45.624 [2024-11-27 17:30:16.719981] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:45.624 17:30:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:45.624 17:30:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=dcbceef0-9cfd-4641-9474-4d79244524a7 00:08:45.624 17:30:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z dcbceef0-9cfd-4641-9474-4d79244524a7 ']' 00:08:45.624 17:30:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:08:45.624 17:30:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:45.624 17:30:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:45.624 [2024-11-27 17:30:16.763643] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:08:45.624 [2024-11-27 17:30:16.763719] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:45.624 [2024-11-27 17:30:16.763822] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:45.624 [2024-11-27 17:30:16.763900] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:45.624 [2024-11-27 17:30:16.763952] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:08:45.624 17:30:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:45.624 17:30:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:08:45.624 17:30:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:45.624 17:30:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:45.624 17:30:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:45.624 17:30:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:45.624 17:30:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:08:45.624 17:30:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:08:45.624 17:30:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:08:45.624 17:30:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:08:45.624 17:30:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:45.624 17:30:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:45.624 17:30:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:45.624 17:30:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:08:45.624 17:30:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:08:45.885 17:30:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:45.885 17:30:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:45.885 17:30:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:45.885 17:30:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:08:45.885 17:30:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt3 00:08:45.885 17:30:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:45.885 17:30:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:45.885 17:30:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:45.885 17:30:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:08:45.885 17:30:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:08:45.885 17:30:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:45.885 17:30:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:45.885 17:30:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:45.885 17:30:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:08:45.885 17:30:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:08:45.885 17:30:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@650 -- # local es=0 00:08:45.885 17:30:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:08:45.885 17:30:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:08:45.885 17:30:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:08:45.885 17:30:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:08:45.885 17:30:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:08:45.885 17:30:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:08:45.885 17:30:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:45.885 17:30:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:45.885 [2024-11-27 17:30:16.899449] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:08:45.885 [2024-11-27 17:30:16.901630] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:08:45.885 [2024-11-27 17:30:16.901667] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc3 is claimed 00:08:45.885 [2024-11-27 17:30:16.901726] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:08:45.885 [2024-11-27 17:30:16.901776] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:08:45.885 [2024-11-27 17:30:16.901797] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc3 00:08:45.885 [2024-11-27 17:30:16.901809] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:08:45.885 [2024-11-27 17:30:16.901818] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state configuring 00:08:45.885 request: 00:08:45.885 { 00:08:45.885 "name": "raid_bdev1", 00:08:45.885 "raid_level": "raid0", 00:08:45.885 "base_bdevs": [ 00:08:45.885 "malloc1", 00:08:45.885 "malloc2", 00:08:45.885 "malloc3" 00:08:45.885 ], 00:08:45.885 "strip_size_kb": 64, 00:08:45.885 "superblock": false, 00:08:45.885 "method": "bdev_raid_create", 00:08:45.885 "req_id": 1 00:08:45.885 } 00:08:45.885 Got JSON-RPC error response 00:08:45.885 response: 00:08:45.885 { 00:08:45.885 "code": -17, 00:08:45.885 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:08:45.885 } 00:08:45.885 17:30:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:08:45.885 17:30:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@653 -- # es=1 00:08:45.885 17:30:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:08:45.885 17:30:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:08:45.885 17:30:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:08:45.885 17:30:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:08:45.885 17:30:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:45.885 17:30:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:45.885 17:30:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:45.885 17:30:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:45.885 17:30:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:08:45.885 17:30:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:08:45.885 17:30:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:08:45.885 17:30:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:45.885 17:30:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:45.885 [2024-11-27 17:30:16.959315] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:08:45.885 [2024-11-27 17:30:16.959419] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:45.885 [2024-11-27 17:30:16.959478] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:08:45.885 [2024-11-27 17:30:16.959513] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:45.885 [2024-11-27 17:30:16.962037] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:45.885 [2024-11-27 17:30:16.962109] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:08:45.885 [2024-11-27 17:30:16.962210] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:08:45.885 [2024-11-27 17:30:16.962284] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:08:45.885 pt1 00:08:45.885 17:30:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:45.885 17:30:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid0 64 3 00:08:45.885 17:30:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:45.885 17:30:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:45.885 17:30:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:45.885 17:30:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:45.885 17:30:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:45.885 17:30:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:45.885 17:30:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:45.885 17:30:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:45.885 17:30:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:45.885 17:30:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:45.885 17:30:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:45.885 17:30:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:45.885 17:30:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:45.885 17:30:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:45.885 17:30:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:45.885 "name": "raid_bdev1", 00:08:45.885 "uuid": "dcbceef0-9cfd-4641-9474-4d79244524a7", 00:08:45.885 "strip_size_kb": 64, 00:08:45.885 "state": "configuring", 00:08:45.885 "raid_level": "raid0", 00:08:45.885 "superblock": true, 00:08:45.885 "num_base_bdevs": 3, 00:08:45.885 "num_base_bdevs_discovered": 1, 00:08:45.885 "num_base_bdevs_operational": 3, 00:08:45.885 "base_bdevs_list": [ 00:08:45.885 { 00:08:45.886 "name": "pt1", 00:08:45.886 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:45.886 "is_configured": true, 00:08:45.886 "data_offset": 2048, 00:08:45.886 "data_size": 63488 00:08:45.886 }, 00:08:45.886 { 00:08:45.886 "name": null, 00:08:45.886 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:45.886 "is_configured": false, 00:08:45.886 "data_offset": 2048, 00:08:45.886 "data_size": 63488 00:08:45.886 }, 00:08:45.886 { 00:08:45.886 "name": null, 00:08:45.886 "uuid": "00000000-0000-0000-0000-000000000003", 00:08:45.886 "is_configured": false, 00:08:45.886 "data_offset": 2048, 00:08:45.886 "data_size": 63488 00:08:45.886 } 00:08:45.886 ] 00:08:45.886 }' 00:08:45.886 17:30:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:45.886 17:30:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:46.455 17:30:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 3 -gt 2 ']' 00:08:46.455 17:30:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@472 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:08:46.455 17:30:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:46.455 17:30:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:46.455 [2024-11-27 17:30:17.430571] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:08:46.455 [2024-11-27 17:30:17.430679] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:46.455 [2024-11-27 17:30:17.430719] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:08:46.455 [2024-11-27 17:30:17.430758] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:46.455 [2024-11-27 17:30:17.431225] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:46.455 [2024-11-27 17:30:17.431291] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:08:46.455 [2024-11-27 17:30:17.431369] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:08:46.455 [2024-11-27 17:30:17.431398] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:08:46.455 pt2 00:08:46.455 17:30:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:46.455 17:30:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@473 -- # rpc_cmd bdev_passthru_delete pt2 00:08:46.455 17:30:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:46.455 17:30:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:46.455 [2024-11-27 17:30:17.442550] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: pt2 00:08:46.455 17:30:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:46.455 17:30:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@474 -- # verify_raid_bdev_state raid_bdev1 configuring raid0 64 3 00:08:46.455 17:30:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:46.455 17:30:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:46.455 17:30:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:46.455 17:30:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:46.455 17:30:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:46.455 17:30:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:46.455 17:30:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:46.455 17:30:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:46.455 17:30:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:46.456 17:30:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:46.456 17:30:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:46.456 17:30:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:46.456 17:30:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:46.456 17:30:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:46.456 17:30:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:46.456 "name": "raid_bdev1", 00:08:46.456 "uuid": "dcbceef0-9cfd-4641-9474-4d79244524a7", 00:08:46.456 "strip_size_kb": 64, 00:08:46.456 "state": "configuring", 00:08:46.456 "raid_level": "raid0", 00:08:46.456 "superblock": true, 00:08:46.456 "num_base_bdevs": 3, 00:08:46.456 "num_base_bdevs_discovered": 1, 00:08:46.456 "num_base_bdevs_operational": 3, 00:08:46.456 "base_bdevs_list": [ 00:08:46.456 { 00:08:46.456 "name": "pt1", 00:08:46.456 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:46.456 "is_configured": true, 00:08:46.456 "data_offset": 2048, 00:08:46.456 "data_size": 63488 00:08:46.456 }, 00:08:46.456 { 00:08:46.456 "name": null, 00:08:46.456 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:46.456 "is_configured": false, 00:08:46.456 "data_offset": 0, 00:08:46.456 "data_size": 63488 00:08:46.456 }, 00:08:46.456 { 00:08:46.456 "name": null, 00:08:46.456 "uuid": "00000000-0000-0000-0000-000000000003", 00:08:46.456 "is_configured": false, 00:08:46.456 "data_offset": 2048, 00:08:46.456 "data_size": 63488 00:08:46.456 } 00:08:46.456 ] 00:08:46.456 }' 00:08:46.456 17:30:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:46.456 17:30:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:46.716 17:30:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:08:46.716 17:30:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:08:46.716 17:30:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:08:46.716 17:30:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:46.716 17:30:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:46.716 [2024-11-27 17:30:17.877780] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:08:46.716 [2024-11-27 17:30:17.877890] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:46.716 [2024-11-27 17:30:17.877928] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:08:46.716 [2024-11-27 17:30:17.877955] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:46.716 [2024-11-27 17:30:17.878407] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:46.716 [2024-11-27 17:30:17.878458] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:08:46.716 [2024-11-27 17:30:17.878549] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:08:46.716 [2024-11-27 17:30:17.878595] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:08:46.716 pt2 00:08:46.716 17:30:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:46.716 17:30:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:08:46.716 17:30:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:08:46.716 17:30:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:08:46.716 17:30:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:46.716 17:30:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:46.716 [2024-11-27 17:30:17.889757] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:08:46.716 [2024-11-27 17:30:17.889832] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:46.716 [2024-11-27 17:30:17.889881] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:08:46.716 [2024-11-27 17:30:17.889907] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:46.716 [2024-11-27 17:30:17.890284] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:46.716 [2024-11-27 17:30:17.890334] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:08:46.716 [2024-11-27 17:30:17.890412] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:08:46.716 [2024-11-27 17:30:17.890468] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:08:46.716 [2024-11-27 17:30:17.890593] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:08:46.716 [2024-11-27 17:30:17.890628] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:08:46.716 [2024-11-27 17:30:17.890901] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:08:46.716 [2024-11-27 17:30:17.891043] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:08:46.716 [2024-11-27 17:30:17.891082] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:08:46.716 [2024-11-27 17:30:17.891232] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:46.716 pt3 00:08:46.716 17:30:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:46.716 17:30:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:08:46.716 17:30:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:08:46.716 17:30:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 3 00:08:46.716 17:30:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:46.716 17:30:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:46.716 17:30:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:46.716 17:30:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:46.716 17:30:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:46.716 17:30:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:46.716 17:30:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:46.716 17:30:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:46.716 17:30:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:46.716 17:30:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:46.716 17:30:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:46.716 17:30:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:46.716 17:30:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:46.977 17:30:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:46.977 17:30:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:46.977 "name": "raid_bdev1", 00:08:46.977 "uuid": "dcbceef0-9cfd-4641-9474-4d79244524a7", 00:08:46.977 "strip_size_kb": 64, 00:08:46.977 "state": "online", 00:08:46.977 "raid_level": "raid0", 00:08:46.977 "superblock": true, 00:08:46.977 "num_base_bdevs": 3, 00:08:46.977 "num_base_bdevs_discovered": 3, 00:08:46.977 "num_base_bdevs_operational": 3, 00:08:46.977 "base_bdevs_list": [ 00:08:46.977 { 00:08:46.977 "name": "pt1", 00:08:46.977 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:46.977 "is_configured": true, 00:08:46.977 "data_offset": 2048, 00:08:46.977 "data_size": 63488 00:08:46.977 }, 00:08:46.977 { 00:08:46.977 "name": "pt2", 00:08:46.977 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:46.977 "is_configured": true, 00:08:46.977 "data_offset": 2048, 00:08:46.977 "data_size": 63488 00:08:46.977 }, 00:08:46.977 { 00:08:46.977 "name": "pt3", 00:08:46.977 "uuid": "00000000-0000-0000-0000-000000000003", 00:08:46.977 "is_configured": true, 00:08:46.977 "data_offset": 2048, 00:08:46.977 "data_size": 63488 00:08:46.977 } 00:08:46.977 ] 00:08:46.977 }' 00:08:46.977 17:30:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:46.977 17:30:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:47.237 17:30:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:08:47.237 17:30:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:08:47.237 17:30:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:47.237 17:30:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:47.237 17:30:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:08:47.237 17:30:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:47.237 17:30:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:08:47.237 17:30:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:47.237 17:30:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:47.237 17:30:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:47.237 [2024-11-27 17:30:18.385225] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:47.237 17:30:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:47.237 17:30:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:47.237 "name": "raid_bdev1", 00:08:47.237 "aliases": [ 00:08:47.237 "dcbceef0-9cfd-4641-9474-4d79244524a7" 00:08:47.237 ], 00:08:47.237 "product_name": "Raid Volume", 00:08:47.237 "block_size": 512, 00:08:47.237 "num_blocks": 190464, 00:08:47.237 "uuid": "dcbceef0-9cfd-4641-9474-4d79244524a7", 00:08:47.237 "assigned_rate_limits": { 00:08:47.237 "rw_ios_per_sec": 0, 00:08:47.237 "rw_mbytes_per_sec": 0, 00:08:47.237 "r_mbytes_per_sec": 0, 00:08:47.237 "w_mbytes_per_sec": 0 00:08:47.237 }, 00:08:47.237 "claimed": false, 00:08:47.237 "zoned": false, 00:08:47.237 "supported_io_types": { 00:08:47.237 "read": true, 00:08:47.237 "write": true, 00:08:47.237 "unmap": true, 00:08:47.237 "flush": true, 00:08:47.237 "reset": true, 00:08:47.237 "nvme_admin": false, 00:08:47.237 "nvme_io": false, 00:08:47.237 "nvme_io_md": false, 00:08:47.237 "write_zeroes": true, 00:08:47.237 "zcopy": false, 00:08:47.237 "get_zone_info": false, 00:08:47.237 "zone_management": false, 00:08:47.237 "zone_append": false, 00:08:47.237 "compare": false, 00:08:47.237 "compare_and_write": false, 00:08:47.237 "abort": false, 00:08:47.237 "seek_hole": false, 00:08:47.237 "seek_data": false, 00:08:47.237 "copy": false, 00:08:47.237 "nvme_iov_md": false 00:08:47.237 }, 00:08:47.237 "memory_domains": [ 00:08:47.237 { 00:08:47.237 "dma_device_id": "system", 00:08:47.237 "dma_device_type": 1 00:08:47.237 }, 00:08:47.237 { 00:08:47.237 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:47.237 "dma_device_type": 2 00:08:47.237 }, 00:08:47.237 { 00:08:47.237 "dma_device_id": "system", 00:08:47.237 "dma_device_type": 1 00:08:47.237 }, 00:08:47.237 { 00:08:47.237 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:47.237 "dma_device_type": 2 00:08:47.237 }, 00:08:47.237 { 00:08:47.237 "dma_device_id": "system", 00:08:47.237 "dma_device_type": 1 00:08:47.237 }, 00:08:47.237 { 00:08:47.237 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:47.237 "dma_device_type": 2 00:08:47.237 } 00:08:47.237 ], 00:08:47.237 "driver_specific": { 00:08:47.237 "raid": { 00:08:47.237 "uuid": "dcbceef0-9cfd-4641-9474-4d79244524a7", 00:08:47.237 "strip_size_kb": 64, 00:08:47.237 "state": "online", 00:08:47.237 "raid_level": "raid0", 00:08:47.237 "superblock": true, 00:08:47.237 "num_base_bdevs": 3, 00:08:47.237 "num_base_bdevs_discovered": 3, 00:08:47.237 "num_base_bdevs_operational": 3, 00:08:47.237 "base_bdevs_list": [ 00:08:47.237 { 00:08:47.237 "name": "pt1", 00:08:47.237 "uuid": "00000000-0000-0000-0000-000000000001", 00:08:47.237 "is_configured": true, 00:08:47.237 "data_offset": 2048, 00:08:47.237 "data_size": 63488 00:08:47.237 }, 00:08:47.237 { 00:08:47.237 "name": "pt2", 00:08:47.237 "uuid": "00000000-0000-0000-0000-000000000002", 00:08:47.237 "is_configured": true, 00:08:47.237 "data_offset": 2048, 00:08:47.237 "data_size": 63488 00:08:47.237 }, 00:08:47.237 { 00:08:47.237 "name": "pt3", 00:08:47.237 "uuid": "00000000-0000-0000-0000-000000000003", 00:08:47.237 "is_configured": true, 00:08:47.237 "data_offset": 2048, 00:08:47.237 "data_size": 63488 00:08:47.237 } 00:08:47.237 ] 00:08:47.237 } 00:08:47.237 } 00:08:47.237 }' 00:08:47.237 17:30:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:47.497 17:30:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:08:47.497 pt2 00:08:47.497 pt3' 00:08:47.497 17:30:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:47.498 17:30:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:47.498 17:30:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:47.498 17:30:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:47.498 17:30:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:08:47.498 17:30:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:47.498 17:30:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:47.498 17:30:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:47.498 17:30:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:47.498 17:30:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:47.498 17:30:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:47.498 17:30:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:08:47.498 17:30:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:47.498 17:30:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:47.498 17:30:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:47.498 17:30:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:47.498 17:30:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:47.498 17:30:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:47.498 17:30:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:47.498 17:30:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:08:47.498 17:30:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:47.498 17:30:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:47.498 17:30:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:47.498 17:30:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:47.498 17:30:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:47.498 17:30:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:47.498 17:30:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:08:47.498 17:30:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:47.498 17:30:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:08:47.498 17:30:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:47.498 [2024-11-27 17:30:18.648710] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:47.498 17:30:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:47.758 17:30:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' dcbceef0-9cfd-4641-9474-4d79244524a7 '!=' dcbceef0-9cfd-4641-9474-4d79244524a7 ']' 00:08:47.758 17:30:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy raid0 00:08:47.758 17:30:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:08:47.758 17:30:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # return 1 00:08:47.758 17:30:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 76064 00:08:47.758 17:30:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@950 -- # '[' -z 76064 ']' 00:08:47.758 17:30:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # kill -0 76064 00:08:47.758 17:30:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@955 -- # uname 00:08:47.758 17:30:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:08:47.758 17:30:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 76064 00:08:47.758 17:30:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:08:47.758 17:30:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:08:47.758 17:30:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 76064' 00:08:47.758 killing process with pid 76064 00:08:47.758 17:30:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@969 -- # kill 76064 00:08:47.758 [2024-11-27 17:30:18.719804] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:47.758 [2024-11-27 17:30:18.719959] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:47.758 17:30:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@974 -- # wait 76064 00:08:47.758 [2024-11-27 17:30:18.720067] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:47.758 [2024-11-27 17:30:18.720079] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:08:47.758 [2024-11-27 17:30:18.780474] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:48.018 17:30:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:08:48.018 00:08:48.018 real 0m4.183s 00:08:48.018 user 0m6.404s 00:08:48.018 sys 0m0.948s 00:08:48.018 17:30:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:08:48.018 17:30:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:08:48.018 ************************************ 00:08:48.018 END TEST raid_superblock_test 00:08:48.018 ************************************ 00:08:48.290 17:30:19 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test raid0 3 read 00:08:48.290 17:30:19 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:08:48.290 17:30:19 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:08:48.290 17:30:19 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:48.290 ************************************ 00:08:48.290 START TEST raid_read_error_test 00:08:48.290 ************************************ 00:08:48.290 17:30:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1125 -- # raid_io_error_test raid0 3 read 00:08:48.290 17:30:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid0 00:08:48.290 17:30:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=3 00:08:48.290 17:30:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:08:48.290 17:30:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:08:48.290 17:30:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:48.290 17:30:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:08:48.290 17:30:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:08:48.290 17:30:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:48.290 17:30:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:08:48.290 17:30:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:08:48.290 17:30:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:48.290 17:30:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:08:48.290 17:30:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:08:48.290 17:30:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:48.290 17:30:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:08:48.290 17:30:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:08:48.290 17:30:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:08:48.290 17:30:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:08:48.290 17:30:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:08:48.290 17:30:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:08:48.290 17:30:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:08:48.290 17:30:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid0 '!=' raid1 ']' 00:08:48.290 17:30:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:08:48.290 17:30:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:08:48.290 17:30:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:08:48.290 17:30:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.XHleTRieJD 00:08:48.290 17:30:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=76306 00:08:48.291 17:30:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:08:48.291 17:30:19 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 76306 00:08:48.291 17:30:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@831 -- # '[' -z 76306 ']' 00:08:48.291 17:30:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:48.291 17:30:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:08:48.291 17:30:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:48.291 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:48.291 17:30:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:08:48.291 17:30:19 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:48.291 [2024-11-27 17:30:19.334714] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:08:48.291 [2024-11-27 17:30:19.334885] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid76306 ] 00:08:48.571 [2024-11-27 17:30:19.480640] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:48.571 [2024-11-27 17:30:19.551383] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:08:48.571 [2024-11-27 17:30:19.629033] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:48.571 [2024-11-27 17:30:19.629071] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:49.141 17:30:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:08:49.141 17:30:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@864 -- # return 0 00:08:49.141 17:30:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:08:49.141 17:30:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:08:49.141 17:30:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:49.141 17:30:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:49.141 BaseBdev1_malloc 00:08:49.141 17:30:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:49.141 17:30:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:08:49.141 17:30:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:49.141 17:30:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:49.141 true 00:08:49.141 17:30:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:49.141 17:30:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:08:49.141 17:30:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:49.141 17:30:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:49.141 [2024-11-27 17:30:20.195021] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:08:49.141 [2024-11-27 17:30:20.195098] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:49.141 [2024-11-27 17:30:20.195128] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:08:49.141 [2024-11-27 17:30:20.195149] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:49.141 [2024-11-27 17:30:20.197544] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:49.141 [2024-11-27 17:30:20.197578] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:08:49.141 BaseBdev1 00:08:49.141 17:30:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:49.141 17:30:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:08:49.141 17:30:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:08:49.141 17:30:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:49.141 17:30:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:49.141 BaseBdev2_malloc 00:08:49.141 17:30:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:49.141 17:30:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:08:49.141 17:30:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:49.141 17:30:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:49.141 true 00:08:49.141 17:30:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:49.141 17:30:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:08:49.141 17:30:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:49.141 17:30:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:49.141 [2024-11-27 17:30:20.257752] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:08:49.141 [2024-11-27 17:30:20.257840] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:49.141 [2024-11-27 17:30:20.257875] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:08:49.142 [2024-11-27 17:30:20.257891] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:49.142 [2024-11-27 17:30:20.261476] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:49.142 [2024-11-27 17:30:20.261526] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:08:49.142 BaseBdev2 00:08:49.142 17:30:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:49.142 17:30:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:08:49.142 17:30:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:08:49.142 17:30:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:49.142 17:30:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:49.142 BaseBdev3_malloc 00:08:49.142 17:30:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:49.142 17:30:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:08:49.142 17:30:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:49.142 17:30:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:49.142 true 00:08:49.142 17:30:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:49.142 17:30:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:08:49.142 17:30:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:49.142 17:30:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:49.142 [2024-11-27 17:30:20.304248] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:08:49.142 [2024-11-27 17:30:20.304350] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:49.142 [2024-11-27 17:30:20.304404] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008780 00:08:49.142 [2024-11-27 17:30:20.304432] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:49.142 [2024-11-27 17:30:20.306795] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:49.142 [2024-11-27 17:30:20.306864] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:08:49.142 BaseBdev3 00:08:49.142 17:30:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:49.142 17:30:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n raid_bdev1 -s 00:08:49.142 17:30:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:49.142 17:30:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:49.142 [2024-11-27 17:30:20.316329] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:49.142 [2024-11-27 17:30:20.318472] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:49.142 [2024-11-27 17:30:20.318584] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:49.142 [2024-11-27 17:30:20.318812] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:08:49.142 [2024-11-27 17:30:20.318862] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:08:49.142 [2024-11-27 17:30:20.319148] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:08:49.142 [2024-11-27 17:30:20.319334] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:08:49.142 [2024-11-27 17:30:20.319375] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001c80 00:08:49.142 [2024-11-27 17:30:20.319552] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:49.142 17:30:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:49.142 17:30:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 3 00:08:49.142 17:30:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:49.142 17:30:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:49.142 17:30:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:49.142 17:30:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:49.142 17:30:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:49.142 17:30:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:49.142 17:30:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:49.142 17:30:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:49.142 17:30:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:49.142 17:30:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:49.142 17:30:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:49.142 17:30:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:49.142 17:30:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:49.401 17:30:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:49.401 17:30:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:49.401 "name": "raid_bdev1", 00:08:49.401 "uuid": "ed058b4e-a7f1-40b3-9b32-0418fa98160a", 00:08:49.401 "strip_size_kb": 64, 00:08:49.401 "state": "online", 00:08:49.401 "raid_level": "raid0", 00:08:49.401 "superblock": true, 00:08:49.401 "num_base_bdevs": 3, 00:08:49.401 "num_base_bdevs_discovered": 3, 00:08:49.401 "num_base_bdevs_operational": 3, 00:08:49.401 "base_bdevs_list": [ 00:08:49.401 { 00:08:49.401 "name": "BaseBdev1", 00:08:49.401 "uuid": "95a84dde-2e81-50d2-aa29-8cf6ea7e619e", 00:08:49.401 "is_configured": true, 00:08:49.401 "data_offset": 2048, 00:08:49.401 "data_size": 63488 00:08:49.401 }, 00:08:49.401 { 00:08:49.401 "name": "BaseBdev2", 00:08:49.401 "uuid": "cfadaff9-f6a3-5963-938a-93fdd5a1e4a1", 00:08:49.401 "is_configured": true, 00:08:49.401 "data_offset": 2048, 00:08:49.401 "data_size": 63488 00:08:49.401 }, 00:08:49.401 { 00:08:49.401 "name": "BaseBdev3", 00:08:49.401 "uuid": "94d33efa-1ef1-5d59-a398-440c34c7bbbf", 00:08:49.401 "is_configured": true, 00:08:49.401 "data_offset": 2048, 00:08:49.401 "data_size": 63488 00:08:49.401 } 00:08:49.401 ] 00:08:49.401 }' 00:08:49.401 17:30:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:49.401 17:30:20 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:49.660 17:30:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:08:49.660 17:30:20 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:08:49.920 [2024-11-27 17:30:20.855864] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:08:50.861 17:30:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:08:50.861 17:30:21 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:50.861 17:30:21 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:50.861 17:30:21 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:50.861 17:30:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:08:50.861 17:30:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid0 = \r\a\i\d\1 ]] 00:08:50.861 17:30:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=3 00:08:50.861 17:30:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 3 00:08:50.861 17:30:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:50.861 17:30:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:50.861 17:30:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:50.861 17:30:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:50.861 17:30:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:50.861 17:30:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:50.861 17:30:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:50.861 17:30:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:50.861 17:30:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:50.861 17:30:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:50.861 17:30:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:50.861 17:30:21 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:50.861 17:30:21 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:50.861 17:30:21 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:50.861 17:30:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:50.861 "name": "raid_bdev1", 00:08:50.861 "uuid": "ed058b4e-a7f1-40b3-9b32-0418fa98160a", 00:08:50.861 "strip_size_kb": 64, 00:08:50.861 "state": "online", 00:08:50.861 "raid_level": "raid0", 00:08:50.861 "superblock": true, 00:08:50.861 "num_base_bdevs": 3, 00:08:50.861 "num_base_bdevs_discovered": 3, 00:08:50.861 "num_base_bdevs_operational": 3, 00:08:50.861 "base_bdevs_list": [ 00:08:50.861 { 00:08:50.861 "name": "BaseBdev1", 00:08:50.861 "uuid": "95a84dde-2e81-50d2-aa29-8cf6ea7e619e", 00:08:50.861 "is_configured": true, 00:08:50.861 "data_offset": 2048, 00:08:50.861 "data_size": 63488 00:08:50.861 }, 00:08:50.861 { 00:08:50.861 "name": "BaseBdev2", 00:08:50.861 "uuid": "cfadaff9-f6a3-5963-938a-93fdd5a1e4a1", 00:08:50.861 "is_configured": true, 00:08:50.861 "data_offset": 2048, 00:08:50.861 "data_size": 63488 00:08:50.861 }, 00:08:50.861 { 00:08:50.861 "name": "BaseBdev3", 00:08:50.861 "uuid": "94d33efa-1ef1-5d59-a398-440c34c7bbbf", 00:08:50.861 "is_configured": true, 00:08:50.861 "data_offset": 2048, 00:08:50.861 "data_size": 63488 00:08:50.861 } 00:08:50.861 ] 00:08:50.861 }' 00:08:50.861 17:30:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:50.861 17:30:21 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:51.121 17:30:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:08:51.121 17:30:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:51.121 17:30:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:51.121 [2024-11-27 17:30:22.244188] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:08:51.121 [2024-11-27 17:30:22.244301] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:51.121 [2024-11-27 17:30:22.246938] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:51.121 [2024-11-27 17:30:22.247027] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:51.121 [2024-11-27 17:30:22.247103] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:51.121 [2024-11-27 17:30:22.247199] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name raid_bdev1, state offline 00:08:51.121 { 00:08:51.121 "results": [ 00:08:51.121 { 00:08:51.121 "job": "raid_bdev1", 00:08:51.121 "core_mask": "0x1", 00:08:51.121 "workload": "randrw", 00:08:51.121 "percentage": 50, 00:08:51.121 "status": "finished", 00:08:51.121 "queue_depth": 1, 00:08:51.121 "io_size": 131072, 00:08:51.121 "runtime": 1.389049, 00:08:51.121 "iops": 15098.099491090667, 00:08:51.121 "mibps": 1887.2624363863333, 00:08:51.121 "io_failed": 1, 00:08:51.121 "io_timeout": 0, 00:08:51.121 "avg_latency_us": 92.90748558606334, 00:08:51.121 "min_latency_us": 24.929257641921396, 00:08:51.121 "max_latency_us": 1402.2986899563318 00:08:51.121 } 00:08:51.121 ], 00:08:51.121 "core_count": 1 00:08:51.121 } 00:08:51.121 17:30:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:51.121 17:30:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 76306 00:08:51.121 17:30:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@950 -- # '[' -z 76306 ']' 00:08:51.121 17:30:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # kill -0 76306 00:08:51.121 17:30:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@955 -- # uname 00:08:51.121 17:30:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:08:51.121 17:30:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 76306 00:08:51.121 17:30:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:08:51.121 17:30:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:08:51.121 killing process with pid 76306 00:08:51.122 17:30:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 76306' 00:08:51.122 17:30:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@969 -- # kill 76306 00:08:51.122 [2024-11-27 17:30:22.294507] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:51.122 17:30:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@974 -- # wait 76306 00:08:51.381 [2024-11-27 17:30:22.340783] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:51.641 17:30:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.XHleTRieJD 00:08:51.641 17:30:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:08:51.641 17:30:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:08:51.641 17:30:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.72 00:08:51.641 17:30:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid0 00:08:51.641 ************************************ 00:08:51.641 END TEST raid_read_error_test 00:08:51.641 ************************************ 00:08:51.641 17:30:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:08:51.641 17:30:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:08:51.641 17:30:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.72 != \0\.\0\0 ]] 00:08:51.641 00:08:51.641 real 0m3.487s 00:08:51.641 user 0m4.262s 00:08:51.641 sys 0m0.659s 00:08:51.641 17:30:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:08:51.641 17:30:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:51.641 17:30:22 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test raid0 3 write 00:08:51.641 17:30:22 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:08:51.641 17:30:22 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:08:51.641 17:30:22 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:51.641 ************************************ 00:08:51.641 START TEST raid_write_error_test 00:08:51.641 ************************************ 00:08:51.642 17:30:22 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1125 -- # raid_io_error_test raid0 3 write 00:08:51.642 17:30:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid0 00:08:51.642 17:30:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=3 00:08:51.642 17:30:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:08:51.642 17:30:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:08:51.642 17:30:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:51.642 17:30:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:08:51.642 17:30:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:08:51.642 17:30:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:51.642 17:30:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:08:51.642 17:30:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:08:51.642 17:30:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:51.642 17:30:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:08:51.642 17:30:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:08:51.642 17:30:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:08:51.642 17:30:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:08:51.642 17:30:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:08:51.642 17:30:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:08:51.642 17:30:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:08:51.642 17:30:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:08:51.642 17:30:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:08:51.642 17:30:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:08:51.642 17:30:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid0 '!=' raid1 ']' 00:08:51.642 17:30:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:08:51.642 17:30:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:08:51.642 17:30:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:08:51.642 17:30:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.KChe1ePlFi 00:08:51.642 17:30:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=76439 00:08:51.642 17:30:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 76439 00:08:51.642 17:30:22 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:08:51.642 17:30:22 bdev_raid.raid_write_error_test -- common/autotest_common.sh@831 -- # '[' -z 76439 ']' 00:08:51.642 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:51.642 17:30:22 bdev_raid.raid_write_error_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:51.642 17:30:22 bdev_raid.raid_write_error_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:08:51.642 17:30:22 bdev_raid.raid_write_error_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:51.642 17:30:22 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:08:51.642 17:30:22 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:51.901 [2024-11-27 17:30:22.885132] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:08:51.901 [2024-11-27 17:30:22.885264] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid76439 ] 00:08:51.901 [2024-11-27 17:30:23.029927] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:52.161 [2024-11-27 17:30:23.101708] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:08:52.161 [2024-11-27 17:30:23.179530] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:52.161 [2024-11-27 17:30:23.179575] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:52.731 17:30:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:08:52.731 17:30:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@864 -- # return 0 00:08:52.731 17:30:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:08:52.731 17:30:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:08:52.731 17:30:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:52.731 17:30:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:52.731 BaseBdev1_malloc 00:08:52.731 17:30:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:52.731 17:30:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:08:52.731 17:30:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:52.731 17:30:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:52.731 true 00:08:52.731 17:30:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:52.731 17:30:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:08:52.731 17:30:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:52.731 17:30:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:52.731 [2024-11-27 17:30:23.741896] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:08:52.731 [2024-11-27 17:30:23.742022] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:52.731 [2024-11-27 17:30:23.742089] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:08:52.731 [2024-11-27 17:30:23.742123] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:52.731 [2024-11-27 17:30:23.744565] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:52.731 [2024-11-27 17:30:23.744635] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:08:52.731 BaseBdev1 00:08:52.731 17:30:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:52.731 17:30:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:08:52.731 17:30:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:08:52.731 17:30:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:52.731 17:30:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:52.731 BaseBdev2_malloc 00:08:52.731 17:30:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:52.731 17:30:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:08:52.731 17:30:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:52.731 17:30:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:52.731 true 00:08:52.731 17:30:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:52.731 17:30:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:08:52.731 17:30:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:52.731 17:30:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:52.731 [2024-11-27 17:30:23.802679] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:08:52.731 [2024-11-27 17:30:23.802816] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:52.731 [2024-11-27 17:30:23.802864] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:08:52.731 [2024-11-27 17:30:23.802903] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:52.731 [2024-11-27 17:30:23.805557] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:52.731 [2024-11-27 17:30:23.805635] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:08:52.731 BaseBdev2 00:08:52.731 17:30:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:52.731 17:30:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:08:52.731 17:30:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:08:52.731 17:30:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:52.731 17:30:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:52.731 BaseBdev3_malloc 00:08:52.731 17:30:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:52.731 17:30:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:08:52.731 17:30:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:52.731 17:30:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:52.731 true 00:08:52.731 17:30:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:52.731 17:30:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:08:52.731 17:30:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:52.731 17:30:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:52.731 [2024-11-27 17:30:23.849573] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:08:52.732 [2024-11-27 17:30:23.849673] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:52.732 [2024-11-27 17:30:23.849712] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008780 00:08:52.732 [2024-11-27 17:30:23.849745] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:52.732 [2024-11-27 17:30:23.852219] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:52.732 [2024-11-27 17:30:23.852285] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:08:52.732 BaseBdev3 00:08:52.732 17:30:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:52.732 17:30:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n raid_bdev1 -s 00:08:52.732 17:30:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:52.732 17:30:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:52.732 [2024-11-27 17:30:23.861642] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:52.732 [2024-11-27 17:30:23.863754] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:52.732 [2024-11-27 17:30:23.863870] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:52.732 [2024-11-27 17:30:23.864065] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:08:52.732 [2024-11-27 17:30:23.864114] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:08:52.732 [2024-11-27 17:30:23.864392] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:08:52.732 [2024-11-27 17:30:23.864583] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:08:52.732 [2024-11-27 17:30:23.864623] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001c80 00:08:52.732 [2024-11-27 17:30:23.864791] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:52.732 17:30:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:52.732 17:30:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 3 00:08:52.732 17:30:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:52.732 17:30:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:52.732 17:30:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:52.732 17:30:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:52.732 17:30:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:52.732 17:30:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:52.732 17:30:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:52.732 17:30:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:52.732 17:30:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:52.732 17:30:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:52.732 17:30:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:52.732 17:30:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:52.732 17:30:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:52.732 17:30:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:52.732 17:30:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:52.732 "name": "raid_bdev1", 00:08:52.732 "uuid": "ca08761d-6cbe-4d38-a802-82b96d024357", 00:08:52.732 "strip_size_kb": 64, 00:08:52.732 "state": "online", 00:08:52.732 "raid_level": "raid0", 00:08:52.732 "superblock": true, 00:08:52.732 "num_base_bdevs": 3, 00:08:52.732 "num_base_bdevs_discovered": 3, 00:08:52.732 "num_base_bdevs_operational": 3, 00:08:52.732 "base_bdevs_list": [ 00:08:52.732 { 00:08:52.732 "name": "BaseBdev1", 00:08:52.732 "uuid": "fbabe8ba-02ac-5207-9625-b1d0c93f53fb", 00:08:52.732 "is_configured": true, 00:08:52.732 "data_offset": 2048, 00:08:52.732 "data_size": 63488 00:08:52.732 }, 00:08:52.732 { 00:08:52.732 "name": "BaseBdev2", 00:08:52.732 "uuid": "8f829372-5c09-5358-ad40-846dcdefa400", 00:08:52.732 "is_configured": true, 00:08:52.732 "data_offset": 2048, 00:08:52.732 "data_size": 63488 00:08:52.732 }, 00:08:52.732 { 00:08:52.732 "name": "BaseBdev3", 00:08:52.732 "uuid": "f596b92c-cc21-5960-bae7-51eb568bbaf8", 00:08:52.732 "is_configured": true, 00:08:52.732 "data_offset": 2048, 00:08:52.732 "data_size": 63488 00:08:52.732 } 00:08:52.732 ] 00:08:52.732 }' 00:08:52.732 17:30:23 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:52.732 17:30:23 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:53.301 17:30:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:08:53.301 17:30:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:08:53.301 [2024-11-27 17:30:24.397218] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:08:54.241 17:30:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:08:54.241 17:30:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:54.241 17:30:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:54.241 17:30:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:54.241 17:30:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:08:54.241 17:30:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid0 = \r\a\i\d\1 ]] 00:08:54.241 17:30:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=3 00:08:54.241 17:30:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 3 00:08:54.241 17:30:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:08:54.241 17:30:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:54.241 17:30:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:08:54.241 17:30:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:54.241 17:30:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:54.241 17:30:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:54.241 17:30:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:54.241 17:30:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:54.241 17:30:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:54.241 17:30:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:54.241 17:30:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:08:54.241 17:30:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:54.241 17:30:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:54.241 17:30:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:54.241 17:30:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:54.241 "name": "raid_bdev1", 00:08:54.241 "uuid": "ca08761d-6cbe-4d38-a802-82b96d024357", 00:08:54.241 "strip_size_kb": 64, 00:08:54.241 "state": "online", 00:08:54.241 "raid_level": "raid0", 00:08:54.241 "superblock": true, 00:08:54.241 "num_base_bdevs": 3, 00:08:54.241 "num_base_bdevs_discovered": 3, 00:08:54.241 "num_base_bdevs_operational": 3, 00:08:54.241 "base_bdevs_list": [ 00:08:54.241 { 00:08:54.241 "name": "BaseBdev1", 00:08:54.241 "uuid": "fbabe8ba-02ac-5207-9625-b1d0c93f53fb", 00:08:54.241 "is_configured": true, 00:08:54.241 "data_offset": 2048, 00:08:54.241 "data_size": 63488 00:08:54.241 }, 00:08:54.241 { 00:08:54.241 "name": "BaseBdev2", 00:08:54.241 "uuid": "8f829372-5c09-5358-ad40-846dcdefa400", 00:08:54.241 "is_configured": true, 00:08:54.241 "data_offset": 2048, 00:08:54.241 "data_size": 63488 00:08:54.241 }, 00:08:54.241 { 00:08:54.241 "name": "BaseBdev3", 00:08:54.241 "uuid": "f596b92c-cc21-5960-bae7-51eb568bbaf8", 00:08:54.241 "is_configured": true, 00:08:54.241 "data_offset": 2048, 00:08:54.241 "data_size": 63488 00:08:54.241 } 00:08:54.241 ] 00:08:54.241 }' 00:08:54.241 17:30:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:54.241 17:30:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:54.811 17:30:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:08:54.811 17:30:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:54.811 17:30:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:54.811 [2024-11-27 17:30:25.777868] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:08:54.811 [2024-11-27 17:30:25.777982] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:54.811 [2024-11-27 17:30:25.780534] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:54.811 [2024-11-27 17:30:25.780623] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:54.811 [2024-11-27 17:30:25.780679] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:08:54.811 [2024-11-27 17:30:25.780720] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name raid_bdev1, state offline 00:08:54.811 { 00:08:54.811 "results": [ 00:08:54.811 { 00:08:54.811 "job": "raid_bdev1", 00:08:54.811 "core_mask": "0x1", 00:08:54.811 "workload": "randrw", 00:08:54.811 "percentage": 50, 00:08:54.811 "status": "finished", 00:08:54.811 "queue_depth": 1, 00:08:54.811 "io_size": 131072, 00:08:54.811 "runtime": 1.381283, 00:08:54.811 "iops": 15215.56408064097, 00:08:54.811 "mibps": 1901.9455100801213, 00:08:54.811 "io_failed": 1, 00:08:54.811 "io_timeout": 0, 00:08:54.811 "avg_latency_us": 92.19945906212226, 00:08:54.811 "min_latency_us": 20.010480349344977, 00:08:54.811 "max_latency_us": 1387.989519650655 00:08:54.811 } 00:08:54.811 ], 00:08:54.811 "core_count": 1 00:08:54.811 } 00:08:54.811 17:30:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:54.811 17:30:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 76439 00:08:54.811 17:30:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@950 -- # '[' -z 76439 ']' 00:08:54.811 17:30:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # kill -0 76439 00:08:54.811 17:30:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@955 -- # uname 00:08:54.811 17:30:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:08:54.811 17:30:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 76439 00:08:54.811 killing process with pid 76439 00:08:54.811 17:30:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:08:54.811 17:30:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:08:54.811 17:30:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 76439' 00:08:54.811 17:30:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@969 -- # kill 76439 00:08:54.811 [2024-11-27 17:30:25.826316] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:08:54.811 17:30:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@974 -- # wait 76439 00:08:54.811 [2024-11-27 17:30:25.872647] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:08:55.072 17:30:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.KChe1ePlFi 00:08:55.072 17:30:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:08:55.072 17:30:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:08:55.072 17:30:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.72 00:08:55.072 17:30:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid0 00:08:55.072 ************************************ 00:08:55.072 END TEST raid_write_error_test 00:08:55.072 ************************************ 00:08:55.072 17:30:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:08:55.072 17:30:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:08:55.072 17:30:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.72 != \0\.\0\0 ]] 00:08:55.072 00:08:55.072 real 0m3.460s 00:08:55.072 user 0m4.219s 00:08:55.072 sys 0m0.633s 00:08:55.072 17:30:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:08:55.072 17:30:26 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:08:55.332 17:30:26 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:08:55.332 17:30:26 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test concat 3 false 00:08:55.332 17:30:26 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:08:55.332 17:30:26 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:08:55.332 17:30:26 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:08:55.332 ************************************ 00:08:55.332 START TEST raid_state_function_test 00:08:55.332 ************************************ 00:08:55.332 17:30:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1125 -- # raid_state_function_test concat 3 false 00:08:55.332 17:30:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=concat 00:08:55.332 17:30:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=3 00:08:55.332 17:30:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:08:55.332 17:30:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:08:55.332 17:30:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:08:55.332 17:30:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:55.332 17:30:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:08:55.332 17:30:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:55.332 17:30:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:55.332 17:30:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:08:55.332 17:30:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:55.332 17:30:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:55.332 17:30:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:08:55.332 17:30:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:08:55.332 17:30:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:08:55.332 17:30:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:08:55.332 17:30:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:08:55.332 17:30:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:08:55.332 17:30:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:08:55.332 17:30:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:08:55.332 17:30:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:08:55.332 17:30:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' concat '!=' raid1 ']' 00:08:55.332 17:30:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:08:55.332 17:30:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:08:55.332 17:30:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:08:55.332 17:30:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:08:55.333 17:30:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=76573 00:08:55.333 17:30:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:08:55.333 17:30:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 76573' 00:08:55.333 Process raid pid: 76573 00:08:55.333 17:30:26 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 76573 00:08:55.333 17:30:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@831 -- # '[' -z 76573 ']' 00:08:55.333 17:30:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:55.333 17:30:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:08:55.333 17:30:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:55.333 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:55.333 17:30:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:08:55.333 17:30:26 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:55.333 [2024-11-27 17:30:26.411008] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:08:55.333 [2024-11-27 17:30:26.411201] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:08:55.593 [2024-11-27 17:30:26.555197] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:55.593 [2024-11-27 17:30:26.623550] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:08:55.593 [2024-11-27 17:30:26.699309] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:55.593 [2024-11-27 17:30:26.699450] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:08:56.162 17:30:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:08:56.162 17:30:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@864 -- # return 0 00:08:56.162 17:30:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:56.162 17:30:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:56.162 17:30:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:56.162 [2024-11-27 17:30:27.234618] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:56.162 [2024-11-27 17:30:27.234723] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:56.162 [2024-11-27 17:30:27.234766] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:56.162 [2024-11-27 17:30:27.234791] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:56.162 [2024-11-27 17:30:27.234840] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:08:56.162 [2024-11-27 17:30:27.234866] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:08:56.162 17:30:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:56.162 17:30:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:56.162 17:30:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:56.162 17:30:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:56.162 17:30:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:56.162 17:30:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:56.162 17:30:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:56.162 17:30:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:56.162 17:30:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:56.162 17:30:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:56.162 17:30:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:56.162 17:30:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:56.162 17:30:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:56.162 17:30:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:56.162 17:30:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:56.162 17:30:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:56.162 17:30:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:56.162 "name": "Existed_Raid", 00:08:56.162 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:56.162 "strip_size_kb": 64, 00:08:56.162 "state": "configuring", 00:08:56.162 "raid_level": "concat", 00:08:56.162 "superblock": false, 00:08:56.162 "num_base_bdevs": 3, 00:08:56.162 "num_base_bdevs_discovered": 0, 00:08:56.162 "num_base_bdevs_operational": 3, 00:08:56.162 "base_bdevs_list": [ 00:08:56.162 { 00:08:56.162 "name": "BaseBdev1", 00:08:56.162 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:56.162 "is_configured": false, 00:08:56.162 "data_offset": 0, 00:08:56.162 "data_size": 0 00:08:56.162 }, 00:08:56.162 { 00:08:56.162 "name": "BaseBdev2", 00:08:56.162 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:56.162 "is_configured": false, 00:08:56.162 "data_offset": 0, 00:08:56.162 "data_size": 0 00:08:56.162 }, 00:08:56.162 { 00:08:56.162 "name": "BaseBdev3", 00:08:56.162 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:56.162 "is_configured": false, 00:08:56.162 "data_offset": 0, 00:08:56.162 "data_size": 0 00:08:56.162 } 00:08:56.162 ] 00:08:56.162 }' 00:08:56.162 17:30:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:56.162 17:30:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:56.733 17:30:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:56.733 17:30:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:56.733 17:30:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:56.733 [2024-11-27 17:30:27.665648] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:56.733 [2024-11-27 17:30:27.665734] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:08:56.733 17:30:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:56.733 17:30:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:56.733 17:30:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:56.733 17:30:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:56.733 [2024-11-27 17:30:27.677648] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:56.733 [2024-11-27 17:30:27.677740] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:56.733 [2024-11-27 17:30:27.677766] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:56.733 [2024-11-27 17:30:27.677789] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:56.733 [2024-11-27 17:30:27.677806] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:08:56.733 [2024-11-27 17:30:27.677827] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:08:56.733 17:30:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:56.733 17:30:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:08:56.733 17:30:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:56.733 17:30:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:56.733 [2024-11-27 17:30:27.704497] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:56.733 BaseBdev1 00:08:56.733 17:30:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:56.733 17:30:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:08:56.733 17:30:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:08:56.733 17:30:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:08:56.733 17:30:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:08:56.733 17:30:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:08:56.733 17:30:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:08:56.733 17:30:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:08:56.733 17:30:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:56.733 17:30:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:56.733 17:30:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:56.733 17:30:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:08:56.733 17:30:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:56.733 17:30:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:56.733 [ 00:08:56.733 { 00:08:56.733 "name": "BaseBdev1", 00:08:56.733 "aliases": [ 00:08:56.733 "dca85b57-f6f6-4fa4-a384-a5d6748d5f65" 00:08:56.733 ], 00:08:56.733 "product_name": "Malloc disk", 00:08:56.733 "block_size": 512, 00:08:56.733 "num_blocks": 65536, 00:08:56.733 "uuid": "dca85b57-f6f6-4fa4-a384-a5d6748d5f65", 00:08:56.733 "assigned_rate_limits": { 00:08:56.733 "rw_ios_per_sec": 0, 00:08:56.733 "rw_mbytes_per_sec": 0, 00:08:56.733 "r_mbytes_per_sec": 0, 00:08:56.733 "w_mbytes_per_sec": 0 00:08:56.733 }, 00:08:56.733 "claimed": true, 00:08:56.733 "claim_type": "exclusive_write", 00:08:56.733 "zoned": false, 00:08:56.733 "supported_io_types": { 00:08:56.733 "read": true, 00:08:56.733 "write": true, 00:08:56.733 "unmap": true, 00:08:56.733 "flush": true, 00:08:56.733 "reset": true, 00:08:56.733 "nvme_admin": false, 00:08:56.733 "nvme_io": false, 00:08:56.733 "nvme_io_md": false, 00:08:56.733 "write_zeroes": true, 00:08:56.733 "zcopy": true, 00:08:56.733 "get_zone_info": false, 00:08:56.733 "zone_management": false, 00:08:56.733 "zone_append": false, 00:08:56.733 "compare": false, 00:08:56.733 "compare_and_write": false, 00:08:56.733 "abort": true, 00:08:56.733 "seek_hole": false, 00:08:56.733 "seek_data": false, 00:08:56.733 "copy": true, 00:08:56.733 "nvme_iov_md": false 00:08:56.733 }, 00:08:56.733 "memory_domains": [ 00:08:56.733 { 00:08:56.733 "dma_device_id": "system", 00:08:56.733 "dma_device_type": 1 00:08:56.733 }, 00:08:56.733 { 00:08:56.733 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:56.733 "dma_device_type": 2 00:08:56.733 } 00:08:56.733 ], 00:08:56.733 "driver_specific": {} 00:08:56.733 } 00:08:56.733 ] 00:08:56.733 17:30:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:56.733 17:30:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:08:56.733 17:30:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:56.733 17:30:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:56.733 17:30:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:56.733 17:30:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:56.733 17:30:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:56.733 17:30:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:56.733 17:30:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:56.733 17:30:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:56.733 17:30:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:56.733 17:30:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:56.733 17:30:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:56.733 17:30:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:56.733 17:30:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:56.733 17:30:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:56.733 17:30:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:56.733 17:30:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:56.733 "name": "Existed_Raid", 00:08:56.733 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:56.733 "strip_size_kb": 64, 00:08:56.733 "state": "configuring", 00:08:56.733 "raid_level": "concat", 00:08:56.733 "superblock": false, 00:08:56.733 "num_base_bdevs": 3, 00:08:56.733 "num_base_bdevs_discovered": 1, 00:08:56.733 "num_base_bdevs_operational": 3, 00:08:56.733 "base_bdevs_list": [ 00:08:56.733 { 00:08:56.733 "name": "BaseBdev1", 00:08:56.733 "uuid": "dca85b57-f6f6-4fa4-a384-a5d6748d5f65", 00:08:56.733 "is_configured": true, 00:08:56.733 "data_offset": 0, 00:08:56.733 "data_size": 65536 00:08:56.733 }, 00:08:56.733 { 00:08:56.733 "name": "BaseBdev2", 00:08:56.733 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:56.733 "is_configured": false, 00:08:56.733 "data_offset": 0, 00:08:56.733 "data_size": 0 00:08:56.733 }, 00:08:56.733 { 00:08:56.733 "name": "BaseBdev3", 00:08:56.733 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:56.733 "is_configured": false, 00:08:56.733 "data_offset": 0, 00:08:56.733 "data_size": 0 00:08:56.733 } 00:08:56.733 ] 00:08:56.733 }' 00:08:56.733 17:30:27 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:56.733 17:30:27 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:56.993 17:30:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:08:56.993 17:30:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:56.993 17:30:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:56.993 [2024-11-27 17:30:28.151741] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:08:56.993 [2024-11-27 17:30:28.151833] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:08:56.993 17:30:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:56.993 17:30:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:56.993 17:30:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:56.993 17:30:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:56.993 [2024-11-27 17:30:28.163776] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:08:56.993 [2024-11-27 17:30:28.165916] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:08:56.993 [2024-11-27 17:30:28.165988] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:08:56.993 [2024-11-27 17:30:28.166031] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:08:56.993 [2024-11-27 17:30:28.166054] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:08:56.993 17:30:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:56.994 17:30:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:08:56.994 17:30:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:56.994 17:30:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:56.994 17:30:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:56.994 17:30:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:56.994 17:30:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:56.994 17:30:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:56.994 17:30:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:56.994 17:30:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:56.994 17:30:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:56.994 17:30:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:56.994 17:30:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:56.994 17:30:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:56.994 17:30:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:56.994 17:30:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:56.994 17:30:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:57.254 17:30:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:57.254 17:30:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:57.254 "name": "Existed_Raid", 00:08:57.254 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:57.254 "strip_size_kb": 64, 00:08:57.254 "state": "configuring", 00:08:57.254 "raid_level": "concat", 00:08:57.254 "superblock": false, 00:08:57.254 "num_base_bdevs": 3, 00:08:57.254 "num_base_bdevs_discovered": 1, 00:08:57.254 "num_base_bdevs_operational": 3, 00:08:57.254 "base_bdevs_list": [ 00:08:57.254 { 00:08:57.254 "name": "BaseBdev1", 00:08:57.254 "uuid": "dca85b57-f6f6-4fa4-a384-a5d6748d5f65", 00:08:57.254 "is_configured": true, 00:08:57.254 "data_offset": 0, 00:08:57.254 "data_size": 65536 00:08:57.254 }, 00:08:57.254 { 00:08:57.254 "name": "BaseBdev2", 00:08:57.254 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:57.254 "is_configured": false, 00:08:57.254 "data_offset": 0, 00:08:57.254 "data_size": 0 00:08:57.254 }, 00:08:57.254 { 00:08:57.254 "name": "BaseBdev3", 00:08:57.254 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:57.254 "is_configured": false, 00:08:57.254 "data_offset": 0, 00:08:57.254 "data_size": 0 00:08:57.254 } 00:08:57.254 ] 00:08:57.254 }' 00:08:57.254 17:30:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:57.254 17:30:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:57.514 17:30:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:08:57.514 17:30:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:57.514 17:30:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:57.514 [2024-11-27 17:30:28.620726] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:57.514 BaseBdev2 00:08:57.514 17:30:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:57.514 17:30:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:08:57.514 17:30:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:08:57.514 17:30:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:08:57.514 17:30:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:08:57.514 17:30:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:08:57.514 17:30:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:08:57.514 17:30:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:08:57.514 17:30:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:57.514 17:30:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:57.514 17:30:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:57.514 17:30:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:08:57.514 17:30:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:57.514 17:30:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:57.514 [ 00:08:57.514 { 00:08:57.514 "name": "BaseBdev2", 00:08:57.514 "aliases": [ 00:08:57.514 "02e10902-c36c-4afc-bcad-e9df550cc139" 00:08:57.514 ], 00:08:57.514 "product_name": "Malloc disk", 00:08:57.514 "block_size": 512, 00:08:57.514 "num_blocks": 65536, 00:08:57.514 "uuid": "02e10902-c36c-4afc-bcad-e9df550cc139", 00:08:57.514 "assigned_rate_limits": { 00:08:57.514 "rw_ios_per_sec": 0, 00:08:57.514 "rw_mbytes_per_sec": 0, 00:08:57.514 "r_mbytes_per_sec": 0, 00:08:57.514 "w_mbytes_per_sec": 0 00:08:57.514 }, 00:08:57.514 "claimed": true, 00:08:57.514 "claim_type": "exclusive_write", 00:08:57.514 "zoned": false, 00:08:57.514 "supported_io_types": { 00:08:57.514 "read": true, 00:08:57.514 "write": true, 00:08:57.514 "unmap": true, 00:08:57.514 "flush": true, 00:08:57.514 "reset": true, 00:08:57.514 "nvme_admin": false, 00:08:57.514 "nvme_io": false, 00:08:57.514 "nvme_io_md": false, 00:08:57.514 "write_zeroes": true, 00:08:57.514 "zcopy": true, 00:08:57.514 "get_zone_info": false, 00:08:57.514 "zone_management": false, 00:08:57.514 "zone_append": false, 00:08:57.514 "compare": false, 00:08:57.514 "compare_and_write": false, 00:08:57.514 "abort": true, 00:08:57.514 "seek_hole": false, 00:08:57.514 "seek_data": false, 00:08:57.514 "copy": true, 00:08:57.514 "nvme_iov_md": false 00:08:57.514 }, 00:08:57.514 "memory_domains": [ 00:08:57.514 { 00:08:57.514 "dma_device_id": "system", 00:08:57.514 "dma_device_type": 1 00:08:57.514 }, 00:08:57.514 { 00:08:57.514 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:57.514 "dma_device_type": 2 00:08:57.514 } 00:08:57.514 ], 00:08:57.514 "driver_specific": {} 00:08:57.514 } 00:08:57.514 ] 00:08:57.514 17:30:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:57.514 17:30:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:08:57.514 17:30:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:08:57.514 17:30:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:57.514 17:30:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:57.514 17:30:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:57.514 17:30:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:57.514 17:30:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:57.514 17:30:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:57.514 17:30:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:57.514 17:30:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:57.514 17:30:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:57.514 17:30:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:57.514 17:30:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:57.514 17:30:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:57.514 17:30:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:57.514 17:30:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:57.514 17:30:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:57.514 17:30:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:57.774 17:30:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:57.774 "name": "Existed_Raid", 00:08:57.774 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:57.774 "strip_size_kb": 64, 00:08:57.774 "state": "configuring", 00:08:57.774 "raid_level": "concat", 00:08:57.774 "superblock": false, 00:08:57.774 "num_base_bdevs": 3, 00:08:57.774 "num_base_bdevs_discovered": 2, 00:08:57.774 "num_base_bdevs_operational": 3, 00:08:57.774 "base_bdevs_list": [ 00:08:57.774 { 00:08:57.774 "name": "BaseBdev1", 00:08:57.774 "uuid": "dca85b57-f6f6-4fa4-a384-a5d6748d5f65", 00:08:57.774 "is_configured": true, 00:08:57.774 "data_offset": 0, 00:08:57.774 "data_size": 65536 00:08:57.774 }, 00:08:57.774 { 00:08:57.774 "name": "BaseBdev2", 00:08:57.774 "uuid": "02e10902-c36c-4afc-bcad-e9df550cc139", 00:08:57.774 "is_configured": true, 00:08:57.774 "data_offset": 0, 00:08:57.774 "data_size": 65536 00:08:57.774 }, 00:08:57.774 { 00:08:57.774 "name": "BaseBdev3", 00:08:57.774 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:57.774 "is_configured": false, 00:08:57.774 "data_offset": 0, 00:08:57.774 "data_size": 0 00:08:57.774 } 00:08:57.774 ] 00:08:57.774 }' 00:08:57.774 17:30:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:57.774 17:30:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:58.034 17:30:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:08:58.034 17:30:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:58.034 17:30:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:58.034 [2024-11-27 17:30:29.121123] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:58.034 [2024-11-27 17:30:29.121268] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:08:58.034 [2024-11-27 17:30:29.121300] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 196608, blocklen 512 00:08:58.034 [2024-11-27 17:30:29.121673] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:08:58.034 [2024-11-27 17:30:29.121869] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:08:58.034 [2024-11-27 17:30:29.121912] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:08:58.034 [2024-11-27 17:30:29.122195] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:08:58.034 BaseBdev3 00:08:58.034 17:30:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:58.034 17:30:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:08:58.034 17:30:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:08:58.034 17:30:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:08:58.034 17:30:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:08:58.034 17:30:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:08:58.034 17:30:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:08:58.034 17:30:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:08:58.034 17:30:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:58.034 17:30:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:58.034 17:30:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:58.034 17:30:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:08:58.034 17:30:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:58.034 17:30:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:58.034 [ 00:08:58.034 { 00:08:58.034 "name": "BaseBdev3", 00:08:58.034 "aliases": [ 00:08:58.034 "363d5d80-ca2f-4b1f-8518-998e3dfa078e" 00:08:58.034 ], 00:08:58.034 "product_name": "Malloc disk", 00:08:58.034 "block_size": 512, 00:08:58.034 "num_blocks": 65536, 00:08:58.034 "uuid": "363d5d80-ca2f-4b1f-8518-998e3dfa078e", 00:08:58.034 "assigned_rate_limits": { 00:08:58.034 "rw_ios_per_sec": 0, 00:08:58.034 "rw_mbytes_per_sec": 0, 00:08:58.034 "r_mbytes_per_sec": 0, 00:08:58.034 "w_mbytes_per_sec": 0 00:08:58.034 }, 00:08:58.034 "claimed": true, 00:08:58.034 "claim_type": "exclusive_write", 00:08:58.034 "zoned": false, 00:08:58.034 "supported_io_types": { 00:08:58.034 "read": true, 00:08:58.034 "write": true, 00:08:58.034 "unmap": true, 00:08:58.034 "flush": true, 00:08:58.034 "reset": true, 00:08:58.034 "nvme_admin": false, 00:08:58.034 "nvme_io": false, 00:08:58.034 "nvme_io_md": false, 00:08:58.034 "write_zeroes": true, 00:08:58.034 "zcopy": true, 00:08:58.034 "get_zone_info": false, 00:08:58.034 "zone_management": false, 00:08:58.034 "zone_append": false, 00:08:58.034 "compare": false, 00:08:58.034 "compare_and_write": false, 00:08:58.034 "abort": true, 00:08:58.034 "seek_hole": false, 00:08:58.034 "seek_data": false, 00:08:58.034 "copy": true, 00:08:58.034 "nvme_iov_md": false 00:08:58.034 }, 00:08:58.034 "memory_domains": [ 00:08:58.034 { 00:08:58.034 "dma_device_id": "system", 00:08:58.034 "dma_device_type": 1 00:08:58.034 }, 00:08:58.034 { 00:08:58.034 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:58.034 "dma_device_type": 2 00:08:58.034 } 00:08:58.034 ], 00:08:58.034 "driver_specific": {} 00:08:58.034 } 00:08:58.034 ] 00:08:58.034 17:30:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:58.034 17:30:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:08:58.034 17:30:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:08:58.034 17:30:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:08:58.034 17:30:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online concat 64 3 00:08:58.034 17:30:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:58.034 17:30:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:08:58.034 17:30:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:58.034 17:30:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:58.034 17:30:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:58.034 17:30:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:58.034 17:30:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:58.034 17:30:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:58.034 17:30:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:58.034 17:30:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:58.034 17:30:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:58.034 17:30:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:58.034 17:30:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:58.034 17:30:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:58.034 17:30:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:58.034 "name": "Existed_Raid", 00:08:58.035 "uuid": "4d94c322-cbab-4fa5-94d8-f513594138e3", 00:08:58.035 "strip_size_kb": 64, 00:08:58.035 "state": "online", 00:08:58.035 "raid_level": "concat", 00:08:58.035 "superblock": false, 00:08:58.035 "num_base_bdevs": 3, 00:08:58.035 "num_base_bdevs_discovered": 3, 00:08:58.035 "num_base_bdevs_operational": 3, 00:08:58.035 "base_bdevs_list": [ 00:08:58.035 { 00:08:58.035 "name": "BaseBdev1", 00:08:58.035 "uuid": "dca85b57-f6f6-4fa4-a384-a5d6748d5f65", 00:08:58.035 "is_configured": true, 00:08:58.035 "data_offset": 0, 00:08:58.035 "data_size": 65536 00:08:58.035 }, 00:08:58.035 { 00:08:58.035 "name": "BaseBdev2", 00:08:58.035 "uuid": "02e10902-c36c-4afc-bcad-e9df550cc139", 00:08:58.035 "is_configured": true, 00:08:58.035 "data_offset": 0, 00:08:58.035 "data_size": 65536 00:08:58.035 }, 00:08:58.035 { 00:08:58.035 "name": "BaseBdev3", 00:08:58.035 "uuid": "363d5d80-ca2f-4b1f-8518-998e3dfa078e", 00:08:58.035 "is_configured": true, 00:08:58.035 "data_offset": 0, 00:08:58.035 "data_size": 65536 00:08:58.035 } 00:08:58.035 ] 00:08:58.035 }' 00:08:58.035 17:30:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:58.035 17:30:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:58.603 17:30:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:08:58.603 17:30:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:08:58.603 17:30:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:08:58.603 17:30:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:08:58.603 17:30:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:08:58.603 17:30:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:08:58.603 17:30:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:08:58.603 17:30:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:08:58.603 17:30:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:58.603 17:30:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:58.603 [2024-11-27 17:30:29.616591] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:08:58.603 17:30:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:58.603 17:30:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:08:58.603 "name": "Existed_Raid", 00:08:58.603 "aliases": [ 00:08:58.603 "4d94c322-cbab-4fa5-94d8-f513594138e3" 00:08:58.603 ], 00:08:58.603 "product_name": "Raid Volume", 00:08:58.603 "block_size": 512, 00:08:58.603 "num_blocks": 196608, 00:08:58.603 "uuid": "4d94c322-cbab-4fa5-94d8-f513594138e3", 00:08:58.603 "assigned_rate_limits": { 00:08:58.603 "rw_ios_per_sec": 0, 00:08:58.603 "rw_mbytes_per_sec": 0, 00:08:58.603 "r_mbytes_per_sec": 0, 00:08:58.603 "w_mbytes_per_sec": 0 00:08:58.603 }, 00:08:58.603 "claimed": false, 00:08:58.603 "zoned": false, 00:08:58.603 "supported_io_types": { 00:08:58.603 "read": true, 00:08:58.603 "write": true, 00:08:58.603 "unmap": true, 00:08:58.603 "flush": true, 00:08:58.603 "reset": true, 00:08:58.603 "nvme_admin": false, 00:08:58.603 "nvme_io": false, 00:08:58.603 "nvme_io_md": false, 00:08:58.603 "write_zeroes": true, 00:08:58.603 "zcopy": false, 00:08:58.603 "get_zone_info": false, 00:08:58.603 "zone_management": false, 00:08:58.603 "zone_append": false, 00:08:58.603 "compare": false, 00:08:58.603 "compare_and_write": false, 00:08:58.603 "abort": false, 00:08:58.603 "seek_hole": false, 00:08:58.603 "seek_data": false, 00:08:58.603 "copy": false, 00:08:58.603 "nvme_iov_md": false 00:08:58.603 }, 00:08:58.603 "memory_domains": [ 00:08:58.603 { 00:08:58.603 "dma_device_id": "system", 00:08:58.603 "dma_device_type": 1 00:08:58.603 }, 00:08:58.603 { 00:08:58.603 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:58.603 "dma_device_type": 2 00:08:58.603 }, 00:08:58.603 { 00:08:58.603 "dma_device_id": "system", 00:08:58.603 "dma_device_type": 1 00:08:58.603 }, 00:08:58.603 { 00:08:58.603 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:58.603 "dma_device_type": 2 00:08:58.603 }, 00:08:58.603 { 00:08:58.603 "dma_device_id": "system", 00:08:58.603 "dma_device_type": 1 00:08:58.603 }, 00:08:58.603 { 00:08:58.603 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:58.603 "dma_device_type": 2 00:08:58.603 } 00:08:58.603 ], 00:08:58.603 "driver_specific": { 00:08:58.603 "raid": { 00:08:58.603 "uuid": "4d94c322-cbab-4fa5-94d8-f513594138e3", 00:08:58.603 "strip_size_kb": 64, 00:08:58.603 "state": "online", 00:08:58.603 "raid_level": "concat", 00:08:58.603 "superblock": false, 00:08:58.603 "num_base_bdevs": 3, 00:08:58.603 "num_base_bdevs_discovered": 3, 00:08:58.603 "num_base_bdevs_operational": 3, 00:08:58.603 "base_bdevs_list": [ 00:08:58.603 { 00:08:58.603 "name": "BaseBdev1", 00:08:58.603 "uuid": "dca85b57-f6f6-4fa4-a384-a5d6748d5f65", 00:08:58.603 "is_configured": true, 00:08:58.603 "data_offset": 0, 00:08:58.603 "data_size": 65536 00:08:58.603 }, 00:08:58.603 { 00:08:58.603 "name": "BaseBdev2", 00:08:58.603 "uuid": "02e10902-c36c-4afc-bcad-e9df550cc139", 00:08:58.603 "is_configured": true, 00:08:58.604 "data_offset": 0, 00:08:58.604 "data_size": 65536 00:08:58.604 }, 00:08:58.604 { 00:08:58.604 "name": "BaseBdev3", 00:08:58.604 "uuid": "363d5d80-ca2f-4b1f-8518-998e3dfa078e", 00:08:58.604 "is_configured": true, 00:08:58.604 "data_offset": 0, 00:08:58.604 "data_size": 65536 00:08:58.604 } 00:08:58.604 ] 00:08:58.604 } 00:08:58.604 } 00:08:58.604 }' 00:08:58.604 17:30:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:08:58.604 17:30:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:08:58.604 BaseBdev2 00:08:58.604 BaseBdev3' 00:08:58.604 17:30:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:58.604 17:30:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:08:58.604 17:30:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:58.604 17:30:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:58.604 17:30:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:08:58.604 17:30:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:58.604 17:30:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:58.604 17:30:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:58.864 17:30:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:58.864 17:30:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:58.864 17:30:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:58.864 17:30:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:58.864 17:30:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:08:58.864 17:30:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:58.864 17:30:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:58.864 17:30:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:58.864 17:30:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:58.864 17:30:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:58.864 17:30:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:08:58.864 17:30:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:08:58.864 17:30:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:08:58.864 17:30:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:58.864 17:30:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:58.864 17:30:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:58.864 17:30:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:08:58.864 17:30:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:08:58.864 17:30:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:08:58.864 17:30:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:58.864 17:30:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:58.864 [2024-11-27 17:30:29.911828] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:08:58.864 [2024-11-27 17:30:29.911903] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:08:58.864 [2024-11-27 17:30:29.911996] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:08:58.864 17:30:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:58.864 17:30:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:08:58.864 17:30:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy concat 00:08:58.864 17:30:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:08:58.864 17:30:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # return 1 00:08:58.864 17:30:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:08:58.864 17:30:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline concat 64 2 00:08:58.864 17:30:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:58.864 17:30:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:08:58.864 17:30:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:58.864 17:30:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:58.864 17:30:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:08:58.864 17:30:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:58.864 17:30:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:58.864 17:30:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:58.864 17:30:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:58.864 17:30:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:58.864 17:30:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:58.864 17:30:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:58.864 17:30:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:58.864 17:30:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:58.864 17:30:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:58.864 "name": "Existed_Raid", 00:08:58.864 "uuid": "4d94c322-cbab-4fa5-94d8-f513594138e3", 00:08:58.864 "strip_size_kb": 64, 00:08:58.864 "state": "offline", 00:08:58.864 "raid_level": "concat", 00:08:58.864 "superblock": false, 00:08:58.864 "num_base_bdevs": 3, 00:08:58.864 "num_base_bdevs_discovered": 2, 00:08:58.864 "num_base_bdevs_operational": 2, 00:08:58.864 "base_bdevs_list": [ 00:08:58.864 { 00:08:58.864 "name": null, 00:08:58.864 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:58.864 "is_configured": false, 00:08:58.864 "data_offset": 0, 00:08:58.864 "data_size": 65536 00:08:58.864 }, 00:08:58.864 { 00:08:58.864 "name": "BaseBdev2", 00:08:58.864 "uuid": "02e10902-c36c-4afc-bcad-e9df550cc139", 00:08:58.864 "is_configured": true, 00:08:58.864 "data_offset": 0, 00:08:58.864 "data_size": 65536 00:08:58.864 }, 00:08:58.864 { 00:08:58.864 "name": "BaseBdev3", 00:08:58.864 "uuid": "363d5d80-ca2f-4b1f-8518-998e3dfa078e", 00:08:58.864 "is_configured": true, 00:08:58.864 "data_offset": 0, 00:08:58.864 "data_size": 65536 00:08:58.864 } 00:08:58.864 ] 00:08:58.865 }' 00:08:58.865 17:30:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:58.865 17:30:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:59.434 17:30:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:08:59.434 17:30:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:59.434 17:30:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:08:59.435 17:30:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:59.435 17:30:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:59.435 17:30:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:59.435 17:30:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:59.435 17:30:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:08:59.435 17:30:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:08:59.435 17:30:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:08:59.435 17:30:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:59.435 17:30:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:59.435 [2024-11-27 17:30:30.459248] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:08:59.435 17:30:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:59.435 17:30:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:08:59.435 17:30:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:59.435 17:30:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:59.435 17:30:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:08:59.435 17:30:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:59.435 17:30:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:59.435 17:30:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:59.435 17:30:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:08:59.435 17:30:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:08:59.435 17:30:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:08:59.435 17:30:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:59.435 17:30:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:59.435 [2024-11-27 17:30:30.535636] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:08:59.435 [2024-11-27 17:30:30.535745] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:08:59.435 17:30:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:59.435 17:30:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:08:59.435 17:30:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:08:59.435 17:30:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:59.435 17:30:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:59.435 17:30:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:59.435 17:30:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:08:59.435 17:30:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:59.435 17:30:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:08:59.435 17:30:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:08:59.435 17:30:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 3 -gt 2 ']' 00:08:59.435 17:30:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:08:59.435 17:30:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:08:59.435 17:30:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:08:59.435 17:30:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:59.435 17:30:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:59.435 BaseBdev2 00:08:59.435 17:30:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:59.435 17:30:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:08:59.435 17:30:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:08:59.698 17:30:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:08:59.698 17:30:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:08:59.698 17:30:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:08:59.698 17:30:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:08:59.698 17:30:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:08:59.698 17:30:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:59.698 17:30:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:59.698 17:30:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:59.698 17:30:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:08:59.698 17:30:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:59.698 17:30:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:59.698 [ 00:08:59.698 { 00:08:59.698 "name": "BaseBdev2", 00:08:59.698 "aliases": [ 00:08:59.698 "082d5c93-dcdc-45c1-834f-144fd1d13a2a" 00:08:59.698 ], 00:08:59.698 "product_name": "Malloc disk", 00:08:59.698 "block_size": 512, 00:08:59.698 "num_blocks": 65536, 00:08:59.698 "uuid": "082d5c93-dcdc-45c1-834f-144fd1d13a2a", 00:08:59.698 "assigned_rate_limits": { 00:08:59.698 "rw_ios_per_sec": 0, 00:08:59.698 "rw_mbytes_per_sec": 0, 00:08:59.698 "r_mbytes_per_sec": 0, 00:08:59.698 "w_mbytes_per_sec": 0 00:08:59.698 }, 00:08:59.698 "claimed": false, 00:08:59.698 "zoned": false, 00:08:59.698 "supported_io_types": { 00:08:59.698 "read": true, 00:08:59.698 "write": true, 00:08:59.698 "unmap": true, 00:08:59.698 "flush": true, 00:08:59.698 "reset": true, 00:08:59.698 "nvme_admin": false, 00:08:59.698 "nvme_io": false, 00:08:59.698 "nvme_io_md": false, 00:08:59.698 "write_zeroes": true, 00:08:59.698 "zcopy": true, 00:08:59.698 "get_zone_info": false, 00:08:59.698 "zone_management": false, 00:08:59.698 "zone_append": false, 00:08:59.698 "compare": false, 00:08:59.698 "compare_and_write": false, 00:08:59.698 "abort": true, 00:08:59.698 "seek_hole": false, 00:08:59.698 "seek_data": false, 00:08:59.698 "copy": true, 00:08:59.698 "nvme_iov_md": false 00:08:59.698 }, 00:08:59.698 "memory_domains": [ 00:08:59.698 { 00:08:59.698 "dma_device_id": "system", 00:08:59.698 "dma_device_type": 1 00:08:59.698 }, 00:08:59.698 { 00:08:59.698 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:59.698 "dma_device_type": 2 00:08:59.698 } 00:08:59.698 ], 00:08:59.698 "driver_specific": {} 00:08:59.698 } 00:08:59.698 ] 00:08:59.698 17:30:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:59.698 17:30:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:08:59.698 17:30:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:08:59.698 17:30:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:08:59.698 17:30:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:08:59.698 17:30:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:59.698 17:30:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:59.698 BaseBdev3 00:08:59.698 17:30:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:59.698 17:30:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:08:59.698 17:30:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:08:59.698 17:30:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:08:59.698 17:30:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:08:59.698 17:30:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:08:59.698 17:30:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:08:59.698 17:30:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:08:59.698 17:30:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:59.698 17:30:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:59.698 17:30:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:59.698 17:30:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:08:59.698 17:30:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:59.698 17:30:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:59.698 [ 00:08:59.698 { 00:08:59.698 "name": "BaseBdev3", 00:08:59.698 "aliases": [ 00:08:59.699 "f5fb9610-c286-4dd4-b1a4-3a962c4e323d" 00:08:59.699 ], 00:08:59.699 "product_name": "Malloc disk", 00:08:59.699 "block_size": 512, 00:08:59.699 "num_blocks": 65536, 00:08:59.699 "uuid": "f5fb9610-c286-4dd4-b1a4-3a962c4e323d", 00:08:59.699 "assigned_rate_limits": { 00:08:59.699 "rw_ios_per_sec": 0, 00:08:59.699 "rw_mbytes_per_sec": 0, 00:08:59.699 "r_mbytes_per_sec": 0, 00:08:59.699 "w_mbytes_per_sec": 0 00:08:59.699 }, 00:08:59.699 "claimed": false, 00:08:59.699 "zoned": false, 00:08:59.699 "supported_io_types": { 00:08:59.699 "read": true, 00:08:59.699 "write": true, 00:08:59.699 "unmap": true, 00:08:59.699 "flush": true, 00:08:59.699 "reset": true, 00:08:59.699 "nvme_admin": false, 00:08:59.699 "nvme_io": false, 00:08:59.699 "nvme_io_md": false, 00:08:59.699 "write_zeroes": true, 00:08:59.699 "zcopy": true, 00:08:59.699 "get_zone_info": false, 00:08:59.699 "zone_management": false, 00:08:59.699 "zone_append": false, 00:08:59.699 "compare": false, 00:08:59.699 "compare_and_write": false, 00:08:59.699 "abort": true, 00:08:59.699 "seek_hole": false, 00:08:59.699 "seek_data": false, 00:08:59.699 "copy": true, 00:08:59.699 "nvme_iov_md": false 00:08:59.699 }, 00:08:59.699 "memory_domains": [ 00:08:59.699 { 00:08:59.699 "dma_device_id": "system", 00:08:59.699 "dma_device_type": 1 00:08:59.699 }, 00:08:59.699 { 00:08:59.699 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:59.699 "dma_device_type": 2 00:08:59.699 } 00:08:59.699 ], 00:08:59.699 "driver_specific": {} 00:08:59.699 } 00:08:59.699 ] 00:08:59.699 17:30:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:59.699 17:30:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:08:59.699 17:30:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:08:59.699 17:30:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:08:59.699 17:30:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:08:59.699 17:30:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:59.699 17:30:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:59.699 [2024-11-27 17:30:30.724859] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:08:59.699 [2024-11-27 17:30:30.724985] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:08:59.699 [2024-11-27 17:30:30.725028] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:08:59.699 [2024-11-27 17:30:30.727181] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:08:59.699 17:30:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:59.699 17:30:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:08:59.699 17:30:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:08:59.699 17:30:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:08:59.699 17:30:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:08:59.699 17:30:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:08:59.699 17:30:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:08:59.699 17:30:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:08:59.699 17:30:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:08:59.699 17:30:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:08:59.699 17:30:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:08:59.699 17:30:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:08:59.699 17:30:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:59.699 17:30:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:08:59.699 17:30:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:08:59.699 17:30:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:59.699 17:30:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:08:59.699 "name": "Existed_Raid", 00:08:59.699 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:59.699 "strip_size_kb": 64, 00:08:59.699 "state": "configuring", 00:08:59.699 "raid_level": "concat", 00:08:59.699 "superblock": false, 00:08:59.699 "num_base_bdevs": 3, 00:08:59.699 "num_base_bdevs_discovered": 2, 00:08:59.699 "num_base_bdevs_operational": 3, 00:08:59.699 "base_bdevs_list": [ 00:08:59.699 { 00:08:59.699 "name": "BaseBdev1", 00:08:59.699 "uuid": "00000000-0000-0000-0000-000000000000", 00:08:59.699 "is_configured": false, 00:08:59.699 "data_offset": 0, 00:08:59.699 "data_size": 0 00:08:59.699 }, 00:08:59.699 { 00:08:59.699 "name": "BaseBdev2", 00:08:59.699 "uuid": "082d5c93-dcdc-45c1-834f-144fd1d13a2a", 00:08:59.699 "is_configured": true, 00:08:59.699 "data_offset": 0, 00:08:59.699 "data_size": 65536 00:08:59.699 }, 00:08:59.699 { 00:08:59.699 "name": "BaseBdev3", 00:08:59.699 "uuid": "f5fb9610-c286-4dd4-b1a4-3a962c4e323d", 00:08:59.699 "is_configured": true, 00:08:59.699 "data_offset": 0, 00:08:59.699 "data_size": 65536 00:08:59.699 } 00:08:59.699 ] 00:08:59.699 }' 00:08:59.699 17:30:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:08:59.699 17:30:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:00.296 17:30:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:09:00.296 17:30:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:00.296 17:30:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:00.296 [2024-11-27 17:30:31.168105] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:09:00.296 17:30:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:00.296 17:30:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:09:00.296 17:30:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:00.296 17:30:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:00.296 17:30:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:00.296 17:30:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:00.296 17:30:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:00.296 17:30:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:00.296 17:30:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:00.296 17:30:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:00.296 17:30:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:00.296 17:30:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:00.296 17:30:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:00.296 17:30:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:00.296 17:30:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:00.296 17:30:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:00.296 17:30:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:00.296 "name": "Existed_Raid", 00:09:00.296 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:00.296 "strip_size_kb": 64, 00:09:00.296 "state": "configuring", 00:09:00.296 "raid_level": "concat", 00:09:00.296 "superblock": false, 00:09:00.296 "num_base_bdevs": 3, 00:09:00.296 "num_base_bdevs_discovered": 1, 00:09:00.296 "num_base_bdevs_operational": 3, 00:09:00.296 "base_bdevs_list": [ 00:09:00.296 { 00:09:00.296 "name": "BaseBdev1", 00:09:00.296 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:00.296 "is_configured": false, 00:09:00.296 "data_offset": 0, 00:09:00.296 "data_size": 0 00:09:00.296 }, 00:09:00.296 { 00:09:00.296 "name": null, 00:09:00.296 "uuid": "082d5c93-dcdc-45c1-834f-144fd1d13a2a", 00:09:00.296 "is_configured": false, 00:09:00.296 "data_offset": 0, 00:09:00.296 "data_size": 65536 00:09:00.296 }, 00:09:00.296 { 00:09:00.296 "name": "BaseBdev3", 00:09:00.296 "uuid": "f5fb9610-c286-4dd4-b1a4-3a962c4e323d", 00:09:00.296 "is_configured": true, 00:09:00.296 "data_offset": 0, 00:09:00.296 "data_size": 65536 00:09:00.296 } 00:09:00.296 ] 00:09:00.296 }' 00:09:00.296 17:30:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:00.296 17:30:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:00.557 17:30:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:09:00.557 17:30:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:00.557 17:30:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:00.557 17:30:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:00.557 17:30:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:00.557 17:30:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:09:00.557 17:30:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:09:00.557 17:30:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:00.557 17:30:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:00.557 [2024-11-27 17:30:31.639989] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:00.557 BaseBdev1 00:09:00.557 17:30:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:00.557 17:30:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:09:00.557 17:30:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:09:00.557 17:30:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:09:00.557 17:30:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:09:00.557 17:30:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:09:00.557 17:30:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:09:00.557 17:30:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:09:00.557 17:30:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:00.557 17:30:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:00.557 17:30:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:00.557 17:30:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:09:00.557 17:30:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:00.557 17:30:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:00.557 [ 00:09:00.557 { 00:09:00.557 "name": "BaseBdev1", 00:09:00.557 "aliases": [ 00:09:00.557 "e067392b-1451-42a2-b81e-481d0eded166" 00:09:00.557 ], 00:09:00.557 "product_name": "Malloc disk", 00:09:00.557 "block_size": 512, 00:09:00.557 "num_blocks": 65536, 00:09:00.557 "uuid": "e067392b-1451-42a2-b81e-481d0eded166", 00:09:00.557 "assigned_rate_limits": { 00:09:00.557 "rw_ios_per_sec": 0, 00:09:00.557 "rw_mbytes_per_sec": 0, 00:09:00.557 "r_mbytes_per_sec": 0, 00:09:00.557 "w_mbytes_per_sec": 0 00:09:00.557 }, 00:09:00.557 "claimed": true, 00:09:00.557 "claim_type": "exclusive_write", 00:09:00.557 "zoned": false, 00:09:00.557 "supported_io_types": { 00:09:00.557 "read": true, 00:09:00.557 "write": true, 00:09:00.557 "unmap": true, 00:09:00.557 "flush": true, 00:09:00.557 "reset": true, 00:09:00.557 "nvme_admin": false, 00:09:00.557 "nvme_io": false, 00:09:00.557 "nvme_io_md": false, 00:09:00.557 "write_zeroes": true, 00:09:00.557 "zcopy": true, 00:09:00.557 "get_zone_info": false, 00:09:00.557 "zone_management": false, 00:09:00.557 "zone_append": false, 00:09:00.557 "compare": false, 00:09:00.557 "compare_and_write": false, 00:09:00.557 "abort": true, 00:09:00.557 "seek_hole": false, 00:09:00.557 "seek_data": false, 00:09:00.557 "copy": true, 00:09:00.557 "nvme_iov_md": false 00:09:00.557 }, 00:09:00.557 "memory_domains": [ 00:09:00.557 { 00:09:00.557 "dma_device_id": "system", 00:09:00.557 "dma_device_type": 1 00:09:00.557 }, 00:09:00.557 { 00:09:00.557 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:00.557 "dma_device_type": 2 00:09:00.557 } 00:09:00.557 ], 00:09:00.557 "driver_specific": {} 00:09:00.557 } 00:09:00.557 ] 00:09:00.557 17:30:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:00.557 17:30:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:09:00.557 17:30:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:09:00.557 17:30:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:00.557 17:30:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:00.557 17:30:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:00.557 17:30:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:00.557 17:30:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:00.557 17:30:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:00.557 17:30:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:00.557 17:30:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:00.557 17:30:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:00.557 17:30:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:00.557 17:30:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:00.557 17:30:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:00.557 17:30:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:00.557 17:30:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:00.557 17:30:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:00.557 "name": "Existed_Raid", 00:09:00.557 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:00.557 "strip_size_kb": 64, 00:09:00.557 "state": "configuring", 00:09:00.557 "raid_level": "concat", 00:09:00.557 "superblock": false, 00:09:00.557 "num_base_bdevs": 3, 00:09:00.557 "num_base_bdevs_discovered": 2, 00:09:00.557 "num_base_bdevs_operational": 3, 00:09:00.557 "base_bdevs_list": [ 00:09:00.557 { 00:09:00.557 "name": "BaseBdev1", 00:09:00.557 "uuid": "e067392b-1451-42a2-b81e-481d0eded166", 00:09:00.557 "is_configured": true, 00:09:00.557 "data_offset": 0, 00:09:00.557 "data_size": 65536 00:09:00.557 }, 00:09:00.557 { 00:09:00.557 "name": null, 00:09:00.557 "uuid": "082d5c93-dcdc-45c1-834f-144fd1d13a2a", 00:09:00.557 "is_configured": false, 00:09:00.557 "data_offset": 0, 00:09:00.557 "data_size": 65536 00:09:00.557 }, 00:09:00.557 { 00:09:00.557 "name": "BaseBdev3", 00:09:00.557 "uuid": "f5fb9610-c286-4dd4-b1a4-3a962c4e323d", 00:09:00.557 "is_configured": true, 00:09:00.557 "data_offset": 0, 00:09:00.557 "data_size": 65536 00:09:00.557 } 00:09:00.557 ] 00:09:00.557 }' 00:09:00.557 17:30:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:00.557 17:30:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:01.127 17:30:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:01.127 17:30:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:01.127 17:30:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:01.127 17:30:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:09:01.127 17:30:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:01.127 17:30:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:09:01.127 17:30:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:09:01.127 17:30:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:01.127 17:30:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:01.127 [2024-11-27 17:30:32.155169] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:09:01.127 17:30:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:01.127 17:30:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:09:01.127 17:30:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:01.127 17:30:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:01.127 17:30:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:01.127 17:30:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:01.127 17:30:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:01.127 17:30:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:01.127 17:30:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:01.127 17:30:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:01.127 17:30:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:01.127 17:30:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:01.127 17:30:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:01.127 17:30:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:01.127 17:30:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:01.127 17:30:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:01.127 17:30:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:01.127 "name": "Existed_Raid", 00:09:01.127 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:01.127 "strip_size_kb": 64, 00:09:01.127 "state": "configuring", 00:09:01.127 "raid_level": "concat", 00:09:01.127 "superblock": false, 00:09:01.127 "num_base_bdevs": 3, 00:09:01.127 "num_base_bdevs_discovered": 1, 00:09:01.127 "num_base_bdevs_operational": 3, 00:09:01.127 "base_bdevs_list": [ 00:09:01.127 { 00:09:01.127 "name": "BaseBdev1", 00:09:01.127 "uuid": "e067392b-1451-42a2-b81e-481d0eded166", 00:09:01.127 "is_configured": true, 00:09:01.127 "data_offset": 0, 00:09:01.127 "data_size": 65536 00:09:01.127 }, 00:09:01.127 { 00:09:01.127 "name": null, 00:09:01.127 "uuid": "082d5c93-dcdc-45c1-834f-144fd1d13a2a", 00:09:01.127 "is_configured": false, 00:09:01.127 "data_offset": 0, 00:09:01.127 "data_size": 65536 00:09:01.127 }, 00:09:01.127 { 00:09:01.127 "name": null, 00:09:01.127 "uuid": "f5fb9610-c286-4dd4-b1a4-3a962c4e323d", 00:09:01.127 "is_configured": false, 00:09:01.127 "data_offset": 0, 00:09:01.127 "data_size": 65536 00:09:01.127 } 00:09:01.127 ] 00:09:01.127 }' 00:09:01.128 17:30:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:01.128 17:30:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:01.697 17:30:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:01.697 17:30:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:01.697 17:30:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:09:01.697 17:30:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:01.697 17:30:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:01.697 17:30:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:09:01.697 17:30:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:09:01.697 17:30:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:01.697 17:30:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:01.697 [2024-11-27 17:30:32.642385] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:01.697 17:30:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:01.697 17:30:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:09:01.697 17:30:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:01.697 17:30:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:01.697 17:30:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:01.697 17:30:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:01.697 17:30:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:01.697 17:30:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:01.697 17:30:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:01.697 17:30:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:01.697 17:30:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:01.697 17:30:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:01.697 17:30:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:01.697 17:30:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:01.697 17:30:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:01.697 17:30:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:01.697 17:30:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:01.697 "name": "Existed_Raid", 00:09:01.697 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:01.697 "strip_size_kb": 64, 00:09:01.697 "state": "configuring", 00:09:01.697 "raid_level": "concat", 00:09:01.697 "superblock": false, 00:09:01.697 "num_base_bdevs": 3, 00:09:01.697 "num_base_bdevs_discovered": 2, 00:09:01.697 "num_base_bdevs_operational": 3, 00:09:01.697 "base_bdevs_list": [ 00:09:01.697 { 00:09:01.697 "name": "BaseBdev1", 00:09:01.697 "uuid": "e067392b-1451-42a2-b81e-481d0eded166", 00:09:01.697 "is_configured": true, 00:09:01.697 "data_offset": 0, 00:09:01.697 "data_size": 65536 00:09:01.697 }, 00:09:01.697 { 00:09:01.697 "name": null, 00:09:01.697 "uuid": "082d5c93-dcdc-45c1-834f-144fd1d13a2a", 00:09:01.697 "is_configured": false, 00:09:01.697 "data_offset": 0, 00:09:01.697 "data_size": 65536 00:09:01.697 }, 00:09:01.697 { 00:09:01.697 "name": "BaseBdev3", 00:09:01.697 "uuid": "f5fb9610-c286-4dd4-b1a4-3a962c4e323d", 00:09:01.697 "is_configured": true, 00:09:01.697 "data_offset": 0, 00:09:01.697 "data_size": 65536 00:09:01.697 } 00:09:01.697 ] 00:09:01.697 }' 00:09:01.697 17:30:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:01.697 17:30:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:01.957 17:30:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:01.957 17:30:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:01.957 17:30:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:01.957 17:30:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:09:01.957 17:30:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:01.957 17:30:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:09:01.957 17:30:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:09:01.957 17:30:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:01.957 17:30:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:01.957 [2024-11-27 17:30:33.089651] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:09:01.957 17:30:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:01.957 17:30:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:09:01.957 17:30:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:01.957 17:30:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:01.957 17:30:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:01.957 17:30:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:01.957 17:30:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:01.957 17:30:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:01.957 17:30:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:01.957 17:30:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:01.957 17:30:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:01.957 17:30:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:01.957 17:30:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:01.957 17:30:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:01.957 17:30:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:01.957 17:30:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:02.216 17:30:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:02.216 "name": "Existed_Raid", 00:09:02.216 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:02.216 "strip_size_kb": 64, 00:09:02.216 "state": "configuring", 00:09:02.216 "raid_level": "concat", 00:09:02.216 "superblock": false, 00:09:02.216 "num_base_bdevs": 3, 00:09:02.216 "num_base_bdevs_discovered": 1, 00:09:02.216 "num_base_bdevs_operational": 3, 00:09:02.216 "base_bdevs_list": [ 00:09:02.216 { 00:09:02.216 "name": null, 00:09:02.216 "uuid": "e067392b-1451-42a2-b81e-481d0eded166", 00:09:02.216 "is_configured": false, 00:09:02.216 "data_offset": 0, 00:09:02.216 "data_size": 65536 00:09:02.216 }, 00:09:02.216 { 00:09:02.216 "name": null, 00:09:02.216 "uuid": "082d5c93-dcdc-45c1-834f-144fd1d13a2a", 00:09:02.216 "is_configured": false, 00:09:02.216 "data_offset": 0, 00:09:02.216 "data_size": 65536 00:09:02.216 }, 00:09:02.216 { 00:09:02.216 "name": "BaseBdev3", 00:09:02.216 "uuid": "f5fb9610-c286-4dd4-b1a4-3a962c4e323d", 00:09:02.216 "is_configured": true, 00:09:02.216 "data_offset": 0, 00:09:02.216 "data_size": 65536 00:09:02.216 } 00:09:02.216 ] 00:09:02.216 }' 00:09:02.216 17:30:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:02.216 17:30:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:02.475 17:30:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:02.475 17:30:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:09:02.475 17:30:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:02.475 17:30:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:02.475 17:30:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:02.475 17:30:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:09:02.475 17:30:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:09:02.475 17:30:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:02.475 17:30:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:02.476 [2024-11-27 17:30:33.636251] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:02.476 17:30:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:02.476 17:30:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:09:02.476 17:30:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:02.476 17:30:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:02.476 17:30:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:02.476 17:30:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:02.476 17:30:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:02.476 17:30:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:02.476 17:30:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:02.476 17:30:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:02.476 17:30:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:02.476 17:30:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:02.476 17:30:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:02.476 17:30:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:02.476 17:30:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:02.736 17:30:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:02.736 17:30:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:02.736 "name": "Existed_Raid", 00:09:02.736 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:02.736 "strip_size_kb": 64, 00:09:02.736 "state": "configuring", 00:09:02.736 "raid_level": "concat", 00:09:02.736 "superblock": false, 00:09:02.736 "num_base_bdevs": 3, 00:09:02.736 "num_base_bdevs_discovered": 2, 00:09:02.736 "num_base_bdevs_operational": 3, 00:09:02.736 "base_bdevs_list": [ 00:09:02.736 { 00:09:02.736 "name": null, 00:09:02.736 "uuid": "e067392b-1451-42a2-b81e-481d0eded166", 00:09:02.736 "is_configured": false, 00:09:02.736 "data_offset": 0, 00:09:02.736 "data_size": 65536 00:09:02.736 }, 00:09:02.736 { 00:09:02.736 "name": "BaseBdev2", 00:09:02.736 "uuid": "082d5c93-dcdc-45c1-834f-144fd1d13a2a", 00:09:02.736 "is_configured": true, 00:09:02.736 "data_offset": 0, 00:09:02.736 "data_size": 65536 00:09:02.736 }, 00:09:02.736 { 00:09:02.736 "name": "BaseBdev3", 00:09:02.736 "uuid": "f5fb9610-c286-4dd4-b1a4-3a962c4e323d", 00:09:02.736 "is_configured": true, 00:09:02.736 "data_offset": 0, 00:09:02.736 "data_size": 65536 00:09:02.736 } 00:09:02.736 ] 00:09:02.736 }' 00:09:02.736 17:30:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:02.736 17:30:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:02.996 17:30:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:02.996 17:30:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:09:02.996 17:30:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:02.996 17:30:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:02.996 17:30:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:02.996 17:30:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:09:02.996 17:30:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:02.996 17:30:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:02.996 17:30:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:02.996 17:30:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:09:02.996 17:30:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:02.996 17:30:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u e067392b-1451-42a2-b81e-481d0eded166 00:09:02.996 17:30:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:02.996 17:30:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:02.996 [2024-11-27 17:30:34.179982] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:09:02.996 [2024-11-27 17:30:34.180085] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:09:02.996 [2024-11-27 17:30:34.180114] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 196608, blocklen 512 00:09:02.996 [2024-11-27 17:30:34.180452] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002870 00:09:02.996 [2024-11-27 17:30:34.180633] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:09:02.996 [2024-11-27 17:30:34.180673] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001c80 00:09:02.996 [2024-11-27 17:30:34.180930] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:02.996 NewBaseBdev 00:09:02.996 17:30:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:02.996 17:30:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:09:02.996 17:30:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=NewBaseBdev 00:09:02.996 17:30:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:09:02.996 17:30:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:09:02.997 17:30:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:09:02.997 17:30:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:09:02.997 17:30:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:09:02.997 17:30:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:02.997 17:30:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:03.257 17:30:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:03.257 17:30:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:09:03.257 17:30:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:03.257 17:30:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:03.257 [ 00:09:03.257 { 00:09:03.257 "name": "NewBaseBdev", 00:09:03.257 "aliases": [ 00:09:03.258 "e067392b-1451-42a2-b81e-481d0eded166" 00:09:03.258 ], 00:09:03.258 "product_name": "Malloc disk", 00:09:03.258 "block_size": 512, 00:09:03.258 "num_blocks": 65536, 00:09:03.258 "uuid": "e067392b-1451-42a2-b81e-481d0eded166", 00:09:03.258 "assigned_rate_limits": { 00:09:03.258 "rw_ios_per_sec": 0, 00:09:03.258 "rw_mbytes_per_sec": 0, 00:09:03.258 "r_mbytes_per_sec": 0, 00:09:03.258 "w_mbytes_per_sec": 0 00:09:03.258 }, 00:09:03.258 "claimed": true, 00:09:03.258 "claim_type": "exclusive_write", 00:09:03.258 "zoned": false, 00:09:03.258 "supported_io_types": { 00:09:03.258 "read": true, 00:09:03.258 "write": true, 00:09:03.258 "unmap": true, 00:09:03.258 "flush": true, 00:09:03.258 "reset": true, 00:09:03.258 "nvme_admin": false, 00:09:03.258 "nvme_io": false, 00:09:03.258 "nvme_io_md": false, 00:09:03.258 "write_zeroes": true, 00:09:03.258 "zcopy": true, 00:09:03.258 "get_zone_info": false, 00:09:03.258 "zone_management": false, 00:09:03.258 "zone_append": false, 00:09:03.258 "compare": false, 00:09:03.258 "compare_and_write": false, 00:09:03.258 "abort": true, 00:09:03.258 "seek_hole": false, 00:09:03.258 "seek_data": false, 00:09:03.258 "copy": true, 00:09:03.258 "nvme_iov_md": false 00:09:03.258 }, 00:09:03.258 "memory_domains": [ 00:09:03.258 { 00:09:03.258 "dma_device_id": "system", 00:09:03.258 "dma_device_type": 1 00:09:03.258 }, 00:09:03.258 { 00:09:03.258 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:03.258 "dma_device_type": 2 00:09:03.258 } 00:09:03.258 ], 00:09:03.258 "driver_specific": {} 00:09:03.258 } 00:09:03.258 ] 00:09:03.258 17:30:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:03.258 17:30:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:09:03.258 17:30:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online concat 64 3 00:09:03.258 17:30:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:03.258 17:30:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:03.258 17:30:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:03.258 17:30:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:03.258 17:30:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:03.258 17:30:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:03.258 17:30:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:03.258 17:30:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:03.258 17:30:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:03.258 17:30:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:03.258 17:30:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:03.258 17:30:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:03.258 17:30:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:03.258 17:30:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:03.258 17:30:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:03.258 "name": "Existed_Raid", 00:09:03.258 "uuid": "0446256a-37ba-4e2c-b599-5d78dd70a9e1", 00:09:03.258 "strip_size_kb": 64, 00:09:03.258 "state": "online", 00:09:03.258 "raid_level": "concat", 00:09:03.258 "superblock": false, 00:09:03.258 "num_base_bdevs": 3, 00:09:03.258 "num_base_bdevs_discovered": 3, 00:09:03.258 "num_base_bdevs_operational": 3, 00:09:03.258 "base_bdevs_list": [ 00:09:03.258 { 00:09:03.258 "name": "NewBaseBdev", 00:09:03.258 "uuid": "e067392b-1451-42a2-b81e-481d0eded166", 00:09:03.258 "is_configured": true, 00:09:03.258 "data_offset": 0, 00:09:03.258 "data_size": 65536 00:09:03.258 }, 00:09:03.258 { 00:09:03.258 "name": "BaseBdev2", 00:09:03.258 "uuid": "082d5c93-dcdc-45c1-834f-144fd1d13a2a", 00:09:03.258 "is_configured": true, 00:09:03.258 "data_offset": 0, 00:09:03.258 "data_size": 65536 00:09:03.258 }, 00:09:03.258 { 00:09:03.258 "name": "BaseBdev3", 00:09:03.258 "uuid": "f5fb9610-c286-4dd4-b1a4-3a962c4e323d", 00:09:03.258 "is_configured": true, 00:09:03.258 "data_offset": 0, 00:09:03.258 "data_size": 65536 00:09:03.258 } 00:09:03.258 ] 00:09:03.258 }' 00:09:03.258 17:30:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:03.258 17:30:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:03.518 17:30:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:09:03.518 17:30:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:09:03.518 17:30:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:03.518 17:30:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:03.779 17:30:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:09:03.779 17:30:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:03.779 17:30:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:09:03.779 17:30:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:03.779 17:30:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:03.779 17:30:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:03.779 [2024-11-27 17:30:34.719393] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:03.779 17:30:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:03.779 17:30:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:03.779 "name": "Existed_Raid", 00:09:03.779 "aliases": [ 00:09:03.779 "0446256a-37ba-4e2c-b599-5d78dd70a9e1" 00:09:03.779 ], 00:09:03.779 "product_name": "Raid Volume", 00:09:03.779 "block_size": 512, 00:09:03.779 "num_blocks": 196608, 00:09:03.779 "uuid": "0446256a-37ba-4e2c-b599-5d78dd70a9e1", 00:09:03.779 "assigned_rate_limits": { 00:09:03.779 "rw_ios_per_sec": 0, 00:09:03.779 "rw_mbytes_per_sec": 0, 00:09:03.779 "r_mbytes_per_sec": 0, 00:09:03.779 "w_mbytes_per_sec": 0 00:09:03.779 }, 00:09:03.779 "claimed": false, 00:09:03.779 "zoned": false, 00:09:03.779 "supported_io_types": { 00:09:03.779 "read": true, 00:09:03.779 "write": true, 00:09:03.779 "unmap": true, 00:09:03.779 "flush": true, 00:09:03.779 "reset": true, 00:09:03.779 "nvme_admin": false, 00:09:03.779 "nvme_io": false, 00:09:03.779 "nvme_io_md": false, 00:09:03.779 "write_zeroes": true, 00:09:03.779 "zcopy": false, 00:09:03.779 "get_zone_info": false, 00:09:03.779 "zone_management": false, 00:09:03.779 "zone_append": false, 00:09:03.779 "compare": false, 00:09:03.779 "compare_and_write": false, 00:09:03.779 "abort": false, 00:09:03.779 "seek_hole": false, 00:09:03.779 "seek_data": false, 00:09:03.779 "copy": false, 00:09:03.779 "nvme_iov_md": false 00:09:03.779 }, 00:09:03.779 "memory_domains": [ 00:09:03.779 { 00:09:03.779 "dma_device_id": "system", 00:09:03.779 "dma_device_type": 1 00:09:03.779 }, 00:09:03.779 { 00:09:03.779 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:03.779 "dma_device_type": 2 00:09:03.779 }, 00:09:03.779 { 00:09:03.779 "dma_device_id": "system", 00:09:03.779 "dma_device_type": 1 00:09:03.779 }, 00:09:03.779 { 00:09:03.779 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:03.779 "dma_device_type": 2 00:09:03.779 }, 00:09:03.779 { 00:09:03.779 "dma_device_id": "system", 00:09:03.779 "dma_device_type": 1 00:09:03.779 }, 00:09:03.779 { 00:09:03.779 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:03.779 "dma_device_type": 2 00:09:03.779 } 00:09:03.779 ], 00:09:03.779 "driver_specific": { 00:09:03.779 "raid": { 00:09:03.779 "uuid": "0446256a-37ba-4e2c-b599-5d78dd70a9e1", 00:09:03.779 "strip_size_kb": 64, 00:09:03.779 "state": "online", 00:09:03.779 "raid_level": "concat", 00:09:03.779 "superblock": false, 00:09:03.779 "num_base_bdevs": 3, 00:09:03.779 "num_base_bdevs_discovered": 3, 00:09:03.779 "num_base_bdevs_operational": 3, 00:09:03.779 "base_bdevs_list": [ 00:09:03.779 { 00:09:03.779 "name": "NewBaseBdev", 00:09:03.779 "uuid": "e067392b-1451-42a2-b81e-481d0eded166", 00:09:03.779 "is_configured": true, 00:09:03.779 "data_offset": 0, 00:09:03.779 "data_size": 65536 00:09:03.779 }, 00:09:03.779 { 00:09:03.779 "name": "BaseBdev2", 00:09:03.779 "uuid": "082d5c93-dcdc-45c1-834f-144fd1d13a2a", 00:09:03.779 "is_configured": true, 00:09:03.779 "data_offset": 0, 00:09:03.779 "data_size": 65536 00:09:03.779 }, 00:09:03.779 { 00:09:03.779 "name": "BaseBdev3", 00:09:03.779 "uuid": "f5fb9610-c286-4dd4-b1a4-3a962c4e323d", 00:09:03.779 "is_configured": true, 00:09:03.779 "data_offset": 0, 00:09:03.779 "data_size": 65536 00:09:03.779 } 00:09:03.779 ] 00:09:03.779 } 00:09:03.779 } 00:09:03.779 }' 00:09:03.779 17:30:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:03.779 17:30:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:09:03.779 BaseBdev2 00:09:03.779 BaseBdev3' 00:09:03.779 17:30:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:03.779 17:30:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:03.779 17:30:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:03.779 17:30:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:09:03.779 17:30:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:03.779 17:30:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:03.779 17:30:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:03.779 17:30:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:03.779 17:30:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:03.779 17:30:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:03.779 17:30:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:03.779 17:30:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:09:03.779 17:30:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:03.779 17:30:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:03.779 17:30:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:03.779 17:30:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:03.779 17:30:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:03.779 17:30:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:03.779 17:30:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:03.779 17:30:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:09:03.779 17:30:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:03.779 17:30:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:03.779 17:30:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:03.780 17:30:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:04.040 17:30:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:04.040 17:30:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:04.040 17:30:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:04.040 17:30:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:04.040 17:30:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:04.040 [2024-11-27 17:30:34.986701] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:04.040 [2024-11-27 17:30:34.986793] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:04.040 [2024-11-27 17:30:34.986896] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:04.040 [2024-11-27 17:30:34.986987] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:04.040 [2024-11-27 17:30:34.987116] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name Existed_Raid, state offline 00:09:04.040 17:30:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:04.040 17:30:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 76573 00:09:04.040 17:30:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@950 -- # '[' -z 76573 ']' 00:09:04.040 17:30:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # kill -0 76573 00:09:04.040 17:30:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@955 -- # uname 00:09:04.040 17:30:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:09:04.040 17:30:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 76573 00:09:04.040 17:30:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:09:04.040 17:30:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:09:04.040 17:30:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 76573' 00:09:04.040 killing process with pid 76573 00:09:04.040 17:30:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@969 -- # kill 76573 00:09:04.040 [2024-11-27 17:30:35.039427] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:09:04.040 17:30:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@974 -- # wait 76573 00:09:04.040 [2024-11-27 17:30:35.096303] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:09:04.300 17:30:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:09:04.300 00:09:04.300 real 0m9.146s 00:09:04.300 user 0m15.414s 00:09:04.300 sys 0m1.847s 00:09:04.300 ************************************ 00:09:04.300 END TEST raid_state_function_test 00:09:04.300 ************************************ 00:09:04.300 17:30:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:09:04.300 17:30:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:04.561 17:30:35 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test concat 3 true 00:09:04.561 17:30:35 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:09:04.561 17:30:35 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:09:04.561 17:30:35 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:09:04.561 ************************************ 00:09:04.561 START TEST raid_state_function_test_sb 00:09:04.561 ************************************ 00:09:04.561 17:30:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1125 -- # raid_state_function_test concat 3 true 00:09:04.561 17:30:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=concat 00:09:04.561 17:30:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=3 00:09:04.561 17:30:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:09:04.561 17:30:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:09:04.561 17:30:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:09:04.561 17:30:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:04.561 17:30:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:09:04.561 17:30:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:04.561 17:30:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:04.561 17:30:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:09:04.561 17:30:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:04.561 17:30:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:04.561 17:30:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:09:04.561 17:30:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:04.561 17:30:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:04.561 17:30:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:09:04.561 17:30:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:09:04.561 17:30:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:09:04.561 17:30:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:09:04.561 17:30:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:09:04.561 17:30:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:09:04.561 17:30:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' concat '!=' raid1 ']' 00:09:04.561 17:30:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:09:04.561 17:30:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:09:04.561 17:30:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:09:04.561 17:30:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:09:04.561 17:30:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=77178 00:09:04.561 17:30:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:09:04.561 17:30:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 77178' 00:09:04.561 Process raid pid: 77178 00:09:04.561 17:30:35 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 77178 00:09:04.561 17:30:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@831 -- # '[' -z 77178 ']' 00:09:04.561 17:30:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:04.561 17:30:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@836 -- # local max_retries=100 00:09:04.561 17:30:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:04.561 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:04.561 17:30:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # xtrace_disable 00:09:04.561 17:30:35 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:04.561 [2024-11-27 17:30:35.642522] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:09:04.561 [2024-11-27 17:30:35.642761] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:09:04.821 [2024-11-27 17:30:35.791639] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:04.821 [2024-11-27 17:30:35.859098] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:09:04.821 [2024-11-27 17:30:35.934664] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:04.821 [2024-11-27 17:30:35.934751] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:05.392 17:30:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:09:05.392 17:30:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@864 -- # return 0 00:09:05.392 17:30:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:09:05.392 17:30:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:05.392 17:30:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:05.392 [2024-11-27 17:30:36.465576] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:05.392 [2024-11-27 17:30:36.465637] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:05.392 [2024-11-27 17:30:36.465650] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:05.392 [2024-11-27 17:30:36.465662] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:05.392 [2024-11-27 17:30:36.465668] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:05.392 [2024-11-27 17:30:36.465680] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:05.392 17:30:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:05.392 17:30:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:09:05.392 17:30:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:05.392 17:30:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:05.392 17:30:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:05.392 17:30:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:05.392 17:30:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:05.392 17:30:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:05.392 17:30:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:05.392 17:30:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:05.392 17:30:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:05.392 17:30:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:05.392 17:30:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:05.392 17:30:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:05.392 17:30:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:05.392 17:30:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:05.392 17:30:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:05.392 "name": "Existed_Raid", 00:09:05.392 "uuid": "91ff2d19-995a-4c36-ac84-3ec4d1965a8c", 00:09:05.392 "strip_size_kb": 64, 00:09:05.392 "state": "configuring", 00:09:05.392 "raid_level": "concat", 00:09:05.392 "superblock": true, 00:09:05.392 "num_base_bdevs": 3, 00:09:05.392 "num_base_bdevs_discovered": 0, 00:09:05.392 "num_base_bdevs_operational": 3, 00:09:05.392 "base_bdevs_list": [ 00:09:05.392 { 00:09:05.392 "name": "BaseBdev1", 00:09:05.392 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:05.392 "is_configured": false, 00:09:05.392 "data_offset": 0, 00:09:05.392 "data_size": 0 00:09:05.392 }, 00:09:05.392 { 00:09:05.392 "name": "BaseBdev2", 00:09:05.392 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:05.392 "is_configured": false, 00:09:05.392 "data_offset": 0, 00:09:05.392 "data_size": 0 00:09:05.392 }, 00:09:05.392 { 00:09:05.392 "name": "BaseBdev3", 00:09:05.392 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:05.392 "is_configured": false, 00:09:05.392 "data_offset": 0, 00:09:05.392 "data_size": 0 00:09:05.392 } 00:09:05.392 ] 00:09:05.392 }' 00:09:05.392 17:30:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:05.392 17:30:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:05.961 17:30:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:05.961 17:30:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:05.961 17:30:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:05.961 [2024-11-27 17:30:36.876700] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:05.961 [2024-11-27 17:30:36.876794] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:09:05.961 17:30:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:05.961 17:30:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:09:05.961 17:30:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:05.961 17:30:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:05.961 [2024-11-27 17:30:36.888713] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:05.961 [2024-11-27 17:30:36.888806] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:05.961 [2024-11-27 17:30:36.888833] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:05.961 [2024-11-27 17:30:36.888862] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:05.961 [2024-11-27 17:30:36.888879] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:05.961 [2024-11-27 17:30:36.888900] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:05.961 17:30:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:05.961 17:30:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:09:05.961 17:30:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:05.961 17:30:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:05.961 [2024-11-27 17:30:36.915573] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:05.961 BaseBdev1 00:09:05.961 17:30:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:05.961 17:30:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:09:05.961 17:30:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:09:05.961 17:30:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:09:05.961 17:30:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:09:05.961 17:30:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:09:05.961 17:30:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:09:05.961 17:30:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:09:05.962 17:30:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:05.962 17:30:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:05.962 17:30:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:05.962 17:30:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:09:05.962 17:30:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:05.962 17:30:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:05.962 [ 00:09:05.962 { 00:09:05.962 "name": "BaseBdev1", 00:09:05.962 "aliases": [ 00:09:05.962 "dc037112-fc5e-46de-b45a-8a4ccc2db305" 00:09:05.962 ], 00:09:05.962 "product_name": "Malloc disk", 00:09:05.962 "block_size": 512, 00:09:05.962 "num_blocks": 65536, 00:09:05.962 "uuid": "dc037112-fc5e-46de-b45a-8a4ccc2db305", 00:09:05.962 "assigned_rate_limits": { 00:09:05.962 "rw_ios_per_sec": 0, 00:09:05.962 "rw_mbytes_per_sec": 0, 00:09:05.962 "r_mbytes_per_sec": 0, 00:09:05.962 "w_mbytes_per_sec": 0 00:09:05.962 }, 00:09:05.962 "claimed": true, 00:09:05.962 "claim_type": "exclusive_write", 00:09:05.962 "zoned": false, 00:09:05.962 "supported_io_types": { 00:09:05.962 "read": true, 00:09:05.962 "write": true, 00:09:05.962 "unmap": true, 00:09:05.962 "flush": true, 00:09:05.962 "reset": true, 00:09:05.962 "nvme_admin": false, 00:09:05.962 "nvme_io": false, 00:09:05.962 "nvme_io_md": false, 00:09:05.962 "write_zeroes": true, 00:09:05.962 "zcopy": true, 00:09:05.962 "get_zone_info": false, 00:09:05.962 "zone_management": false, 00:09:05.962 "zone_append": false, 00:09:05.962 "compare": false, 00:09:05.962 "compare_and_write": false, 00:09:05.962 "abort": true, 00:09:05.962 "seek_hole": false, 00:09:05.962 "seek_data": false, 00:09:05.962 "copy": true, 00:09:05.962 "nvme_iov_md": false 00:09:05.962 }, 00:09:05.962 "memory_domains": [ 00:09:05.962 { 00:09:05.962 "dma_device_id": "system", 00:09:05.962 "dma_device_type": 1 00:09:05.962 }, 00:09:05.962 { 00:09:05.962 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:05.962 "dma_device_type": 2 00:09:05.962 } 00:09:05.962 ], 00:09:05.962 "driver_specific": {} 00:09:05.962 } 00:09:05.962 ] 00:09:05.962 17:30:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:05.962 17:30:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:09:05.962 17:30:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:09:05.962 17:30:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:05.962 17:30:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:05.962 17:30:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:05.962 17:30:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:05.962 17:30:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:05.962 17:30:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:05.962 17:30:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:05.962 17:30:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:05.962 17:30:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:05.962 17:30:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:05.962 17:30:36 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:05.962 17:30:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:05.962 17:30:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:05.962 17:30:36 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:05.962 17:30:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:05.962 "name": "Existed_Raid", 00:09:05.962 "uuid": "385855a6-97f7-4ebc-9c13-dfab8431dc7b", 00:09:05.962 "strip_size_kb": 64, 00:09:05.962 "state": "configuring", 00:09:05.962 "raid_level": "concat", 00:09:05.962 "superblock": true, 00:09:05.962 "num_base_bdevs": 3, 00:09:05.962 "num_base_bdevs_discovered": 1, 00:09:05.962 "num_base_bdevs_operational": 3, 00:09:05.962 "base_bdevs_list": [ 00:09:05.962 { 00:09:05.962 "name": "BaseBdev1", 00:09:05.962 "uuid": "dc037112-fc5e-46de-b45a-8a4ccc2db305", 00:09:05.962 "is_configured": true, 00:09:05.962 "data_offset": 2048, 00:09:05.962 "data_size": 63488 00:09:05.962 }, 00:09:05.962 { 00:09:05.962 "name": "BaseBdev2", 00:09:05.962 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:05.962 "is_configured": false, 00:09:05.962 "data_offset": 0, 00:09:05.962 "data_size": 0 00:09:05.962 }, 00:09:05.962 { 00:09:05.962 "name": "BaseBdev3", 00:09:05.962 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:05.962 "is_configured": false, 00:09:05.962 "data_offset": 0, 00:09:05.962 "data_size": 0 00:09:05.962 } 00:09:05.962 ] 00:09:05.962 }' 00:09:05.962 17:30:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:05.962 17:30:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:06.532 17:30:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:06.532 17:30:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:06.532 17:30:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:06.532 [2024-11-27 17:30:37.438688] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:06.532 [2024-11-27 17:30:37.438816] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:09:06.532 17:30:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:06.532 17:30:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:09:06.532 17:30:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:06.532 17:30:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:06.532 [2024-11-27 17:30:37.450744] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:06.532 [2024-11-27 17:30:37.452867] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:06.532 [2024-11-27 17:30:37.452910] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:06.532 [2024-11-27 17:30:37.452920] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:06.532 [2024-11-27 17:30:37.452930] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:06.532 17:30:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:06.532 17:30:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:09:06.532 17:30:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:06.532 17:30:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:09:06.532 17:30:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:06.532 17:30:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:06.532 17:30:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:06.532 17:30:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:06.532 17:30:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:06.532 17:30:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:06.532 17:30:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:06.532 17:30:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:06.532 17:30:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:06.532 17:30:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:06.532 17:30:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:06.532 17:30:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:06.532 17:30:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:06.532 17:30:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:06.532 17:30:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:06.532 "name": "Existed_Raid", 00:09:06.532 "uuid": "7dc84895-f60b-4cfb-9bab-dcf15dfd7be0", 00:09:06.532 "strip_size_kb": 64, 00:09:06.532 "state": "configuring", 00:09:06.532 "raid_level": "concat", 00:09:06.532 "superblock": true, 00:09:06.532 "num_base_bdevs": 3, 00:09:06.532 "num_base_bdevs_discovered": 1, 00:09:06.532 "num_base_bdevs_operational": 3, 00:09:06.532 "base_bdevs_list": [ 00:09:06.532 { 00:09:06.532 "name": "BaseBdev1", 00:09:06.532 "uuid": "dc037112-fc5e-46de-b45a-8a4ccc2db305", 00:09:06.532 "is_configured": true, 00:09:06.532 "data_offset": 2048, 00:09:06.532 "data_size": 63488 00:09:06.532 }, 00:09:06.532 { 00:09:06.532 "name": "BaseBdev2", 00:09:06.532 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:06.532 "is_configured": false, 00:09:06.532 "data_offset": 0, 00:09:06.532 "data_size": 0 00:09:06.532 }, 00:09:06.532 { 00:09:06.532 "name": "BaseBdev3", 00:09:06.532 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:06.532 "is_configured": false, 00:09:06.532 "data_offset": 0, 00:09:06.532 "data_size": 0 00:09:06.532 } 00:09:06.532 ] 00:09:06.532 }' 00:09:06.533 17:30:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:06.533 17:30:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:06.793 17:30:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:09:06.793 17:30:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:06.793 17:30:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:06.793 [2024-11-27 17:30:37.855475] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:06.793 BaseBdev2 00:09:06.793 17:30:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:06.793 17:30:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:09:06.793 17:30:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:09:06.793 17:30:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:09:06.793 17:30:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:09:06.793 17:30:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:09:06.793 17:30:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:09:06.793 17:30:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:09:06.793 17:30:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:06.793 17:30:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:06.793 17:30:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:06.793 17:30:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:09:06.793 17:30:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:06.793 17:30:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:06.793 [ 00:09:06.793 { 00:09:06.793 "name": "BaseBdev2", 00:09:06.793 "aliases": [ 00:09:06.793 "92243fb0-b81e-4e75-8858-7074e93e5e9d" 00:09:06.793 ], 00:09:06.793 "product_name": "Malloc disk", 00:09:06.793 "block_size": 512, 00:09:06.793 "num_blocks": 65536, 00:09:06.793 "uuid": "92243fb0-b81e-4e75-8858-7074e93e5e9d", 00:09:06.793 "assigned_rate_limits": { 00:09:06.793 "rw_ios_per_sec": 0, 00:09:06.793 "rw_mbytes_per_sec": 0, 00:09:06.793 "r_mbytes_per_sec": 0, 00:09:06.793 "w_mbytes_per_sec": 0 00:09:06.793 }, 00:09:06.793 "claimed": true, 00:09:06.793 "claim_type": "exclusive_write", 00:09:06.793 "zoned": false, 00:09:06.793 "supported_io_types": { 00:09:06.793 "read": true, 00:09:06.793 "write": true, 00:09:06.793 "unmap": true, 00:09:06.793 "flush": true, 00:09:06.793 "reset": true, 00:09:06.793 "nvme_admin": false, 00:09:06.793 "nvme_io": false, 00:09:06.793 "nvme_io_md": false, 00:09:06.793 "write_zeroes": true, 00:09:06.793 "zcopy": true, 00:09:06.793 "get_zone_info": false, 00:09:06.793 "zone_management": false, 00:09:06.793 "zone_append": false, 00:09:06.793 "compare": false, 00:09:06.793 "compare_and_write": false, 00:09:06.793 "abort": true, 00:09:06.793 "seek_hole": false, 00:09:06.793 "seek_data": false, 00:09:06.793 "copy": true, 00:09:06.793 "nvme_iov_md": false 00:09:06.793 }, 00:09:06.793 "memory_domains": [ 00:09:06.793 { 00:09:06.793 "dma_device_id": "system", 00:09:06.793 "dma_device_type": 1 00:09:06.793 }, 00:09:06.793 { 00:09:06.793 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:06.793 "dma_device_type": 2 00:09:06.793 } 00:09:06.793 ], 00:09:06.793 "driver_specific": {} 00:09:06.793 } 00:09:06.793 ] 00:09:06.793 17:30:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:06.793 17:30:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:09:06.793 17:30:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:09:06.793 17:30:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:06.793 17:30:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:09:06.793 17:30:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:06.793 17:30:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:06.793 17:30:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:06.793 17:30:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:06.793 17:30:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:06.793 17:30:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:06.793 17:30:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:06.793 17:30:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:06.793 17:30:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:06.793 17:30:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:06.793 17:30:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:06.793 17:30:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:06.793 17:30:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:06.793 17:30:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:06.793 17:30:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:06.793 "name": "Existed_Raid", 00:09:06.793 "uuid": "7dc84895-f60b-4cfb-9bab-dcf15dfd7be0", 00:09:06.793 "strip_size_kb": 64, 00:09:06.793 "state": "configuring", 00:09:06.793 "raid_level": "concat", 00:09:06.793 "superblock": true, 00:09:06.793 "num_base_bdevs": 3, 00:09:06.793 "num_base_bdevs_discovered": 2, 00:09:06.793 "num_base_bdevs_operational": 3, 00:09:06.794 "base_bdevs_list": [ 00:09:06.794 { 00:09:06.794 "name": "BaseBdev1", 00:09:06.794 "uuid": "dc037112-fc5e-46de-b45a-8a4ccc2db305", 00:09:06.794 "is_configured": true, 00:09:06.794 "data_offset": 2048, 00:09:06.794 "data_size": 63488 00:09:06.794 }, 00:09:06.794 { 00:09:06.794 "name": "BaseBdev2", 00:09:06.794 "uuid": "92243fb0-b81e-4e75-8858-7074e93e5e9d", 00:09:06.794 "is_configured": true, 00:09:06.794 "data_offset": 2048, 00:09:06.794 "data_size": 63488 00:09:06.794 }, 00:09:06.794 { 00:09:06.794 "name": "BaseBdev3", 00:09:06.794 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:06.794 "is_configured": false, 00:09:06.794 "data_offset": 0, 00:09:06.794 "data_size": 0 00:09:06.794 } 00:09:06.794 ] 00:09:06.794 }' 00:09:06.794 17:30:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:06.794 17:30:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:07.364 17:30:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:09:07.364 17:30:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:07.364 17:30:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:07.364 [2024-11-27 17:30:38.375408] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:07.364 [2024-11-27 17:30:38.375637] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:09:07.364 [2024-11-27 17:30:38.375659] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:09:07.364 BaseBdev3 00:09:07.364 [2024-11-27 17:30:38.375967] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:09:07.364 [2024-11-27 17:30:38.376114] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:09:07.364 [2024-11-27 17:30:38.376132] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:09:07.364 [2024-11-27 17:30:38.376295] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:07.364 17:30:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:07.364 17:30:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:09:07.364 17:30:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:09:07.364 17:30:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:09:07.364 17:30:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:09:07.364 17:30:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:09:07.364 17:30:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:09:07.364 17:30:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:09:07.364 17:30:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:07.364 17:30:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:07.364 17:30:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:07.364 17:30:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:09:07.364 17:30:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:07.364 17:30:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:07.364 [ 00:09:07.364 { 00:09:07.364 "name": "BaseBdev3", 00:09:07.364 "aliases": [ 00:09:07.364 "70b2a7d8-5f77-4c05-ac62-349d07a515b4" 00:09:07.364 ], 00:09:07.364 "product_name": "Malloc disk", 00:09:07.364 "block_size": 512, 00:09:07.364 "num_blocks": 65536, 00:09:07.364 "uuid": "70b2a7d8-5f77-4c05-ac62-349d07a515b4", 00:09:07.364 "assigned_rate_limits": { 00:09:07.364 "rw_ios_per_sec": 0, 00:09:07.364 "rw_mbytes_per_sec": 0, 00:09:07.364 "r_mbytes_per_sec": 0, 00:09:07.364 "w_mbytes_per_sec": 0 00:09:07.364 }, 00:09:07.364 "claimed": true, 00:09:07.364 "claim_type": "exclusive_write", 00:09:07.364 "zoned": false, 00:09:07.364 "supported_io_types": { 00:09:07.364 "read": true, 00:09:07.364 "write": true, 00:09:07.364 "unmap": true, 00:09:07.364 "flush": true, 00:09:07.364 "reset": true, 00:09:07.364 "nvme_admin": false, 00:09:07.364 "nvme_io": false, 00:09:07.364 "nvme_io_md": false, 00:09:07.364 "write_zeroes": true, 00:09:07.364 "zcopy": true, 00:09:07.364 "get_zone_info": false, 00:09:07.364 "zone_management": false, 00:09:07.364 "zone_append": false, 00:09:07.364 "compare": false, 00:09:07.364 "compare_and_write": false, 00:09:07.364 "abort": true, 00:09:07.364 "seek_hole": false, 00:09:07.364 "seek_data": false, 00:09:07.364 "copy": true, 00:09:07.364 "nvme_iov_md": false 00:09:07.364 }, 00:09:07.364 "memory_domains": [ 00:09:07.364 { 00:09:07.364 "dma_device_id": "system", 00:09:07.364 "dma_device_type": 1 00:09:07.364 }, 00:09:07.364 { 00:09:07.364 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:07.364 "dma_device_type": 2 00:09:07.364 } 00:09:07.364 ], 00:09:07.364 "driver_specific": {} 00:09:07.364 } 00:09:07.364 ] 00:09:07.364 17:30:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:07.364 17:30:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:09:07.364 17:30:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:09:07.364 17:30:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:07.365 17:30:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online concat 64 3 00:09:07.365 17:30:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:07.365 17:30:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:07.365 17:30:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:07.365 17:30:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:07.365 17:30:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:07.365 17:30:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:07.365 17:30:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:07.365 17:30:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:07.365 17:30:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:07.365 17:30:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:07.365 17:30:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:07.365 17:30:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:07.365 17:30:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:07.365 17:30:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:07.365 17:30:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:07.365 "name": "Existed_Raid", 00:09:07.365 "uuid": "7dc84895-f60b-4cfb-9bab-dcf15dfd7be0", 00:09:07.365 "strip_size_kb": 64, 00:09:07.365 "state": "online", 00:09:07.365 "raid_level": "concat", 00:09:07.365 "superblock": true, 00:09:07.365 "num_base_bdevs": 3, 00:09:07.365 "num_base_bdevs_discovered": 3, 00:09:07.365 "num_base_bdevs_operational": 3, 00:09:07.365 "base_bdevs_list": [ 00:09:07.365 { 00:09:07.365 "name": "BaseBdev1", 00:09:07.365 "uuid": "dc037112-fc5e-46de-b45a-8a4ccc2db305", 00:09:07.365 "is_configured": true, 00:09:07.365 "data_offset": 2048, 00:09:07.365 "data_size": 63488 00:09:07.365 }, 00:09:07.365 { 00:09:07.365 "name": "BaseBdev2", 00:09:07.365 "uuid": "92243fb0-b81e-4e75-8858-7074e93e5e9d", 00:09:07.365 "is_configured": true, 00:09:07.365 "data_offset": 2048, 00:09:07.365 "data_size": 63488 00:09:07.365 }, 00:09:07.365 { 00:09:07.365 "name": "BaseBdev3", 00:09:07.365 "uuid": "70b2a7d8-5f77-4c05-ac62-349d07a515b4", 00:09:07.365 "is_configured": true, 00:09:07.365 "data_offset": 2048, 00:09:07.365 "data_size": 63488 00:09:07.365 } 00:09:07.365 ] 00:09:07.365 }' 00:09:07.365 17:30:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:07.365 17:30:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:07.935 17:30:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:09:07.935 17:30:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:09:07.935 17:30:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:07.935 17:30:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:07.935 17:30:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:09:07.935 17:30:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:07.935 17:30:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:09:07.935 17:30:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:07.935 17:30:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:07.935 17:30:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:07.935 [2024-11-27 17:30:38.870991] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:07.936 17:30:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:07.936 17:30:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:07.936 "name": "Existed_Raid", 00:09:07.936 "aliases": [ 00:09:07.936 "7dc84895-f60b-4cfb-9bab-dcf15dfd7be0" 00:09:07.936 ], 00:09:07.936 "product_name": "Raid Volume", 00:09:07.936 "block_size": 512, 00:09:07.936 "num_blocks": 190464, 00:09:07.936 "uuid": "7dc84895-f60b-4cfb-9bab-dcf15dfd7be0", 00:09:07.936 "assigned_rate_limits": { 00:09:07.936 "rw_ios_per_sec": 0, 00:09:07.936 "rw_mbytes_per_sec": 0, 00:09:07.936 "r_mbytes_per_sec": 0, 00:09:07.936 "w_mbytes_per_sec": 0 00:09:07.936 }, 00:09:07.936 "claimed": false, 00:09:07.936 "zoned": false, 00:09:07.936 "supported_io_types": { 00:09:07.936 "read": true, 00:09:07.936 "write": true, 00:09:07.936 "unmap": true, 00:09:07.936 "flush": true, 00:09:07.936 "reset": true, 00:09:07.936 "nvme_admin": false, 00:09:07.936 "nvme_io": false, 00:09:07.936 "nvme_io_md": false, 00:09:07.936 "write_zeroes": true, 00:09:07.936 "zcopy": false, 00:09:07.936 "get_zone_info": false, 00:09:07.936 "zone_management": false, 00:09:07.936 "zone_append": false, 00:09:07.936 "compare": false, 00:09:07.936 "compare_and_write": false, 00:09:07.936 "abort": false, 00:09:07.936 "seek_hole": false, 00:09:07.936 "seek_data": false, 00:09:07.936 "copy": false, 00:09:07.936 "nvme_iov_md": false 00:09:07.936 }, 00:09:07.936 "memory_domains": [ 00:09:07.936 { 00:09:07.936 "dma_device_id": "system", 00:09:07.936 "dma_device_type": 1 00:09:07.936 }, 00:09:07.936 { 00:09:07.936 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:07.936 "dma_device_type": 2 00:09:07.936 }, 00:09:07.936 { 00:09:07.936 "dma_device_id": "system", 00:09:07.936 "dma_device_type": 1 00:09:07.936 }, 00:09:07.936 { 00:09:07.936 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:07.936 "dma_device_type": 2 00:09:07.936 }, 00:09:07.936 { 00:09:07.936 "dma_device_id": "system", 00:09:07.936 "dma_device_type": 1 00:09:07.936 }, 00:09:07.936 { 00:09:07.936 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:07.936 "dma_device_type": 2 00:09:07.936 } 00:09:07.936 ], 00:09:07.936 "driver_specific": { 00:09:07.936 "raid": { 00:09:07.936 "uuid": "7dc84895-f60b-4cfb-9bab-dcf15dfd7be0", 00:09:07.936 "strip_size_kb": 64, 00:09:07.936 "state": "online", 00:09:07.936 "raid_level": "concat", 00:09:07.936 "superblock": true, 00:09:07.936 "num_base_bdevs": 3, 00:09:07.936 "num_base_bdevs_discovered": 3, 00:09:07.936 "num_base_bdevs_operational": 3, 00:09:07.936 "base_bdevs_list": [ 00:09:07.936 { 00:09:07.936 "name": "BaseBdev1", 00:09:07.936 "uuid": "dc037112-fc5e-46de-b45a-8a4ccc2db305", 00:09:07.936 "is_configured": true, 00:09:07.936 "data_offset": 2048, 00:09:07.936 "data_size": 63488 00:09:07.936 }, 00:09:07.936 { 00:09:07.936 "name": "BaseBdev2", 00:09:07.936 "uuid": "92243fb0-b81e-4e75-8858-7074e93e5e9d", 00:09:07.936 "is_configured": true, 00:09:07.936 "data_offset": 2048, 00:09:07.936 "data_size": 63488 00:09:07.936 }, 00:09:07.936 { 00:09:07.936 "name": "BaseBdev3", 00:09:07.936 "uuid": "70b2a7d8-5f77-4c05-ac62-349d07a515b4", 00:09:07.936 "is_configured": true, 00:09:07.936 "data_offset": 2048, 00:09:07.936 "data_size": 63488 00:09:07.936 } 00:09:07.936 ] 00:09:07.936 } 00:09:07.936 } 00:09:07.936 }' 00:09:07.936 17:30:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:07.936 17:30:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:09:07.936 BaseBdev2 00:09:07.936 BaseBdev3' 00:09:07.936 17:30:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:07.936 17:30:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:07.936 17:30:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:07.936 17:30:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:09:07.936 17:30:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:07.936 17:30:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:07.936 17:30:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:07.936 17:30:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:07.936 17:30:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:07.936 17:30:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:07.936 17:30:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:07.936 17:30:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:09:07.936 17:30:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:07.936 17:30:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:07.936 17:30:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:07.936 17:30:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:07.936 17:30:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:07.936 17:30:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:07.936 17:30:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:07.936 17:30:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:09:07.936 17:30:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:07.936 17:30:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:07.936 17:30:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:07.936 17:30:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:07.936 17:30:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:07.936 17:30:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:07.936 17:30:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:09:07.936 17:30:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:07.936 17:30:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:07.936 [2024-11-27 17:30:39.122328] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:09:07.936 [2024-11-27 17:30:39.122357] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:07.936 [2024-11-27 17:30:39.122430] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:08.196 17:30:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:08.196 17:30:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:09:08.196 17:30:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy concat 00:09:08.196 17:30:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:09:08.196 17:30:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # return 1 00:09:08.196 17:30:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:09:08.196 17:30:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline concat 64 2 00:09:08.196 17:30:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:08.196 17:30:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:09:08.196 17:30:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:08.196 17:30:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:08.196 17:30:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:09:08.196 17:30:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:08.196 17:30:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:08.196 17:30:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:08.196 17:30:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:08.196 17:30:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:08.196 17:30:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:08.196 17:30:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:08.196 17:30:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:08.196 17:30:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:08.196 17:30:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:08.196 "name": "Existed_Raid", 00:09:08.196 "uuid": "7dc84895-f60b-4cfb-9bab-dcf15dfd7be0", 00:09:08.196 "strip_size_kb": 64, 00:09:08.196 "state": "offline", 00:09:08.196 "raid_level": "concat", 00:09:08.196 "superblock": true, 00:09:08.196 "num_base_bdevs": 3, 00:09:08.196 "num_base_bdevs_discovered": 2, 00:09:08.196 "num_base_bdevs_operational": 2, 00:09:08.196 "base_bdevs_list": [ 00:09:08.196 { 00:09:08.196 "name": null, 00:09:08.196 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:08.196 "is_configured": false, 00:09:08.196 "data_offset": 0, 00:09:08.196 "data_size": 63488 00:09:08.196 }, 00:09:08.196 { 00:09:08.196 "name": "BaseBdev2", 00:09:08.196 "uuid": "92243fb0-b81e-4e75-8858-7074e93e5e9d", 00:09:08.196 "is_configured": true, 00:09:08.196 "data_offset": 2048, 00:09:08.196 "data_size": 63488 00:09:08.196 }, 00:09:08.196 { 00:09:08.196 "name": "BaseBdev3", 00:09:08.196 "uuid": "70b2a7d8-5f77-4c05-ac62-349d07a515b4", 00:09:08.196 "is_configured": true, 00:09:08.196 "data_offset": 2048, 00:09:08.196 "data_size": 63488 00:09:08.196 } 00:09:08.196 ] 00:09:08.196 }' 00:09:08.196 17:30:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:08.196 17:30:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:08.457 17:30:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:09:08.458 17:30:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:08.458 17:30:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:09:08.458 17:30:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:08.458 17:30:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:08.458 17:30:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:08.458 17:30:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:08.458 17:30:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:09:08.458 17:30:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:09:08.458 17:30:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:09:08.458 17:30:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:08.458 17:30:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:08.458 [2024-11-27 17:30:39.609893] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:09:08.458 17:30:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:08.458 17:30:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:09:08.458 17:30:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:08.458 17:30:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:08.458 17:30:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:09:08.458 17:30:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:08.458 17:30:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:08.717 17:30:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:08.717 17:30:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:09:08.717 17:30:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:09:08.717 17:30:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:09:08.717 17:30:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:08.717 17:30:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:08.717 [2024-11-27 17:30:39.690477] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:09:08.717 [2024-11-27 17:30:39.690595] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:09:08.717 17:30:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:08.717 17:30:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:09:08.717 17:30:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:08.717 17:30:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:08.717 17:30:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:09:08.717 17:30:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:08.717 17:30:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:08.717 17:30:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:08.718 17:30:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:09:08.718 17:30:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:09:08.718 17:30:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 3 -gt 2 ']' 00:09:08.718 17:30:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:09:08.718 17:30:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:08.718 17:30:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:09:08.718 17:30:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:08.718 17:30:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:08.718 BaseBdev2 00:09:08.718 17:30:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:08.718 17:30:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:09:08.718 17:30:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:09:08.718 17:30:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:09:08.718 17:30:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:09:08.718 17:30:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:09:08.718 17:30:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:09:08.718 17:30:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:09:08.718 17:30:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:08.718 17:30:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:08.718 17:30:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:08.718 17:30:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:09:08.718 17:30:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:08.718 17:30:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:08.718 [ 00:09:08.718 { 00:09:08.718 "name": "BaseBdev2", 00:09:08.718 "aliases": [ 00:09:08.718 "7d9d6516-22f1-48e6-a61d-40e01fb5d258" 00:09:08.718 ], 00:09:08.718 "product_name": "Malloc disk", 00:09:08.718 "block_size": 512, 00:09:08.718 "num_blocks": 65536, 00:09:08.718 "uuid": "7d9d6516-22f1-48e6-a61d-40e01fb5d258", 00:09:08.718 "assigned_rate_limits": { 00:09:08.718 "rw_ios_per_sec": 0, 00:09:08.718 "rw_mbytes_per_sec": 0, 00:09:08.718 "r_mbytes_per_sec": 0, 00:09:08.718 "w_mbytes_per_sec": 0 00:09:08.718 }, 00:09:08.718 "claimed": false, 00:09:08.718 "zoned": false, 00:09:08.718 "supported_io_types": { 00:09:08.718 "read": true, 00:09:08.718 "write": true, 00:09:08.718 "unmap": true, 00:09:08.718 "flush": true, 00:09:08.718 "reset": true, 00:09:08.718 "nvme_admin": false, 00:09:08.718 "nvme_io": false, 00:09:08.718 "nvme_io_md": false, 00:09:08.718 "write_zeroes": true, 00:09:08.718 "zcopy": true, 00:09:08.718 "get_zone_info": false, 00:09:08.718 "zone_management": false, 00:09:08.718 "zone_append": false, 00:09:08.718 "compare": false, 00:09:08.718 "compare_and_write": false, 00:09:08.718 "abort": true, 00:09:08.718 "seek_hole": false, 00:09:08.718 "seek_data": false, 00:09:08.718 "copy": true, 00:09:08.718 "nvme_iov_md": false 00:09:08.718 }, 00:09:08.718 "memory_domains": [ 00:09:08.718 { 00:09:08.718 "dma_device_id": "system", 00:09:08.718 "dma_device_type": 1 00:09:08.718 }, 00:09:08.718 { 00:09:08.718 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:08.718 "dma_device_type": 2 00:09:08.718 } 00:09:08.718 ], 00:09:08.718 "driver_specific": {} 00:09:08.718 } 00:09:08.718 ] 00:09:08.718 17:30:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:08.718 17:30:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:09:08.718 17:30:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:09:08.718 17:30:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:08.718 17:30:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:09:08.718 17:30:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:08.718 17:30:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:08.718 BaseBdev3 00:09:08.718 17:30:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:08.718 17:30:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:09:08.718 17:30:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:09:08.718 17:30:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:09:08.718 17:30:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:09:08.718 17:30:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:09:08.718 17:30:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:09:08.718 17:30:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:09:08.718 17:30:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:08.718 17:30:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:08.718 17:30:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:08.718 17:30:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:09:08.718 17:30:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:08.718 17:30:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:08.718 [ 00:09:08.718 { 00:09:08.718 "name": "BaseBdev3", 00:09:08.718 "aliases": [ 00:09:08.718 "a54823ea-108d-46e2-944b-311dcd1183bb" 00:09:08.718 ], 00:09:08.718 "product_name": "Malloc disk", 00:09:08.718 "block_size": 512, 00:09:08.718 "num_blocks": 65536, 00:09:08.718 "uuid": "a54823ea-108d-46e2-944b-311dcd1183bb", 00:09:08.718 "assigned_rate_limits": { 00:09:08.718 "rw_ios_per_sec": 0, 00:09:08.718 "rw_mbytes_per_sec": 0, 00:09:08.718 "r_mbytes_per_sec": 0, 00:09:08.718 "w_mbytes_per_sec": 0 00:09:08.718 }, 00:09:08.718 "claimed": false, 00:09:08.718 "zoned": false, 00:09:08.718 "supported_io_types": { 00:09:08.718 "read": true, 00:09:08.718 "write": true, 00:09:08.718 "unmap": true, 00:09:08.718 "flush": true, 00:09:08.718 "reset": true, 00:09:08.718 "nvme_admin": false, 00:09:08.718 "nvme_io": false, 00:09:08.718 "nvme_io_md": false, 00:09:08.718 "write_zeroes": true, 00:09:08.718 "zcopy": true, 00:09:08.718 "get_zone_info": false, 00:09:08.718 "zone_management": false, 00:09:08.718 "zone_append": false, 00:09:08.718 "compare": false, 00:09:08.718 "compare_and_write": false, 00:09:08.718 "abort": true, 00:09:08.718 "seek_hole": false, 00:09:08.718 "seek_data": false, 00:09:08.718 "copy": true, 00:09:08.718 "nvme_iov_md": false 00:09:08.718 }, 00:09:08.718 "memory_domains": [ 00:09:08.718 { 00:09:08.718 "dma_device_id": "system", 00:09:08.718 "dma_device_type": 1 00:09:08.718 }, 00:09:08.718 { 00:09:08.718 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:08.718 "dma_device_type": 2 00:09:08.718 } 00:09:08.718 ], 00:09:08.718 "driver_specific": {} 00:09:08.718 } 00:09:08.718 ] 00:09:08.718 17:30:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:08.718 17:30:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:09:08.718 17:30:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:09:08.718 17:30:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:08.718 17:30:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:09:08.718 17:30:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:08.718 17:30:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:08.718 [2024-11-27 17:30:39.884145] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:08.718 [2024-11-27 17:30:39.884214] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:08.718 [2024-11-27 17:30:39.884250] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:08.718 [2024-11-27 17:30:39.886344] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:08.718 17:30:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:08.718 17:30:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:09:08.718 17:30:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:08.718 17:30:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:08.718 17:30:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:08.718 17:30:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:08.718 17:30:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:08.718 17:30:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:08.718 17:30:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:08.718 17:30:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:08.718 17:30:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:08.718 17:30:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:08.718 17:30:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:08.719 17:30:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:08.719 17:30:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:08.978 17:30:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:08.978 17:30:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:08.978 "name": "Existed_Raid", 00:09:08.978 "uuid": "e210a160-f556-45f2-8da9-233b04a5dfdf", 00:09:08.978 "strip_size_kb": 64, 00:09:08.978 "state": "configuring", 00:09:08.978 "raid_level": "concat", 00:09:08.978 "superblock": true, 00:09:08.978 "num_base_bdevs": 3, 00:09:08.978 "num_base_bdevs_discovered": 2, 00:09:08.978 "num_base_bdevs_operational": 3, 00:09:08.978 "base_bdevs_list": [ 00:09:08.978 { 00:09:08.978 "name": "BaseBdev1", 00:09:08.978 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:08.978 "is_configured": false, 00:09:08.978 "data_offset": 0, 00:09:08.978 "data_size": 0 00:09:08.978 }, 00:09:08.978 { 00:09:08.978 "name": "BaseBdev2", 00:09:08.978 "uuid": "7d9d6516-22f1-48e6-a61d-40e01fb5d258", 00:09:08.978 "is_configured": true, 00:09:08.978 "data_offset": 2048, 00:09:08.978 "data_size": 63488 00:09:08.978 }, 00:09:08.978 { 00:09:08.978 "name": "BaseBdev3", 00:09:08.978 "uuid": "a54823ea-108d-46e2-944b-311dcd1183bb", 00:09:08.978 "is_configured": true, 00:09:08.978 "data_offset": 2048, 00:09:08.978 "data_size": 63488 00:09:08.978 } 00:09:08.978 ] 00:09:08.978 }' 00:09:08.978 17:30:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:08.978 17:30:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:09.237 17:30:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:09:09.237 17:30:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:09.237 17:30:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:09.237 [2024-11-27 17:30:40.339307] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:09:09.237 17:30:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:09.237 17:30:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:09:09.237 17:30:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:09.237 17:30:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:09.238 17:30:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:09.238 17:30:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:09.238 17:30:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:09.238 17:30:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:09.238 17:30:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:09.238 17:30:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:09.238 17:30:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:09.238 17:30:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:09.238 17:30:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:09.238 17:30:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:09.238 17:30:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:09.238 17:30:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:09.238 17:30:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:09.238 "name": "Existed_Raid", 00:09:09.238 "uuid": "e210a160-f556-45f2-8da9-233b04a5dfdf", 00:09:09.238 "strip_size_kb": 64, 00:09:09.238 "state": "configuring", 00:09:09.238 "raid_level": "concat", 00:09:09.238 "superblock": true, 00:09:09.238 "num_base_bdevs": 3, 00:09:09.238 "num_base_bdevs_discovered": 1, 00:09:09.238 "num_base_bdevs_operational": 3, 00:09:09.238 "base_bdevs_list": [ 00:09:09.238 { 00:09:09.238 "name": "BaseBdev1", 00:09:09.238 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:09.238 "is_configured": false, 00:09:09.238 "data_offset": 0, 00:09:09.238 "data_size": 0 00:09:09.238 }, 00:09:09.238 { 00:09:09.238 "name": null, 00:09:09.238 "uuid": "7d9d6516-22f1-48e6-a61d-40e01fb5d258", 00:09:09.238 "is_configured": false, 00:09:09.238 "data_offset": 0, 00:09:09.238 "data_size": 63488 00:09:09.238 }, 00:09:09.238 { 00:09:09.238 "name": "BaseBdev3", 00:09:09.238 "uuid": "a54823ea-108d-46e2-944b-311dcd1183bb", 00:09:09.238 "is_configured": true, 00:09:09.238 "data_offset": 2048, 00:09:09.238 "data_size": 63488 00:09:09.238 } 00:09:09.238 ] 00:09:09.238 }' 00:09:09.238 17:30:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:09.238 17:30:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:09.807 17:30:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:09.807 17:30:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:09:09.807 17:30:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:09.807 17:30:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:09.807 17:30:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:09.807 17:30:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:09:09.807 17:30:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:09:09.808 17:30:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:09.808 17:30:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:09.808 [2024-11-27 17:30:40.839251] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:09.808 BaseBdev1 00:09:09.808 17:30:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:09.808 17:30:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:09:09.808 17:30:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:09:09.808 17:30:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:09:09.808 17:30:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:09:09.808 17:30:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:09:09.808 17:30:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:09:09.808 17:30:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:09:09.808 17:30:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:09.808 17:30:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:09.808 17:30:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:09.808 17:30:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:09:09.808 17:30:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:09.808 17:30:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:09.808 [ 00:09:09.808 { 00:09:09.808 "name": "BaseBdev1", 00:09:09.808 "aliases": [ 00:09:09.808 "40052452-280c-4dad-91fc-d45b1ad951ba" 00:09:09.808 ], 00:09:09.808 "product_name": "Malloc disk", 00:09:09.808 "block_size": 512, 00:09:09.808 "num_blocks": 65536, 00:09:09.808 "uuid": "40052452-280c-4dad-91fc-d45b1ad951ba", 00:09:09.808 "assigned_rate_limits": { 00:09:09.808 "rw_ios_per_sec": 0, 00:09:09.808 "rw_mbytes_per_sec": 0, 00:09:09.808 "r_mbytes_per_sec": 0, 00:09:09.808 "w_mbytes_per_sec": 0 00:09:09.808 }, 00:09:09.808 "claimed": true, 00:09:09.808 "claim_type": "exclusive_write", 00:09:09.808 "zoned": false, 00:09:09.808 "supported_io_types": { 00:09:09.808 "read": true, 00:09:09.808 "write": true, 00:09:09.808 "unmap": true, 00:09:09.808 "flush": true, 00:09:09.808 "reset": true, 00:09:09.808 "nvme_admin": false, 00:09:09.808 "nvme_io": false, 00:09:09.808 "nvme_io_md": false, 00:09:09.808 "write_zeroes": true, 00:09:09.808 "zcopy": true, 00:09:09.808 "get_zone_info": false, 00:09:09.808 "zone_management": false, 00:09:09.808 "zone_append": false, 00:09:09.808 "compare": false, 00:09:09.808 "compare_and_write": false, 00:09:09.808 "abort": true, 00:09:09.808 "seek_hole": false, 00:09:09.808 "seek_data": false, 00:09:09.808 "copy": true, 00:09:09.808 "nvme_iov_md": false 00:09:09.808 }, 00:09:09.808 "memory_domains": [ 00:09:09.808 { 00:09:09.808 "dma_device_id": "system", 00:09:09.808 "dma_device_type": 1 00:09:09.808 }, 00:09:09.808 { 00:09:09.808 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:09.808 "dma_device_type": 2 00:09:09.808 } 00:09:09.808 ], 00:09:09.808 "driver_specific": {} 00:09:09.808 } 00:09:09.808 ] 00:09:09.808 17:30:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:09.808 17:30:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:09:09.808 17:30:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:09:09.808 17:30:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:09.808 17:30:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:09.808 17:30:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:09.808 17:30:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:09.808 17:30:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:09.808 17:30:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:09.808 17:30:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:09.808 17:30:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:09.808 17:30:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:09.808 17:30:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:09.808 17:30:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:09.808 17:30:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:09.808 17:30:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:09.808 17:30:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:09.808 17:30:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:09.808 "name": "Existed_Raid", 00:09:09.808 "uuid": "e210a160-f556-45f2-8da9-233b04a5dfdf", 00:09:09.808 "strip_size_kb": 64, 00:09:09.808 "state": "configuring", 00:09:09.808 "raid_level": "concat", 00:09:09.808 "superblock": true, 00:09:09.808 "num_base_bdevs": 3, 00:09:09.808 "num_base_bdevs_discovered": 2, 00:09:09.808 "num_base_bdevs_operational": 3, 00:09:09.808 "base_bdevs_list": [ 00:09:09.808 { 00:09:09.808 "name": "BaseBdev1", 00:09:09.808 "uuid": "40052452-280c-4dad-91fc-d45b1ad951ba", 00:09:09.808 "is_configured": true, 00:09:09.808 "data_offset": 2048, 00:09:09.808 "data_size": 63488 00:09:09.808 }, 00:09:09.808 { 00:09:09.808 "name": null, 00:09:09.808 "uuid": "7d9d6516-22f1-48e6-a61d-40e01fb5d258", 00:09:09.808 "is_configured": false, 00:09:09.808 "data_offset": 0, 00:09:09.808 "data_size": 63488 00:09:09.808 }, 00:09:09.808 { 00:09:09.808 "name": "BaseBdev3", 00:09:09.808 "uuid": "a54823ea-108d-46e2-944b-311dcd1183bb", 00:09:09.808 "is_configured": true, 00:09:09.808 "data_offset": 2048, 00:09:09.808 "data_size": 63488 00:09:09.808 } 00:09:09.808 ] 00:09:09.808 }' 00:09:09.808 17:30:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:09.808 17:30:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:10.379 17:30:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:10.379 17:30:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:10.379 17:30:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:10.379 17:30:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:09:10.379 17:30:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:10.379 17:30:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:09:10.379 17:30:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:09:10.379 17:30:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:10.379 17:30:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:10.379 [2024-11-27 17:30:41.338418] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:09:10.379 17:30:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:10.379 17:30:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:09:10.379 17:30:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:10.379 17:30:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:10.379 17:30:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:10.379 17:30:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:10.379 17:30:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:10.379 17:30:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:10.379 17:30:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:10.379 17:30:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:10.379 17:30:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:10.379 17:30:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:10.379 17:30:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:10.379 17:30:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:10.379 17:30:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:10.379 17:30:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:10.379 17:30:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:10.379 "name": "Existed_Raid", 00:09:10.379 "uuid": "e210a160-f556-45f2-8da9-233b04a5dfdf", 00:09:10.379 "strip_size_kb": 64, 00:09:10.379 "state": "configuring", 00:09:10.379 "raid_level": "concat", 00:09:10.379 "superblock": true, 00:09:10.379 "num_base_bdevs": 3, 00:09:10.379 "num_base_bdevs_discovered": 1, 00:09:10.379 "num_base_bdevs_operational": 3, 00:09:10.379 "base_bdevs_list": [ 00:09:10.379 { 00:09:10.379 "name": "BaseBdev1", 00:09:10.379 "uuid": "40052452-280c-4dad-91fc-d45b1ad951ba", 00:09:10.379 "is_configured": true, 00:09:10.379 "data_offset": 2048, 00:09:10.379 "data_size": 63488 00:09:10.379 }, 00:09:10.379 { 00:09:10.379 "name": null, 00:09:10.379 "uuid": "7d9d6516-22f1-48e6-a61d-40e01fb5d258", 00:09:10.379 "is_configured": false, 00:09:10.379 "data_offset": 0, 00:09:10.379 "data_size": 63488 00:09:10.379 }, 00:09:10.379 { 00:09:10.379 "name": null, 00:09:10.379 "uuid": "a54823ea-108d-46e2-944b-311dcd1183bb", 00:09:10.379 "is_configured": false, 00:09:10.379 "data_offset": 0, 00:09:10.379 "data_size": 63488 00:09:10.379 } 00:09:10.379 ] 00:09:10.379 }' 00:09:10.379 17:30:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:10.379 17:30:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:10.639 17:30:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:10.639 17:30:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:10.639 17:30:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:09:10.639 17:30:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:10.639 17:30:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:10.639 17:30:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:09:10.639 17:30:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:09:10.639 17:30:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:10.639 17:30:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:10.639 [2024-11-27 17:30:41.825608] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:10.899 17:30:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:10.899 17:30:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:09:10.899 17:30:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:10.899 17:30:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:10.899 17:30:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:10.899 17:30:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:10.899 17:30:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:10.899 17:30:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:10.899 17:30:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:10.899 17:30:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:10.899 17:30:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:10.899 17:30:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:10.899 17:30:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:10.899 17:30:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:10.899 17:30:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:10.899 17:30:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:10.899 17:30:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:10.899 "name": "Existed_Raid", 00:09:10.899 "uuid": "e210a160-f556-45f2-8da9-233b04a5dfdf", 00:09:10.899 "strip_size_kb": 64, 00:09:10.899 "state": "configuring", 00:09:10.899 "raid_level": "concat", 00:09:10.899 "superblock": true, 00:09:10.899 "num_base_bdevs": 3, 00:09:10.899 "num_base_bdevs_discovered": 2, 00:09:10.899 "num_base_bdevs_operational": 3, 00:09:10.899 "base_bdevs_list": [ 00:09:10.899 { 00:09:10.899 "name": "BaseBdev1", 00:09:10.899 "uuid": "40052452-280c-4dad-91fc-d45b1ad951ba", 00:09:10.899 "is_configured": true, 00:09:10.899 "data_offset": 2048, 00:09:10.899 "data_size": 63488 00:09:10.899 }, 00:09:10.899 { 00:09:10.899 "name": null, 00:09:10.899 "uuid": "7d9d6516-22f1-48e6-a61d-40e01fb5d258", 00:09:10.899 "is_configured": false, 00:09:10.899 "data_offset": 0, 00:09:10.899 "data_size": 63488 00:09:10.899 }, 00:09:10.899 { 00:09:10.899 "name": "BaseBdev3", 00:09:10.899 "uuid": "a54823ea-108d-46e2-944b-311dcd1183bb", 00:09:10.899 "is_configured": true, 00:09:10.899 "data_offset": 2048, 00:09:10.899 "data_size": 63488 00:09:10.899 } 00:09:10.899 ] 00:09:10.899 }' 00:09:10.899 17:30:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:10.899 17:30:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:11.159 17:30:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:11.159 17:30:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:11.159 17:30:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:11.159 17:30:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:09:11.159 17:30:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:11.159 17:30:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:09:11.159 17:30:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:09:11.159 17:30:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:11.159 17:30:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:11.159 [2024-11-27 17:30:42.332790] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:09:11.420 17:30:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:11.420 17:30:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:09:11.420 17:30:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:11.420 17:30:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:11.420 17:30:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:11.420 17:30:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:11.420 17:30:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:11.420 17:30:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:11.420 17:30:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:11.420 17:30:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:11.420 17:30:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:11.420 17:30:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:11.420 17:30:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:11.420 17:30:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:11.420 17:30:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:11.420 17:30:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:11.420 17:30:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:11.420 "name": "Existed_Raid", 00:09:11.420 "uuid": "e210a160-f556-45f2-8da9-233b04a5dfdf", 00:09:11.420 "strip_size_kb": 64, 00:09:11.420 "state": "configuring", 00:09:11.420 "raid_level": "concat", 00:09:11.420 "superblock": true, 00:09:11.420 "num_base_bdevs": 3, 00:09:11.420 "num_base_bdevs_discovered": 1, 00:09:11.420 "num_base_bdevs_operational": 3, 00:09:11.420 "base_bdevs_list": [ 00:09:11.420 { 00:09:11.420 "name": null, 00:09:11.420 "uuid": "40052452-280c-4dad-91fc-d45b1ad951ba", 00:09:11.420 "is_configured": false, 00:09:11.420 "data_offset": 0, 00:09:11.420 "data_size": 63488 00:09:11.420 }, 00:09:11.420 { 00:09:11.420 "name": null, 00:09:11.420 "uuid": "7d9d6516-22f1-48e6-a61d-40e01fb5d258", 00:09:11.420 "is_configured": false, 00:09:11.420 "data_offset": 0, 00:09:11.420 "data_size": 63488 00:09:11.420 }, 00:09:11.420 { 00:09:11.420 "name": "BaseBdev3", 00:09:11.420 "uuid": "a54823ea-108d-46e2-944b-311dcd1183bb", 00:09:11.420 "is_configured": true, 00:09:11.420 "data_offset": 2048, 00:09:11.420 "data_size": 63488 00:09:11.420 } 00:09:11.420 ] 00:09:11.420 }' 00:09:11.420 17:30:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:11.420 17:30:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:11.680 17:30:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:09:11.680 17:30:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:11.680 17:30:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:11.680 17:30:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:11.680 17:30:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:11.680 17:30:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:09:11.680 17:30:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:09:11.681 17:30:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:11.681 17:30:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:11.681 [2024-11-27 17:30:42.855546] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:11.681 17:30:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:11.681 17:30:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 3 00:09:11.681 17:30:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:11.681 17:30:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:11.681 17:30:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:11.681 17:30:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:11.681 17:30:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:11.681 17:30:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:11.681 17:30:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:11.681 17:30:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:11.681 17:30:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:11.681 17:30:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:11.681 17:30:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:11.681 17:30:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:11.681 17:30:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:11.952 17:30:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:11.952 17:30:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:11.952 "name": "Existed_Raid", 00:09:11.952 "uuid": "e210a160-f556-45f2-8da9-233b04a5dfdf", 00:09:11.952 "strip_size_kb": 64, 00:09:11.952 "state": "configuring", 00:09:11.952 "raid_level": "concat", 00:09:11.952 "superblock": true, 00:09:11.952 "num_base_bdevs": 3, 00:09:11.952 "num_base_bdevs_discovered": 2, 00:09:11.952 "num_base_bdevs_operational": 3, 00:09:11.952 "base_bdevs_list": [ 00:09:11.952 { 00:09:11.952 "name": null, 00:09:11.952 "uuid": "40052452-280c-4dad-91fc-d45b1ad951ba", 00:09:11.952 "is_configured": false, 00:09:11.952 "data_offset": 0, 00:09:11.952 "data_size": 63488 00:09:11.952 }, 00:09:11.952 { 00:09:11.952 "name": "BaseBdev2", 00:09:11.952 "uuid": "7d9d6516-22f1-48e6-a61d-40e01fb5d258", 00:09:11.952 "is_configured": true, 00:09:11.952 "data_offset": 2048, 00:09:11.952 "data_size": 63488 00:09:11.952 }, 00:09:11.952 { 00:09:11.952 "name": "BaseBdev3", 00:09:11.952 "uuid": "a54823ea-108d-46e2-944b-311dcd1183bb", 00:09:11.952 "is_configured": true, 00:09:11.952 "data_offset": 2048, 00:09:11.952 "data_size": 63488 00:09:11.952 } 00:09:11.952 ] 00:09:11.952 }' 00:09:11.952 17:30:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:11.952 17:30:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:12.225 17:30:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:12.225 17:30:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:12.225 17:30:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:12.225 17:30:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:09:12.225 17:30:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:12.225 17:30:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:09:12.225 17:30:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:09:12.225 17:30:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:12.225 17:30:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:12.225 17:30:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:12.225 17:30:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:12.225 17:30:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 40052452-280c-4dad-91fc-d45b1ad951ba 00:09:12.225 17:30:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:12.225 17:30:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:12.225 [2024-11-27 17:30:43.367486] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:09:12.225 [2024-11-27 17:30:43.367753] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:09:12.225 [2024-11-27 17:30:43.367804] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:09:12.225 [2024-11-27 17:30:43.368108] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002870 00:09:12.225 NewBaseBdev 00:09:12.225 [2024-11-27 17:30:43.368294] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:09:12.226 [2024-11-27 17:30:43.368352] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001c80 00:09:12.226 [2024-11-27 17:30:43.368507] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:12.226 17:30:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:12.226 17:30:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:09:12.226 17:30:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=NewBaseBdev 00:09:12.226 17:30:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:09:12.226 17:30:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:09:12.226 17:30:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:09:12.226 17:30:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:09:12.226 17:30:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:09:12.226 17:30:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:12.226 17:30:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:12.226 17:30:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:12.226 17:30:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:09:12.226 17:30:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:12.226 17:30:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:12.226 [ 00:09:12.226 { 00:09:12.226 "name": "NewBaseBdev", 00:09:12.226 "aliases": [ 00:09:12.226 "40052452-280c-4dad-91fc-d45b1ad951ba" 00:09:12.226 ], 00:09:12.226 "product_name": "Malloc disk", 00:09:12.226 "block_size": 512, 00:09:12.226 "num_blocks": 65536, 00:09:12.226 "uuid": "40052452-280c-4dad-91fc-d45b1ad951ba", 00:09:12.226 "assigned_rate_limits": { 00:09:12.226 "rw_ios_per_sec": 0, 00:09:12.226 "rw_mbytes_per_sec": 0, 00:09:12.226 "r_mbytes_per_sec": 0, 00:09:12.226 "w_mbytes_per_sec": 0 00:09:12.226 }, 00:09:12.226 "claimed": true, 00:09:12.226 "claim_type": "exclusive_write", 00:09:12.226 "zoned": false, 00:09:12.226 "supported_io_types": { 00:09:12.226 "read": true, 00:09:12.226 "write": true, 00:09:12.226 "unmap": true, 00:09:12.226 "flush": true, 00:09:12.226 "reset": true, 00:09:12.226 "nvme_admin": false, 00:09:12.226 "nvme_io": false, 00:09:12.226 "nvme_io_md": false, 00:09:12.226 "write_zeroes": true, 00:09:12.226 "zcopy": true, 00:09:12.226 "get_zone_info": false, 00:09:12.226 "zone_management": false, 00:09:12.226 "zone_append": false, 00:09:12.226 "compare": false, 00:09:12.226 "compare_and_write": false, 00:09:12.226 "abort": true, 00:09:12.226 "seek_hole": false, 00:09:12.226 "seek_data": false, 00:09:12.226 "copy": true, 00:09:12.226 "nvme_iov_md": false 00:09:12.226 }, 00:09:12.226 "memory_domains": [ 00:09:12.226 { 00:09:12.226 "dma_device_id": "system", 00:09:12.226 "dma_device_type": 1 00:09:12.226 }, 00:09:12.226 { 00:09:12.226 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:12.226 "dma_device_type": 2 00:09:12.226 } 00:09:12.226 ], 00:09:12.226 "driver_specific": {} 00:09:12.226 } 00:09:12.226 ] 00:09:12.226 17:30:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:12.226 17:30:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:09:12.226 17:30:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online concat 64 3 00:09:12.226 17:30:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:12.226 17:30:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:12.226 17:30:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:12.226 17:30:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:12.226 17:30:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:12.226 17:30:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:12.226 17:30:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:12.226 17:30:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:12.226 17:30:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:12.226 17:30:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:12.226 17:30:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:12.226 17:30:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:12.226 17:30:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:12.486 17:30:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:12.486 17:30:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:12.486 "name": "Existed_Raid", 00:09:12.486 "uuid": "e210a160-f556-45f2-8da9-233b04a5dfdf", 00:09:12.486 "strip_size_kb": 64, 00:09:12.486 "state": "online", 00:09:12.486 "raid_level": "concat", 00:09:12.486 "superblock": true, 00:09:12.486 "num_base_bdevs": 3, 00:09:12.486 "num_base_bdevs_discovered": 3, 00:09:12.486 "num_base_bdevs_operational": 3, 00:09:12.486 "base_bdevs_list": [ 00:09:12.486 { 00:09:12.486 "name": "NewBaseBdev", 00:09:12.486 "uuid": "40052452-280c-4dad-91fc-d45b1ad951ba", 00:09:12.486 "is_configured": true, 00:09:12.486 "data_offset": 2048, 00:09:12.486 "data_size": 63488 00:09:12.486 }, 00:09:12.486 { 00:09:12.486 "name": "BaseBdev2", 00:09:12.486 "uuid": "7d9d6516-22f1-48e6-a61d-40e01fb5d258", 00:09:12.486 "is_configured": true, 00:09:12.486 "data_offset": 2048, 00:09:12.486 "data_size": 63488 00:09:12.486 }, 00:09:12.486 { 00:09:12.486 "name": "BaseBdev3", 00:09:12.486 "uuid": "a54823ea-108d-46e2-944b-311dcd1183bb", 00:09:12.486 "is_configured": true, 00:09:12.486 "data_offset": 2048, 00:09:12.486 "data_size": 63488 00:09:12.486 } 00:09:12.486 ] 00:09:12.486 }' 00:09:12.486 17:30:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:12.486 17:30:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:12.746 17:30:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:09:12.746 17:30:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:09:12.746 17:30:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:12.746 17:30:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:12.746 17:30:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:09:12.746 17:30:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:12.746 17:30:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:09:12.746 17:30:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:12.746 17:30:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:12.746 17:30:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:12.746 [2024-11-27 17:30:43.839001] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:12.746 17:30:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:12.746 17:30:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:12.746 "name": "Existed_Raid", 00:09:12.746 "aliases": [ 00:09:12.746 "e210a160-f556-45f2-8da9-233b04a5dfdf" 00:09:12.746 ], 00:09:12.746 "product_name": "Raid Volume", 00:09:12.746 "block_size": 512, 00:09:12.746 "num_blocks": 190464, 00:09:12.746 "uuid": "e210a160-f556-45f2-8da9-233b04a5dfdf", 00:09:12.746 "assigned_rate_limits": { 00:09:12.746 "rw_ios_per_sec": 0, 00:09:12.746 "rw_mbytes_per_sec": 0, 00:09:12.746 "r_mbytes_per_sec": 0, 00:09:12.746 "w_mbytes_per_sec": 0 00:09:12.746 }, 00:09:12.746 "claimed": false, 00:09:12.746 "zoned": false, 00:09:12.746 "supported_io_types": { 00:09:12.746 "read": true, 00:09:12.746 "write": true, 00:09:12.746 "unmap": true, 00:09:12.746 "flush": true, 00:09:12.746 "reset": true, 00:09:12.746 "nvme_admin": false, 00:09:12.746 "nvme_io": false, 00:09:12.746 "nvme_io_md": false, 00:09:12.746 "write_zeroes": true, 00:09:12.746 "zcopy": false, 00:09:12.746 "get_zone_info": false, 00:09:12.746 "zone_management": false, 00:09:12.746 "zone_append": false, 00:09:12.746 "compare": false, 00:09:12.746 "compare_and_write": false, 00:09:12.746 "abort": false, 00:09:12.746 "seek_hole": false, 00:09:12.746 "seek_data": false, 00:09:12.746 "copy": false, 00:09:12.746 "nvme_iov_md": false 00:09:12.746 }, 00:09:12.746 "memory_domains": [ 00:09:12.746 { 00:09:12.746 "dma_device_id": "system", 00:09:12.746 "dma_device_type": 1 00:09:12.746 }, 00:09:12.746 { 00:09:12.746 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:12.746 "dma_device_type": 2 00:09:12.746 }, 00:09:12.746 { 00:09:12.746 "dma_device_id": "system", 00:09:12.746 "dma_device_type": 1 00:09:12.746 }, 00:09:12.746 { 00:09:12.746 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:12.746 "dma_device_type": 2 00:09:12.746 }, 00:09:12.746 { 00:09:12.746 "dma_device_id": "system", 00:09:12.746 "dma_device_type": 1 00:09:12.746 }, 00:09:12.746 { 00:09:12.746 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:12.746 "dma_device_type": 2 00:09:12.746 } 00:09:12.746 ], 00:09:12.746 "driver_specific": { 00:09:12.746 "raid": { 00:09:12.746 "uuid": "e210a160-f556-45f2-8da9-233b04a5dfdf", 00:09:12.746 "strip_size_kb": 64, 00:09:12.746 "state": "online", 00:09:12.746 "raid_level": "concat", 00:09:12.746 "superblock": true, 00:09:12.746 "num_base_bdevs": 3, 00:09:12.746 "num_base_bdevs_discovered": 3, 00:09:12.746 "num_base_bdevs_operational": 3, 00:09:12.746 "base_bdevs_list": [ 00:09:12.746 { 00:09:12.746 "name": "NewBaseBdev", 00:09:12.746 "uuid": "40052452-280c-4dad-91fc-d45b1ad951ba", 00:09:12.746 "is_configured": true, 00:09:12.746 "data_offset": 2048, 00:09:12.746 "data_size": 63488 00:09:12.746 }, 00:09:12.746 { 00:09:12.746 "name": "BaseBdev2", 00:09:12.746 "uuid": "7d9d6516-22f1-48e6-a61d-40e01fb5d258", 00:09:12.746 "is_configured": true, 00:09:12.746 "data_offset": 2048, 00:09:12.746 "data_size": 63488 00:09:12.746 }, 00:09:12.746 { 00:09:12.746 "name": "BaseBdev3", 00:09:12.746 "uuid": "a54823ea-108d-46e2-944b-311dcd1183bb", 00:09:12.746 "is_configured": true, 00:09:12.746 "data_offset": 2048, 00:09:12.746 "data_size": 63488 00:09:12.746 } 00:09:12.746 ] 00:09:12.746 } 00:09:12.746 } 00:09:12.746 }' 00:09:12.746 17:30:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:12.746 17:30:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:09:12.746 BaseBdev2 00:09:12.746 BaseBdev3' 00:09:12.746 17:30:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:13.006 17:30:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:13.006 17:30:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:13.006 17:30:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:09:13.006 17:30:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:13.006 17:30:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:13.006 17:30:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:13.006 17:30:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:13.006 17:30:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:13.006 17:30:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:13.006 17:30:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:13.006 17:30:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:09:13.006 17:30:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:13.006 17:30:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:13.006 17:30:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:13.006 17:30:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:13.006 17:30:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:13.006 17:30:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:13.006 17:30:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:13.006 17:30:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:09:13.006 17:30:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:13.006 17:30:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:13.006 17:30:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:13.006 17:30:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:13.006 17:30:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:13.006 17:30:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:13.006 17:30:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:13.006 17:30:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:13.006 17:30:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:13.006 [2024-11-27 17:30:44.142187] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:13.006 [2024-11-27 17:30:44.142252] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:13.006 [2024-11-27 17:30:44.142348] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:13.006 [2024-11-27 17:30:44.142401] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:13.006 [2024-11-27 17:30:44.142414] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name Existed_Raid, state offline 00:09:13.006 17:30:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:13.006 17:30:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 77178 00:09:13.006 17:30:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@950 -- # '[' -z 77178 ']' 00:09:13.006 17:30:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # kill -0 77178 00:09:13.006 17:30:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@955 -- # uname 00:09:13.006 17:30:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:09:13.006 17:30:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 77178 00:09:13.006 killing process with pid 77178 00:09:13.006 17:30:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:09:13.006 17:30:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:09:13.006 17:30:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@968 -- # echo 'killing process with pid 77178' 00:09:13.006 17:30:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@969 -- # kill 77178 00:09:13.006 [2024-11-27 17:30:44.191410] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:09:13.006 17:30:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@974 -- # wait 77178 00:09:13.266 [2024-11-27 17:30:44.250029] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:09:13.526 17:30:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:09:13.526 00:09:13.526 real 0m9.075s 00:09:13.526 user 0m15.195s 00:09:13.526 sys 0m1.894s 00:09:13.526 17:30:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1126 -- # xtrace_disable 00:09:13.526 17:30:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:13.526 ************************************ 00:09:13.526 END TEST raid_state_function_test_sb 00:09:13.526 ************************************ 00:09:13.526 17:30:44 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test concat 3 00:09:13.526 17:30:44 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:09:13.526 17:30:44 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:09:13.526 17:30:44 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:09:13.526 ************************************ 00:09:13.526 START TEST raid_superblock_test 00:09:13.526 ************************************ 00:09:13.526 17:30:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1125 -- # raid_superblock_test concat 3 00:09:13.526 17:30:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=concat 00:09:13.526 17:30:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=3 00:09:13.526 17:30:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:09:13.526 17:30:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:09:13.526 17:30:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:09:13.526 17:30:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:09:13.526 17:30:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:09:13.526 17:30:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:09:13.526 17:30:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:09:13.526 17:30:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:09:13.526 17:30:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:09:13.526 17:30:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:09:13.526 17:30:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:09:13.526 17:30:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' concat '!=' raid1 ']' 00:09:13.526 17:30:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@405 -- # strip_size=64 00:09:13.526 17:30:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@406 -- # strip_size_create_arg='-z 64' 00:09:13.526 17:30:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=77781 00:09:13.526 17:30:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 77781 00:09:13.526 17:30:44 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:09:13.526 17:30:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@831 -- # '[' -z 77781 ']' 00:09:13.526 17:30:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:13.526 17:30:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:09:13.526 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:13.526 17:30:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:13.526 17:30:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:09:13.526 17:30:44 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:13.786 [2024-11-27 17:30:44.780482] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:09:13.786 [2024-11-27 17:30:44.780712] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid77781 ] 00:09:13.786 [2024-11-27 17:30:44.907124] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:13.786 [2024-11-27 17:30:44.974229] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:09:14.046 [2024-11-27 17:30:45.050193] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:14.046 [2024-11-27 17:30:45.050235] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:14.615 17:30:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:09:14.615 17:30:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@864 -- # return 0 00:09:14.615 17:30:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:09:14.615 17:30:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:09:14.615 17:30:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:09:14.615 17:30:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:09:14.615 17:30:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:09:14.615 17:30:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:09:14.615 17:30:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:09:14.615 17:30:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:09:14.615 17:30:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:09:14.615 17:30:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:14.615 17:30:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:14.615 malloc1 00:09:14.615 17:30:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:14.615 17:30:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:09:14.615 17:30:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:14.615 17:30:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:14.615 [2024-11-27 17:30:45.636357] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:09:14.615 [2024-11-27 17:30:45.636508] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:14.615 [2024-11-27 17:30:45.636534] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:09:14.615 [2024-11-27 17:30:45.636557] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:14.615 [2024-11-27 17:30:45.638992] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:14.615 [2024-11-27 17:30:45.639032] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:09:14.615 pt1 00:09:14.615 17:30:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:14.616 17:30:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:09:14.616 17:30:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:09:14.616 17:30:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:09:14.616 17:30:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:09:14.616 17:30:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:09:14.616 17:30:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:09:14.616 17:30:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:09:14.616 17:30:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:09:14.616 17:30:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:09:14.616 17:30:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:14.616 17:30:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:14.616 malloc2 00:09:14.616 17:30:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:14.616 17:30:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:09:14.616 17:30:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:14.616 17:30:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:14.616 [2024-11-27 17:30:45.679701] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:09:14.616 [2024-11-27 17:30:45.679810] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:14.616 [2024-11-27 17:30:45.679845] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:09:14.616 [2024-11-27 17:30:45.679897] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:14.616 [2024-11-27 17:30:45.682279] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:14.616 [2024-11-27 17:30:45.682358] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:09:14.616 pt2 00:09:14.616 17:30:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:14.616 17:30:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:09:14.616 17:30:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:09:14.616 17:30:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc3 00:09:14.616 17:30:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt3 00:09:14.616 17:30:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000003 00:09:14.616 17:30:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:09:14.616 17:30:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:09:14.616 17:30:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:09:14.616 17:30:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc3 00:09:14.616 17:30:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:14.616 17:30:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:14.616 malloc3 00:09:14.616 17:30:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:14.616 17:30:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:09:14.616 17:30:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:14.616 17:30:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:14.616 [2024-11-27 17:30:45.718089] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:09:14.616 [2024-11-27 17:30:45.718203] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:14.616 [2024-11-27 17:30:45.718255] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:09:14.616 [2024-11-27 17:30:45.718285] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:14.616 [2024-11-27 17:30:45.720691] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:14.616 [2024-11-27 17:30:45.720758] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:09:14.616 pt3 00:09:14.616 17:30:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:14.616 17:30:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:09:14.616 17:30:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:09:14.616 17:30:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''pt1 pt2 pt3'\''' -n raid_bdev1 -s 00:09:14.616 17:30:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:14.616 17:30:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:14.616 [2024-11-27 17:30:45.730173] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:09:14.616 [2024-11-27 17:30:45.732328] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:09:14.616 [2024-11-27 17:30:45.732433] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:09:14.616 [2024-11-27 17:30:45.732625] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:09:14.616 [2024-11-27 17:30:45.732668] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:09:14.616 [2024-11-27 17:30:45.732964] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:09:14.616 [2024-11-27 17:30:45.733153] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:09:14.616 [2024-11-27 17:30:45.733201] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:09:14.616 [2024-11-27 17:30:45.733358] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:14.616 17:30:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:14.616 17:30:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online concat 64 3 00:09:14.616 17:30:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:14.616 17:30:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:14.616 17:30:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:14.616 17:30:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:14.616 17:30:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:14.616 17:30:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:14.616 17:30:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:14.616 17:30:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:14.616 17:30:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:14.616 17:30:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:14.616 17:30:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:14.616 17:30:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:14.616 17:30:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:14.616 17:30:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:14.616 17:30:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:14.616 "name": "raid_bdev1", 00:09:14.616 "uuid": "8ef6a4fd-514a-4e5b-b3e9-80374fcd9b7b", 00:09:14.616 "strip_size_kb": 64, 00:09:14.616 "state": "online", 00:09:14.616 "raid_level": "concat", 00:09:14.616 "superblock": true, 00:09:14.616 "num_base_bdevs": 3, 00:09:14.616 "num_base_bdevs_discovered": 3, 00:09:14.616 "num_base_bdevs_operational": 3, 00:09:14.616 "base_bdevs_list": [ 00:09:14.616 { 00:09:14.616 "name": "pt1", 00:09:14.616 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:14.616 "is_configured": true, 00:09:14.616 "data_offset": 2048, 00:09:14.616 "data_size": 63488 00:09:14.616 }, 00:09:14.616 { 00:09:14.616 "name": "pt2", 00:09:14.616 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:14.616 "is_configured": true, 00:09:14.616 "data_offset": 2048, 00:09:14.616 "data_size": 63488 00:09:14.616 }, 00:09:14.616 { 00:09:14.616 "name": "pt3", 00:09:14.616 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:14.616 "is_configured": true, 00:09:14.616 "data_offset": 2048, 00:09:14.616 "data_size": 63488 00:09:14.616 } 00:09:14.616 ] 00:09:14.616 }' 00:09:14.616 17:30:45 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:14.616 17:30:45 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:15.186 17:30:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:09:15.186 17:30:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:09:15.186 17:30:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:15.186 17:30:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:15.186 17:30:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:09:15.186 17:30:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:15.186 17:30:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:09:15.186 17:30:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:15.186 17:30:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:15.186 17:30:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:15.186 [2024-11-27 17:30:46.157645] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:15.186 17:30:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:15.186 17:30:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:15.186 "name": "raid_bdev1", 00:09:15.186 "aliases": [ 00:09:15.186 "8ef6a4fd-514a-4e5b-b3e9-80374fcd9b7b" 00:09:15.186 ], 00:09:15.186 "product_name": "Raid Volume", 00:09:15.186 "block_size": 512, 00:09:15.186 "num_blocks": 190464, 00:09:15.186 "uuid": "8ef6a4fd-514a-4e5b-b3e9-80374fcd9b7b", 00:09:15.186 "assigned_rate_limits": { 00:09:15.186 "rw_ios_per_sec": 0, 00:09:15.186 "rw_mbytes_per_sec": 0, 00:09:15.186 "r_mbytes_per_sec": 0, 00:09:15.186 "w_mbytes_per_sec": 0 00:09:15.186 }, 00:09:15.186 "claimed": false, 00:09:15.186 "zoned": false, 00:09:15.186 "supported_io_types": { 00:09:15.186 "read": true, 00:09:15.186 "write": true, 00:09:15.186 "unmap": true, 00:09:15.186 "flush": true, 00:09:15.186 "reset": true, 00:09:15.186 "nvme_admin": false, 00:09:15.186 "nvme_io": false, 00:09:15.186 "nvme_io_md": false, 00:09:15.186 "write_zeroes": true, 00:09:15.186 "zcopy": false, 00:09:15.186 "get_zone_info": false, 00:09:15.186 "zone_management": false, 00:09:15.186 "zone_append": false, 00:09:15.186 "compare": false, 00:09:15.186 "compare_and_write": false, 00:09:15.186 "abort": false, 00:09:15.186 "seek_hole": false, 00:09:15.186 "seek_data": false, 00:09:15.186 "copy": false, 00:09:15.186 "nvme_iov_md": false 00:09:15.186 }, 00:09:15.186 "memory_domains": [ 00:09:15.186 { 00:09:15.186 "dma_device_id": "system", 00:09:15.186 "dma_device_type": 1 00:09:15.186 }, 00:09:15.186 { 00:09:15.186 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:15.186 "dma_device_type": 2 00:09:15.186 }, 00:09:15.186 { 00:09:15.186 "dma_device_id": "system", 00:09:15.186 "dma_device_type": 1 00:09:15.186 }, 00:09:15.186 { 00:09:15.186 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:15.186 "dma_device_type": 2 00:09:15.186 }, 00:09:15.186 { 00:09:15.186 "dma_device_id": "system", 00:09:15.186 "dma_device_type": 1 00:09:15.186 }, 00:09:15.186 { 00:09:15.186 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:15.186 "dma_device_type": 2 00:09:15.186 } 00:09:15.186 ], 00:09:15.186 "driver_specific": { 00:09:15.186 "raid": { 00:09:15.186 "uuid": "8ef6a4fd-514a-4e5b-b3e9-80374fcd9b7b", 00:09:15.186 "strip_size_kb": 64, 00:09:15.186 "state": "online", 00:09:15.186 "raid_level": "concat", 00:09:15.186 "superblock": true, 00:09:15.186 "num_base_bdevs": 3, 00:09:15.186 "num_base_bdevs_discovered": 3, 00:09:15.186 "num_base_bdevs_operational": 3, 00:09:15.186 "base_bdevs_list": [ 00:09:15.186 { 00:09:15.186 "name": "pt1", 00:09:15.186 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:15.186 "is_configured": true, 00:09:15.186 "data_offset": 2048, 00:09:15.186 "data_size": 63488 00:09:15.186 }, 00:09:15.186 { 00:09:15.186 "name": "pt2", 00:09:15.186 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:15.186 "is_configured": true, 00:09:15.186 "data_offset": 2048, 00:09:15.186 "data_size": 63488 00:09:15.186 }, 00:09:15.186 { 00:09:15.186 "name": "pt3", 00:09:15.186 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:15.186 "is_configured": true, 00:09:15.186 "data_offset": 2048, 00:09:15.186 "data_size": 63488 00:09:15.186 } 00:09:15.186 ] 00:09:15.186 } 00:09:15.186 } 00:09:15.186 }' 00:09:15.186 17:30:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:15.186 17:30:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:09:15.186 pt2 00:09:15.186 pt3' 00:09:15.186 17:30:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:15.186 17:30:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:15.186 17:30:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:15.186 17:30:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:15.186 17:30:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:09:15.186 17:30:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:15.186 17:30:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:15.187 17:30:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:15.187 17:30:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:15.187 17:30:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:15.187 17:30:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:15.187 17:30:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:09:15.187 17:30:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:15.187 17:30:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:15.187 17:30:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:15.187 17:30:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:15.187 17:30:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:15.187 17:30:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:15.187 17:30:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:15.187 17:30:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:09:15.187 17:30:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:15.187 17:30:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:15.187 17:30:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:15.187 17:30:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:15.187 17:30:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:15.187 17:30:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:15.447 17:30:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:09:15.447 17:30:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:15.447 17:30:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:15.447 17:30:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:09:15.447 [2024-11-27 17:30:46.381253] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:15.447 17:30:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:15.447 17:30:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=8ef6a4fd-514a-4e5b-b3e9-80374fcd9b7b 00:09:15.447 17:30:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z 8ef6a4fd-514a-4e5b-b3e9-80374fcd9b7b ']' 00:09:15.447 17:30:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:09:15.447 17:30:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:15.447 17:30:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:15.447 [2024-11-27 17:30:46.432909] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:15.447 [2024-11-27 17:30:46.432935] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:15.447 [2024-11-27 17:30:46.433015] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:15.447 [2024-11-27 17:30:46.433070] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:15.447 [2024-11-27 17:30:46.433085] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:09:15.447 17:30:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:15.447 17:30:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:15.447 17:30:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:15.447 17:30:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:15.447 17:30:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:09:15.447 17:30:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:15.447 17:30:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:09:15.447 17:30:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:09:15.447 17:30:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:09:15.447 17:30:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:09:15.447 17:30:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:15.447 17:30:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:15.447 17:30:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:15.447 17:30:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:09:15.447 17:30:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:09:15.447 17:30:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:15.447 17:30:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:15.447 17:30:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:15.447 17:30:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:09:15.447 17:30:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt3 00:09:15.447 17:30:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:15.447 17:30:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:15.447 17:30:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:15.447 17:30:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:09:15.447 17:30:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:09:15.447 17:30:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:15.447 17:30:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:15.447 17:30:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:15.447 17:30:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:09:15.447 17:30:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:09:15.447 17:30:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@650 -- # local es=0 00:09:15.447 17:30:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:09:15.447 17:30:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:09:15.447 17:30:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:09:15.447 17:30:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:09:15.447 17:30:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:09:15.447 17:30:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:09:15.447 17:30:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:15.447 17:30:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:15.447 [2024-11-27 17:30:46.580688] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:09:15.447 [2024-11-27 17:30:46.582906] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:09:15.447 [2024-11-27 17:30:46.582950] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc3 is claimed 00:09:15.447 [2024-11-27 17:30:46.583009] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:09:15.447 [2024-11-27 17:30:46.583057] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:09:15.447 [2024-11-27 17:30:46.583079] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc3 00:09:15.447 [2024-11-27 17:30:46.583092] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:15.447 [2024-11-27 17:30:46.583101] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state configuring 00:09:15.447 request: 00:09:15.447 { 00:09:15.447 "name": "raid_bdev1", 00:09:15.447 "raid_level": "concat", 00:09:15.447 "base_bdevs": [ 00:09:15.447 "malloc1", 00:09:15.447 "malloc2", 00:09:15.447 "malloc3" 00:09:15.447 ], 00:09:15.447 "strip_size_kb": 64, 00:09:15.447 "superblock": false, 00:09:15.447 "method": "bdev_raid_create", 00:09:15.447 "req_id": 1 00:09:15.447 } 00:09:15.447 Got JSON-RPC error response 00:09:15.447 response: 00:09:15.447 { 00:09:15.447 "code": -17, 00:09:15.447 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:09:15.447 } 00:09:15.447 17:30:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:09:15.447 17:30:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@653 -- # es=1 00:09:15.447 17:30:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:09:15.447 17:30:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:09:15.447 17:30:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:09:15.447 17:30:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:09:15.447 17:30:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:15.447 17:30:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:15.447 17:30:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:15.447 17:30:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:15.447 17:30:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:09:15.447 17:30:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:09:15.447 17:30:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:09:15.447 17:30:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:15.447 17:30:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:15.447 [2024-11-27 17:30:46.632551] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:09:15.447 [2024-11-27 17:30:46.632638] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:15.447 [2024-11-27 17:30:46.632686] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:09:15.447 [2024-11-27 17:30:46.632717] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:15.447 [2024-11-27 17:30:46.635305] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:15.447 [2024-11-27 17:30:46.635375] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:09:15.447 [2024-11-27 17:30:46.635465] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:09:15.447 [2024-11-27 17:30:46.635530] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:09:15.707 pt1 00:09:15.707 17:30:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:15.707 17:30:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring concat 64 3 00:09:15.707 17:30:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:15.707 17:30:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:15.707 17:30:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:15.707 17:30:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:15.707 17:30:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:15.707 17:30:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:15.707 17:30:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:15.707 17:30:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:15.707 17:30:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:15.707 17:30:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:15.707 17:30:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:15.707 17:30:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:15.707 17:30:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:15.707 17:30:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:15.707 17:30:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:15.707 "name": "raid_bdev1", 00:09:15.707 "uuid": "8ef6a4fd-514a-4e5b-b3e9-80374fcd9b7b", 00:09:15.707 "strip_size_kb": 64, 00:09:15.707 "state": "configuring", 00:09:15.707 "raid_level": "concat", 00:09:15.707 "superblock": true, 00:09:15.707 "num_base_bdevs": 3, 00:09:15.707 "num_base_bdevs_discovered": 1, 00:09:15.707 "num_base_bdevs_operational": 3, 00:09:15.707 "base_bdevs_list": [ 00:09:15.707 { 00:09:15.707 "name": "pt1", 00:09:15.707 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:15.707 "is_configured": true, 00:09:15.707 "data_offset": 2048, 00:09:15.707 "data_size": 63488 00:09:15.707 }, 00:09:15.707 { 00:09:15.707 "name": null, 00:09:15.707 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:15.707 "is_configured": false, 00:09:15.707 "data_offset": 2048, 00:09:15.707 "data_size": 63488 00:09:15.707 }, 00:09:15.707 { 00:09:15.707 "name": null, 00:09:15.707 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:15.707 "is_configured": false, 00:09:15.707 "data_offset": 2048, 00:09:15.707 "data_size": 63488 00:09:15.707 } 00:09:15.707 ] 00:09:15.707 }' 00:09:15.707 17:30:46 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:15.707 17:30:46 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:15.966 17:30:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 3 -gt 2 ']' 00:09:15.966 17:30:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@472 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:09:15.966 17:30:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:15.966 17:30:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:15.966 [2024-11-27 17:30:47.087787] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:09:15.966 [2024-11-27 17:30:47.087853] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:15.966 [2024-11-27 17:30:47.087874] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:09:15.966 [2024-11-27 17:30:47.087887] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:15.966 [2024-11-27 17:30:47.088322] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:15.966 [2024-11-27 17:30:47.088343] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:09:15.966 [2024-11-27 17:30:47.088408] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:09:15.966 [2024-11-27 17:30:47.088434] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:09:15.966 pt2 00:09:15.966 17:30:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:15.966 17:30:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@473 -- # rpc_cmd bdev_passthru_delete pt2 00:09:15.966 17:30:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:15.966 17:30:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:15.966 [2024-11-27 17:30:47.095780] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: pt2 00:09:15.966 17:30:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:15.966 17:30:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@474 -- # verify_raid_bdev_state raid_bdev1 configuring concat 64 3 00:09:15.966 17:30:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:15.966 17:30:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:15.966 17:30:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:15.966 17:30:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:15.966 17:30:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:15.966 17:30:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:15.966 17:30:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:15.966 17:30:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:15.966 17:30:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:15.966 17:30:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:15.966 17:30:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:15.966 17:30:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:15.966 17:30:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:15.966 17:30:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:15.966 17:30:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:15.966 "name": "raid_bdev1", 00:09:15.966 "uuid": "8ef6a4fd-514a-4e5b-b3e9-80374fcd9b7b", 00:09:15.966 "strip_size_kb": 64, 00:09:15.966 "state": "configuring", 00:09:15.966 "raid_level": "concat", 00:09:15.966 "superblock": true, 00:09:15.966 "num_base_bdevs": 3, 00:09:15.966 "num_base_bdevs_discovered": 1, 00:09:15.966 "num_base_bdevs_operational": 3, 00:09:15.966 "base_bdevs_list": [ 00:09:15.966 { 00:09:15.966 "name": "pt1", 00:09:15.966 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:15.966 "is_configured": true, 00:09:15.966 "data_offset": 2048, 00:09:15.966 "data_size": 63488 00:09:15.966 }, 00:09:15.966 { 00:09:15.966 "name": null, 00:09:15.966 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:15.966 "is_configured": false, 00:09:15.966 "data_offset": 0, 00:09:15.966 "data_size": 63488 00:09:15.966 }, 00:09:15.966 { 00:09:15.966 "name": null, 00:09:15.966 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:15.966 "is_configured": false, 00:09:15.966 "data_offset": 2048, 00:09:15.966 "data_size": 63488 00:09:15.966 } 00:09:15.966 ] 00:09:15.966 }' 00:09:15.966 17:30:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:15.966 17:30:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:16.536 17:30:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:09:16.536 17:30:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:09:16.536 17:30:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:09:16.536 17:30:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:16.536 17:30:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:16.536 [2024-11-27 17:30:47.519034] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:09:16.536 [2024-11-27 17:30:47.519127] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:16.536 [2024-11-27 17:30:47.519193] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:09:16.536 [2024-11-27 17:30:47.519243] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:16.536 [2024-11-27 17:30:47.519668] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:16.536 [2024-11-27 17:30:47.519723] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:09:16.536 [2024-11-27 17:30:47.519817] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:09:16.536 [2024-11-27 17:30:47.519863] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:09:16.536 pt2 00:09:16.536 17:30:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:16.536 17:30:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:09:16.536 17:30:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:09:16.536 17:30:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:09:16.536 17:30:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:16.536 17:30:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:16.536 [2024-11-27 17:30:47.531003] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:09:16.536 [2024-11-27 17:30:47.531077] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:16.536 [2024-11-27 17:30:47.531126] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:09:16.536 [2024-11-27 17:30:47.531169] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:16.536 [2024-11-27 17:30:47.531515] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:16.536 [2024-11-27 17:30:47.531566] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:09:16.536 [2024-11-27 17:30:47.531647] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:09:16.536 [2024-11-27 17:30:47.531703] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:09:16.536 [2024-11-27 17:30:47.531829] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:09:16.536 [2024-11-27 17:30:47.531865] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:09:16.536 [2024-11-27 17:30:47.532124] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:09:16.536 [2024-11-27 17:30:47.532275] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:09:16.536 [2024-11-27 17:30:47.532315] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:09:16.536 [2024-11-27 17:30:47.532442] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:16.536 pt3 00:09:16.536 17:30:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:16.536 17:30:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:09:16.536 17:30:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:09:16.536 17:30:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online concat 64 3 00:09:16.536 17:30:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:16.536 17:30:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:16.536 17:30:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:16.536 17:30:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:16.536 17:30:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:16.536 17:30:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:16.536 17:30:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:16.536 17:30:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:16.536 17:30:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:16.536 17:30:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:16.536 17:30:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:16.536 17:30:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:16.536 17:30:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:16.536 17:30:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:16.536 17:30:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:16.536 "name": "raid_bdev1", 00:09:16.536 "uuid": "8ef6a4fd-514a-4e5b-b3e9-80374fcd9b7b", 00:09:16.536 "strip_size_kb": 64, 00:09:16.536 "state": "online", 00:09:16.536 "raid_level": "concat", 00:09:16.536 "superblock": true, 00:09:16.536 "num_base_bdevs": 3, 00:09:16.536 "num_base_bdevs_discovered": 3, 00:09:16.536 "num_base_bdevs_operational": 3, 00:09:16.536 "base_bdevs_list": [ 00:09:16.536 { 00:09:16.536 "name": "pt1", 00:09:16.536 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:16.536 "is_configured": true, 00:09:16.536 "data_offset": 2048, 00:09:16.536 "data_size": 63488 00:09:16.536 }, 00:09:16.536 { 00:09:16.536 "name": "pt2", 00:09:16.536 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:16.536 "is_configured": true, 00:09:16.536 "data_offset": 2048, 00:09:16.536 "data_size": 63488 00:09:16.536 }, 00:09:16.536 { 00:09:16.536 "name": "pt3", 00:09:16.536 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:16.536 "is_configured": true, 00:09:16.536 "data_offset": 2048, 00:09:16.536 "data_size": 63488 00:09:16.536 } 00:09:16.536 ] 00:09:16.536 }' 00:09:16.536 17:30:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:16.536 17:30:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:16.796 17:30:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:09:16.796 17:30:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:09:16.796 17:30:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:16.796 17:30:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:16.796 17:30:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:09:16.796 17:30:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:16.796 17:30:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:09:16.796 17:30:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:16.796 17:30:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:16.796 17:30:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:16.796 [2024-11-27 17:30:47.954761] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:16.796 17:30:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:17.055 17:30:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:17.056 "name": "raid_bdev1", 00:09:17.056 "aliases": [ 00:09:17.056 "8ef6a4fd-514a-4e5b-b3e9-80374fcd9b7b" 00:09:17.056 ], 00:09:17.056 "product_name": "Raid Volume", 00:09:17.056 "block_size": 512, 00:09:17.056 "num_blocks": 190464, 00:09:17.056 "uuid": "8ef6a4fd-514a-4e5b-b3e9-80374fcd9b7b", 00:09:17.056 "assigned_rate_limits": { 00:09:17.056 "rw_ios_per_sec": 0, 00:09:17.056 "rw_mbytes_per_sec": 0, 00:09:17.056 "r_mbytes_per_sec": 0, 00:09:17.056 "w_mbytes_per_sec": 0 00:09:17.056 }, 00:09:17.056 "claimed": false, 00:09:17.056 "zoned": false, 00:09:17.056 "supported_io_types": { 00:09:17.056 "read": true, 00:09:17.056 "write": true, 00:09:17.056 "unmap": true, 00:09:17.056 "flush": true, 00:09:17.056 "reset": true, 00:09:17.056 "nvme_admin": false, 00:09:17.056 "nvme_io": false, 00:09:17.056 "nvme_io_md": false, 00:09:17.056 "write_zeroes": true, 00:09:17.056 "zcopy": false, 00:09:17.056 "get_zone_info": false, 00:09:17.056 "zone_management": false, 00:09:17.056 "zone_append": false, 00:09:17.056 "compare": false, 00:09:17.056 "compare_and_write": false, 00:09:17.056 "abort": false, 00:09:17.056 "seek_hole": false, 00:09:17.056 "seek_data": false, 00:09:17.056 "copy": false, 00:09:17.056 "nvme_iov_md": false 00:09:17.056 }, 00:09:17.056 "memory_domains": [ 00:09:17.056 { 00:09:17.056 "dma_device_id": "system", 00:09:17.056 "dma_device_type": 1 00:09:17.056 }, 00:09:17.056 { 00:09:17.056 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:17.056 "dma_device_type": 2 00:09:17.056 }, 00:09:17.056 { 00:09:17.056 "dma_device_id": "system", 00:09:17.056 "dma_device_type": 1 00:09:17.056 }, 00:09:17.056 { 00:09:17.056 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:17.056 "dma_device_type": 2 00:09:17.056 }, 00:09:17.056 { 00:09:17.056 "dma_device_id": "system", 00:09:17.056 "dma_device_type": 1 00:09:17.056 }, 00:09:17.056 { 00:09:17.056 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:17.056 "dma_device_type": 2 00:09:17.056 } 00:09:17.056 ], 00:09:17.056 "driver_specific": { 00:09:17.056 "raid": { 00:09:17.056 "uuid": "8ef6a4fd-514a-4e5b-b3e9-80374fcd9b7b", 00:09:17.056 "strip_size_kb": 64, 00:09:17.056 "state": "online", 00:09:17.056 "raid_level": "concat", 00:09:17.056 "superblock": true, 00:09:17.056 "num_base_bdevs": 3, 00:09:17.056 "num_base_bdevs_discovered": 3, 00:09:17.056 "num_base_bdevs_operational": 3, 00:09:17.056 "base_bdevs_list": [ 00:09:17.056 { 00:09:17.056 "name": "pt1", 00:09:17.056 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:17.056 "is_configured": true, 00:09:17.056 "data_offset": 2048, 00:09:17.056 "data_size": 63488 00:09:17.056 }, 00:09:17.056 { 00:09:17.056 "name": "pt2", 00:09:17.056 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:17.056 "is_configured": true, 00:09:17.056 "data_offset": 2048, 00:09:17.056 "data_size": 63488 00:09:17.056 }, 00:09:17.056 { 00:09:17.056 "name": "pt3", 00:09:17.056 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:17.056 "is_configured": true, 00:09:17.056 "data_offset": 2048, 00:09:17.056 "data_size": 63488 00:09:17.056 } 00:09:17.056 ] 00:09:17.056 } 00:09:17.056 } 00:09:17.056 }' 00:09:17.056 17:30:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:17.056 17:30:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:09:17.056 pt2 00:09:17.056 pt3' 00:09:17.056 17:30:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:17.056 17:30:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:17.056 17:30:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:17.056 17:30:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:17.056 17:30:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:09:17.056 17:30:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:17.056 17:30:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:17.056 17:30:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:17.056 17:30:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:17.056 17:30:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:17.056 17:30:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:17.056 17:30:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:09:17.056 17:30:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:17.056 17:30:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:17.056 17:30:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:17.056 17:30:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:17.056 17:30:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:17.056 17:30:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:17.056 17:30:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:17.056 17:30:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:09:17.056 17:30:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:17.056 17:30:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:17.056 17:30:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:17.056 17:30:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:17.056 17:30:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:17.056 17:30:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:17.056 17:30:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:09:17.056 17:30:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:09:17.056 17:30:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:17.056 17:30:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:17.056 [2024-11-27 17:30:48.242208] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:17.316 17:30:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:17.316 17:30:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' 8ef6a4fd-514a-4e5b-b3e9-80374fcd9b7b '!=' 8ef6a4fd-514a-4e5b-b3e9-80374fcd9b7b ']' 00:09:17.316 17:30:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy concat 00:09:17.316 17:30:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:09:17.316 17:30:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # return 1 00:09:17.316 17:30:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 77781 00:09:17.316 17:30:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@950 -- # '[' -z 77781 ']' 00:09:17.316 17:30:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # kill -0 77781 00:09:17.316 17:30:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@955 -- # uname 00:09:17.316 17:30:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:09:17.316 17:30:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 77781 00:09:17.316 17:30:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:09:17.316 17:30:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:09:17.316 17:30:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 77781' 00:09:17.316 killing process with pid 77781 00:09:17.316 17:30:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@969 -- # kill 77781 00:09:17.316 [2024-11-27 17:30:48.300841] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:09:17.316 [2024-11-27 17:30:48.300927] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:17.316 [2024-11-27 17:30:48.300996] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:17.316 [2024-11-27 17:30:48.301005] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:09:17.316 17:30:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@974 -- # wait 77781 00:09:17.316 [2024-11-27 17:30:48.361338] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:09:17.577 17:30:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:09:17.577 00:09:17.577 real 0m4.028s 00:09:17.577 user 0m6.147s 00:09:17.577 sys 0m0.901s 00:09:17.577 17:30:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:09:17.577 ************************************ 00:09:17.577 END TEST raid_superblock_test 00:09:17.577 ************************************ 00:09:17.577 17:30:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:17.839 17:30:48 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test concat 3 read 00:09:17.839 17:30:48 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:09:17.839 17:30:48 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:09:17.839 17:30:48 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:09:17.839 ************************************ 00:09:17.839 START TEST raid_read_error_test 00:09:17.839 ************************************ 00:09:17.839 17:30:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1125 -- # raid_io_error_test concat 3 read 00:09:17.839 17:30:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=concat 00:09:17.839 17:30:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=3 00:09:17.839 17:30:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:09:17.839 17:30:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:09:17.839 17:30:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:17.839 17:30:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:09:17.839 17:30:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:17.839 17:30:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:17.839 17:30:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:09:17.839 17:30:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:17.839 17:30:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:17.839 17:30:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:09:17.839 17:30:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:17.839 17:30:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:17.839 17:30:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:09:17.839 17:30:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:09:17.839 17:30:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:09:17.839 17:30:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:09:17.839 17:30:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:09:17.839 17:30:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:09:17.839 17:30:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:09:17.839 17:30:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' concat '!=' raid1 ']' 00:09:17.839 17:30:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:09:17.839 17:30:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:09:17.839 17:30:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:09:17.839 17:30:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.cY3Q2y3lMS 00:09:17.839 17:30:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=78023 00:09:17.839 17:30:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:09:17.839 17:30:48 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 78023 00:09:17.839 17:30:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@831 -- # '[' -z 78023 ']' 00:09:17.839 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:17.839 17:30:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:17.839 17:30:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:09:17.839 17:30:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:17.839 17:30:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:09:17.839 17:30:48 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:17.839 [2024-11-27 17:30:48.898738] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:09:17.839 [2024-11-27 17:30:48.898884] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid78023 ] 00:09:18.099 [2024-11-27 17:30:49.045565] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:18.099 [2024-11-27 17:30:49.114103] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:09:18.099 [2024-11-27 17:30:49.189877] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:18.099 [2024-11-27 17:30:49.189916] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:18.668 17:30:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:09:18.668 17:30:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@864 -- # return 0 00:09:18.668 17:30:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:18.668 17:30:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:09:18.668 17:30:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:18.668 17:30:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:18.668 BaseBdev1_malloc 00:09:18.668 17:30:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:18.668 17:30:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:09:18.668 17:30:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:18.668 17:30:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:18.668 true 00:09:18.669 17:30:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:18.669 17:30:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:09:18.669 17:30:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:18.669 17:30:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:18.669 [2024-11-27 17:30:49.755616] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:09:18.669 [2024-11-27 17:30:49.755684] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:18.669 [2024-11-27 17:30:49.755712] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:09:18.669 [2024-11-27 17:30:49.755723] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:18.669 [2024-11-27 17:30:49.758117] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:18.669 [2024-11-27 17:30:49.758162] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:09:18.669 BaseBdev1 00:09:18.669 17:30:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:18.669 17:30:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:18.669 17:30:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:09:18.669 17:30:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:18.669 17:30:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:18.669 BaseBdev2_malloc 00:09:18.669 17:30:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:18.669 17:30:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:09:18.669 17:30:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:18.669 17:30:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:18.669 true 00:09:18.669 17:30:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:18.669 17:30:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:09:18.669 17:30:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:18.669 17:30:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:18.669 [2024-11-27 17:30:49.819457] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:09:18.669 [2024-11-27 17:30:49.819591] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:18.669 [2024-11-27 17:30:49.819627] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:09:18.669 [2024-11-27 17:30:49.819640] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:18.669 [2024-11-27 17:30:49.822736] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:18.669 [2024-11-27 17:30:49.822835] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:09:18.669 BaseBdev2 00:09:18.669 17:30:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:18.669 17:30:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:18.669 17:30:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:09:18.669 17:30:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:18.669 17:30:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:18.669 BaseBdev3_malloc 00:09:18.669 17:30:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:18.669 17:30:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:09:18.669 17:30:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:18.669 17:30:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:18.929 true 00:09:18.929 17:30:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:18.929 17:30:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:09:18.929 17:30:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:18.929 17:30:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:18.929 [2024-11-27 17:30:49.865999] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:09:18.929 [2024-11-27 17:30:49.866049] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:18.929 [2024-11-27 17:30:49.866070] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008780 00:09:18.929 [2024-11-27 17:30:49.866079] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:18.929 [2024-11-27 17:30:49.868542] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:18.929 [2024-11-27 17:30:49.868614] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:09:18.929 BaseBdev3 00:09:18.929 17:30:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:18.929 17:30:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n raid_bdev1 -s 00:09:18.929 17:30:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:18.929 17:30:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:18.929 [2024-11-27 17:30:49.878097] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:18.929 [2024-11-27 17:30:49.880338] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:18.929 [2024-11-27 17:30:49.880417] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:18.929 [2024-11-27 17:30:49.880603] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:09:18.929 [2024-11-27 17:30:49.880623] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:09:18.929 [2024-11-27 17:30:49.880898] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:09:18.929 [2024-11-27 17:30:49.881034] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:09:18.929 [2024-11-27 17:30:49.881044] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001c80 00:09:18.929 [2024-11-27 17:30:49.881185] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:18.929 17:30:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:18.929 17:30:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online concat 64 3 00:09:18.929 17:30:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:18.929 17:30:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:18.929 17:30:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:18.929 17:30:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:18.929 17:30:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:18.929 17:30:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:18.929 17:30:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:18.929 17:30:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:18.929 17:30:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:18.929 17:30:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:18.929 17:30:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:18.929 17:30:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:18.929 17:30:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:18.929 17:30:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:18.929 17:30:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:18.929 "name": "raid_bdev1", 00:09:18.929 "uuid": "d6200a01-6420-42ad-88d2-6314b04df5a2", 00:09:18.929 "strip_size_kb": 64, 00:09:18.929 "state": "online", 00:09:18.929 "raid_level": "concat", 00:09:18.929 "superblock": true, 00:09:18.929 "num_base_bdevs": 3, 00:09:18.929 "num_base_bdevs_discovered": 3, 00:09:18.929 "num_base_bdevs_operational": 3, 00:09:18.929 "base_bdevs_list": [ 00:09:18.929 { 00:09:18.929 "name": "BaseBdev1", 00:09:18.929 "uuid": "b11eb950-d069-5da7-953a-aa521b9f15ab", 00:09:18.929 "is_configured": true, 00:09:18.929 "data_offset": 2048, 00:09:18.929 "data_size": 63488 00:09:18.929 }, 00:09:18.929 { 00:09:18.929 "name": "BaseBdev2", 00:09:18.930 "uuid": "89f8f163-3129-5399-873c-56f359d2488d", 00:09:18.930 "is_configured": true, 00:09:18.930 "data_offset": 2048, 00:09:18.930 "data_size": 63488 00:09:18.930 }, 00:09:18.930 { 00:09:18.930 "name": "BaseBdev3", 00:09:18.930 "uuid": "e2cf1722-a78b-5dc4-9339-b9700d5df0c2", 00:09:18.930 "is_configured": true, 00:09:18.930 "data_offset": 2048, 00:09:18.930 "data_size": 63488 00:09:18.930 } 00:09:18.930 ] 00:09:18.930 }' 00:09:18.930 17:30:49 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:18.930 17:30:49 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:19.189 17:30:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:09:19.189 17:30:50 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:09:19.189 [2024-11-27 17:30:50.369685] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:09:20.130 17:30:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:09:20.130 17:30:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:20.130 17:30:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:20.130 17:30:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:20.130 17:30:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:09:20.130 17:30:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ concat = \r\a\i\d\1 ]] 00:09:20.130 17:30:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=3 00:09:20.130 17:30:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online concat 64 3 00:09:20.130 17:30:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:20.130 17:30:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:20.130 17:30:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:20.130 17:30:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:20.130 17:30:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:20.130 17:30:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:20.130 17:30:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:20.130 17:30:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:20.130 17:30:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:20.130 17:30:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:20.130 17:30:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:20.130 17:30:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:20.130 17:30:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:20.131 17:30:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:20.391 17:30:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:20.391 "name": "raid_bdev1", 00:09:20.391 "uuid": "d6200a01-6420-42ad-88d2-6314b04df5a2", 00:09:20.391 "strip_size_kb": 64, 00:09:20.391 "state": "online", 00:09:20.391 "raid_level": "concat", 00:09:20.391 "superblock": true, 00:09:20.391 "num_base_bdevs": 3, 00:09:20.391 "num_base_bdevs_discovered": 3, 00:09:20.391 "num_base_bdevs_operational": 3, 00:09:20.391 "base_bdevs_list": [ 00:09:20.391 { 00:09:20.391 "name": "BaseBdev1", 00:09:20.391 "uuid": "b11eb950-d069-5da7-953a-aa521b9f15ab", 00:09:20.391 "is_configured": true, 00:09:20.391 "data_offset": 2048, 00:09:20.391 "data_size": 63488 00:09:20.391 }, 00:09:20.391 { 00:09:20.391 "name": "BaseBdev2", 00:09:20.391 "uuid": "89f8f163-3129-5399-873c-56f359d2488d", 00:09:20.391 "is_configured": true, 00:09:20.391 "data_offset": 2048, 00:09:20.391 "data_size": 63488 00:09:20.391 }, 00:09:20.391 { 00:09:20.391 "name": "BaseBdev3", 00:09:20.391 "uuid": "e2cf1722-a78b-5dc4-9339-b9700d5df0c2", 00:09:20.391 "is_configured": true, 00:09:20.391 "data_offset": 2048, 00:09:20.391 "data_size": 63488 00:09:20.391 } 00:09:20.391 ] 00:09:20.391 }' 00:09:20.391 17:30:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:20.391 17:30:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:20.651 17:30:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:09:20.651 17:30:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:20.651 17:30:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:20.651 [2024-11-27 17:30:51.754364] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:20.651 [2024-11-27 17:30:51.754472] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:20.651 [2024-11-27 17:30:51.757022] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:20.651 [2024-11-27 17:30:51.757126] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:20.651 [2024-11-27 17:30:51.757202] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:20.651 [2024-11-27 17:30:51.757257] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name raid_bdev1, state offline 00:09:20.651 { 00:09:20.651 "results": [ 00:09:20.651 { 00:09:20.651 "job": "raid_bdev1", 00:09:20.651 "core_mask": "0x1", 00:09:20.651 "workload": "randrw", 00:09:20.651 "percentage": 50, 00:09:20.651 "status": "finished", 00:09:20.651 "queue_depth": 1, 00:09:20.651 "io_size": 131072, 00:09:20.651 "runtime": 1.385297, 00:09:20.651 "iops": 14955.637671921617, 00:09:20.651 "mibps": 1869.454708990202, 00:09:20.651 "io_failed": 1, 00:09:20.651 "io_timeout": 0, 00:09:20.651 "avg_latency_us": 93.81425339819515, 00:09:20.651 "min_latency_us": 25.152838427947597, 00:09:20.651 "max_latency_us": 1409.4532751091704 00:09:20.651 } 00:09:20.651 ], 00:09:20.651 "core_count": 1 00:09:20.651 } 00:09:20.651 17:30:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:20.651 17:30:51 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 78023 00:09:20.651 17:30:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@950 -- # '[' -z 78023 ']' 00:09:20.651 17:30:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # kill -0 78023 00:09:20.651 17:30:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@955 -- # uname 00:09:20.651 17:30:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:09:20.651 17:30:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 78023 00:09:20.651 17:30:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:09:20.651 17:30:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:09:20.651 17:30:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 78023' 00:09:20.651 killing process with pid 78023 00:09:20.651 17:30:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@969 -- # kill 78023 00:09:20.651 17:30:51 bdev_raid.raid_read_error_test -- common/autotest_common.sh@974 -- # wait 78023 00:09:20.651 [2024-11-27 17:30:51.804686] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:09:20.912 [2024-11-27 17:30:51.852028] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:09:21.172 17:30:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.cY3Q2y3lMS 00:09:21.172 17:30:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:09:21.172 17:30:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:09:21.172 17:30:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.72 00:09:21.172 17:30:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy concat 00:09:21.172 17:30:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:09:21.172 17:30:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:09:21.172 17:30:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.72 != \0\.\0\0 ]] 00:09:21.172 00:09:21.172 real 0m3.436s 00:09:21.172 user 0m4.144s 00:09:21.172 sys 0m0.653s 00:09:21.172 17:30:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:09:21.172 17:30:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:21.172 ************************************ 00:09:21.172 END TEST raid_read_error_test 00:09:21.172 ************************************ 00:09:21.172 17:30:52 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test concat 3 write 00:09:21.172 17:30:52 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:09:21.172 17:30:52 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:09:21.172 17:30:52 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:09:21.172 ************************************ 00:09:21.172 START TEST raid_write_error_test 00:09:21.172 ************************************ 00:09:21.172 17:30:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1125 -- # raid_io_error_test concat 3 write 00:09:21.172 17:30:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=concat 00:09:21.172 17:30:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=3 00:09:21.172 17:30:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:09:21.172 17:30:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:09:21.172 17:30:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:21.172 17:30:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:09:21.172 17:30:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:21.172 17:30:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:21.172 17:30:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:09:21.172 17:30:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:21.172 17:30:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:21.172 17:30:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:09:21.172 17:30:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:21.172 17:30:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:21.172 17:30:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:09:21.172 17:30:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:09:21.172 17:30:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:09:21.172 17:30:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:09:21.172 17:30:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:09:21.172 17:30:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:09:21.172 17:30:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:09:21.172 17:30:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' concat '!=' raid1 ']' 00:09:21.172 17:30:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:09:21.172 17:30:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:09:21.172 17:30:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:09:21.172 17:30:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.NzlY30qsT2 00:09:21.172 17:30:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=78158 00:09:21.172 17:30:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 78158 00:09:21.172 17:30:52 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:09:21.172 17:30:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@831 -- # '[' -z 78158 ']' 00:09:21.172 17:30:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:21.172 17:30:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:09:21.172 17:30:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:21.172 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:21.172 17:30:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:09:21.173 17:30:52 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:21.433 [2024-11-27 17:30:52.411842] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:09:21.433 [2024-11-27 17:30:52.412060] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid78158 ] 00:09:21.433 [2024-11-27 17:30:52.553438] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:21.433 [2024-11-27 17:30:52.621199] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:09:21.694 [2024-11-27 17:30:52.697200] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:21.694 [2024-11-27 17:30:52.697240] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:22.266 17:30:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:09:22.266 17:30:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@864 -- # return 0 00:09:22.266 17:30:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:22.266 17:30:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:09:22.266 17:30:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:22.266 17:30:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:22.266 BaseBdev1_malloc 00:09:22.266 17:30:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:22.266 17:30:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:09:22.266 17:30:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:22.266 17:30:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:22.266 true 00:09:22.266 17:30:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:22.266 17:30:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:09:22.266 17:30:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:22.266 17:30:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:22.266 [2024-11-27 17:30:53.271328] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:09:22.266 [2024-11-27 17:30:53.271434] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:22.266 [2024-11-27 17:30:53.271466] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:09:22.266 [2024-11-27 17:30:53.271476] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:22.266 [2024-11-27 17:30:53.273914] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:22.266 [2024-11-27 17:30:53.273953] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:09:22.266 BaseBdev1 00:09:22.266 17:30:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:22.266 17:30:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:22.266 17:30:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:09:22.266 17:30:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:22.266 17:30:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:22.266 BaseBdev2_malloc 00:09:22.266 17:30:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:22.266 17:30:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:09:22.266 17:30:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:22.266 17:30:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:22.266 true 00:09:22.266 17:30:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:22.266 17:30:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:09:22.266 17:30:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:22.266 17:30:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:22.266 [2024-11-27 17:30:53.337500] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:09:22.266 [2024-11-27 17:30:53.337636] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:22.266 [2024-11-27 17:30:53.337669] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:09:22.266 [2024-11-27 17:30:53.337680] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:22.266 [2024-11-27 17:30:53.340430] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:22.266 [2024-11-27 17:30:53.340492] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:09:22.266 BaseBdev2 00:09:22.266 17:30:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:22.266 17:30:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:22.266 17:30:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:09:22.266 17:30:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:22.266 17:30:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:22.266 BaseBdev3_malloc 00:09:22.266 17:30:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:22.266 17:30:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:09:22.266 17:30:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:22.266 17:30:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:22.266 true 00:09:22.266 17:30:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:22.266 17:30:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:09:22.266 17:30:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:22.266 17:30:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:22.266 [2024-11-27 17:30:53.383907] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:09:22.266 [2024-11-27 17:30:53.384005] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:22.266 [2024-11-27 17:30:53.384046] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008780 00:09:22.266 [2024-11-27 17:30:53.384055] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:22.266 [2024-11-27 17:30:53.386413] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:22.266 [2024-11-27 17:30:53.386445] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:09:22.266 BaseBdev3 00:09:22.266 17:30:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:22.266 17:30:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n raid_bdev1 -s 00:09:22.266 17:30:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:22.266 17:30:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:22.266 [2024-11-27 17:30:53.395989] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:22.266 [2024-11-27 17:30:53.398061] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:22.266 [2024-11-27 17:30:53.398139] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:22.266 [2024-11-27 17:30:53.398316] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:09:22.266 [2024-11-27 17:30:53.398330] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:09:22.266 [2024-11-27 17:30:53.398566] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:09:22.266 [2024-11-27 17:30:53.398705] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:09:22.266 [2024-11-27 17:30:53.398715] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001c80 00:09:22.266 [2024-11-27 17:30:53.398851] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:22.266 17:30:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:22.266 17:30:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online concat 64 3 00:09:22.266 17:30:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:22.266 17:30:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:22.266 17:30:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:22.266 17:30:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:22.266 17:30:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:22.267 17:30:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:22.267 17:30:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:22.267 17:30:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:22.267 17:30:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:22.267 17:30:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:22.267 17:30:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:22.267 17:30:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:22.267 17:30:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:22.267 17:30:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:22.527 17:30:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:22.527 "name": "raid_bdev1", 00:09:22.527 "uuid": "75e272e1-9eaf-42ba-b36a-66687e1c82cd", 00:09:22.527 "strip_size_kb": 64, 00:09:22.527 "state": "online", 00:09:22.527 "raid_level": "concat", 00:09:22.527 "superblock": true, 00:09:22.527 "num_base_bdevs": 3, 00:09:22.527 "num_base_bdevs_discovered": 3, 00:09:22.527 "num_base_bdevs_operational": 3, 00:09:22.527 "base_bdevs_list": [ 00:09:22.527 { 00:09:22.527 "name": "BaseBdev1", 00:09:22.527 "uuid": "10e68b39-f56b-50b5-98aa-f5eedb73c675", 00:09:22.527 "is_configured": true, 00:09:22.527 "data_offset": 2048, 00:09:22.527 "data_size": 63488 00:09:22.527 }, 00:09:22.527 { 00:09:22.527 "name": "BaseBdev2", 00:09:22.527 "uuid": "1c0794c8-7d39-575b-b988-4681183e05b4", 00:09:22.527 "is_configured": true, 00:09:22.527 "data_offset": 2048, 00:09:22.527 "data_size": 63488 00:09:22.527 }, 00:09:22.527 { 00:09:22.527 "name": "BaseBdev3", 00:09:22.527 "uuid": "eefde5e3-f49c-50fa-8313-bff6273ebd57", 00:09:22.527 "is_configured": true, 00:09:22.527 "data_offset": 2048, 00:09:22.527 "data_size": 63488 00:09:22.527 } 00:09:22.527 ] 00:09:22.527 }' 00:09:22.527 17:30:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:22.527 17:30:53 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:22.787 17:30:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:09:22.787 17:30:53 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:09:22.787 [2024-11-27 17:30:53.943535] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:09:23.726 17:30:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:09:23.727 17:30:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:23.727 17:30:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:23.727 17:30:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:23.727 17:30:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:09:23.727 17:30:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ concat = \r\a\i\d\1 ]] 00:09:23.727 17:30:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=3 00:09:23.727 17:30:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online concat 64 3 00:09:23.727 17:30:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:23.727 17:30:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:23.727 17:30:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:09:23.727 17:30:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:23.727 17:30:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:23.727 17:30:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:23.727 17:30:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:23.727 17:30:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:23.727 17:30:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:23.727 17:30:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:23.727 17:30:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:23.727 17:30:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:23.727 17:30:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:23.727 17:30:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:23.727 17:30:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:23.727 "name": "raid_bdev1", 00:09:23.727 "uuid": "75e272e1-9eaf-42ba-b36a-66687e1c82cd", 00:09:23.727 "strip_size_kb": 64, 00:09:23.727 "state": "online", 00:09:23.727 "raid_level": "concat", 00:09:23.727 "superblock": true, 00:09:23.727 "num_base_bdevs": 3, 00:09:23.727 "num_base_bdevs_discovered": 3, 00:09:23.727 "num_base_bdevs_operational": 3, 00:09:23.727 "base_bdevs_list": [ 00:09:23.727 { 00:09:23.727 "name": "BaseBdev1", 00:09:23.727 "uuid": "10e68b39-f56b-50b5-98aa-f5eedb73c675", 00:09:23.727 "is_configured": true, 00:09:23.727 "data_offset": 2048, 00:09:23.727 "data_size": 63488 00:09:23.727 }, 00:09:23.727 { 00:09:23.727 "name": "BaseBdev2", 00:09:23.727 "uuid": "1c0794c8-7d39-575b-b988-4681183e05b4", 00:09:23.727 "is_configured": true, 00:09:23.727 "data_offset": 2048, 00:09:23.727 "data_size": 63488 00:09:23.727 }, 00:09:23.727 { 00:09:23.727 "name": "BaseBdev3", 00:09:23.727 "uuid": "eefde5e3-f49c-50fa-8313-bff6273ebd57", 00:09:23.727 "is_configured": true, 00:09:23.727 "data_offset": 2048, 00:09:23.727 "data_size": 63488 00:09:23.727 } 00:09:23.727 ] 00:09:23.727 }' 00:09:23.727 17:30:54 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:23.727 17:30:54 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:24.297 17:30:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:09:24.297 17:30:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:24.297 17:30:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:24.297 [2024-11-27 17:30:55.336005] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:24.297 [2024-11-27 17:30:55.336188] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:24.297 [2024-11-27 17:30:55.338756] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:24.297 [2024-11-27 17:30:55.338850] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:24.297 [2024-11-27 17:30:55.338907] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:24.297 [2024-11-27 17:30:55.339004] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name raid_bdev1, state offline 00:09:24.297 { 00:09:24.297 "results": [ 00:09:24.297 { 00:09:24.297 "job": "raid_bdev1", 00:09:24.297 "core_mask": "0x1", 00:09:24.297 "workload": "randrw", 00:09:24.297 "percentage": 50, 00:09:24.297 "status": "finished", 00:09:24.297 "queue_depth": 1, 00:09:24.297 "io_size": 131072, 00:09:24.297 "runtime": 1.393162, 00:09:24.297 "iops": 14973.850851516192, 00:09:24.297 "mibps": 1871.731356439524, 00:09:24.297 "io_failed": 1, 00:09:24.297 "io_timeout": 0, 00:09:24.297 "avg_latency_us": 93.63117378958168, 00:09:24.297 "min_latency_us": 25.041048034934498, 00:09:24.298 "max_latency_us": 1395.1441048034935 00:09:24.298 } 00:09:24.298 ], 00:09:24.298 "core_count": 1 00:09:24.298 } 00:09:24.298 17:30:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:24.298 17:30:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 78158 00:09:24.298 17:30:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@950 -- # '[' -z 78158 ']' 00:09:24.298 17:30:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # kill -0 78158 00:09:24.298 17:30:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@955 -- # uname 00:09:24.298 17:30:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:09:24.298 17:30:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 78158 00:09:24.298 17:30:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:09:24.298 killing process with pid 78158 00:09:24.298 17:30:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:09:24.298 17:30:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 78158' 00:09:24.298 17:30:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@969 -- # kill 78158 00:09:24.298 [2024-11-27 17:30:55.388573] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:09:24.298 17:30:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@974 -- # wait 78158 00:09:24.298 [2024-11-27 17:30:55.435561] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:09:24.869 17:30:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.NzlY30qsT2 00:09:24.869 17:30:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:09:24.869 17:30:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:09:24.869 ************************************ 00:09:24.869 END TEST raid_write_error_test 00:09:24.869 ************************************ 00:09:24.869 17:30:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.72 00:09:24.869 17:30:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy concat 00:09:24.869 17:30:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:09:24.869 17:30:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:09:24.869 17:30:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.72 != \0\.\0\0 ]] 00:09:24.869 00:09:24.869 real 0m3.507s 00:09:24.869 user 0m4.309s 00:09:24.869 sys 0m0.630s 00:09:24.869 17:30:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:09:24.869 17:30:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:24.869 17:30:55 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:09:24.869 17:30:55 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test raid1 3 false 00:09:24.869 17:30:55 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:09:24.869 17:30:55 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:09:24.869 17:30:55 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:09:24.869 ************************************ 00:09:24.869 START TEST raid_state_function_test 00:09:24.869 ************************************ 00:09:24.869 17:30:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1125 -- # raid_state_function_test raid1 3 false 00:09:24.869 17:30:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:09:24.869 17:30:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=3 00:09:24.869 17:30:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:09:24.869 17:30:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:09:24.869 17:30:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:09:24.869 17:30:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:24.869 17:30:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:09:24.869 17:30:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:24.869 17:30:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:24.869 17:30:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:09:24.869 17:30:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:24.869 17:30:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:24.869 17:30:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:09:24.869 17:30:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:24.869 17:30:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:24.869 17:30:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:09:24.869 17:30:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:09:24.869 17:30:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:09:24.869 17:30:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:09:24.869 17:30:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:09:24.869 17:30:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:09:24.869 17:30:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:09:24.869 17:30:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:09:24.869 17:30:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:09:24.869 17:30:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:09:24.869 Process raid pid: 78285 00:09:24.869 17:30:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=78285 00:09:24.869 17:30:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:09:24.869 17:30:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 78285' 00:09:24.869 17:30:55 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 78285 00:09:24.869 17:30:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@831 -- # '[' -z 78285 ']' 00:09:24.869 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:24.869 17:30:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:24.869 17:30:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:09:24.869 17:30:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:24.869 17:30:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:09:24.869 17:30:55 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:24.869 [2024-11-27 17:30:55.986502] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:09:24.869 [2024-11-27 17:30:55.986647] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:09:25.129 [2024-11-27 17:30:56.133710] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:25.130 [2024-11-27 17:30:56.201618] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:09:25.130 [2024-11-27 17:30:56.277732] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:25.130 [2024-11-27 17:30:56.277868] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:25.700 17:30:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:09:25.700 17:30:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@864 -- # return 0 00:09:25.700 17:30:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:09:25.700 17:30:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:25.700 17:30:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:25.700 [2024-11-27 17:30:56.817183] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:25.700 [2024-11-27 17:30:56.817348] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:25.700 [2024-11-27 17:30:56.817366] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:25.700 [2024-11-27 17:30:56.817375] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:25.700 [2024-11-27 17:30:56.817382] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:25.700 [2024-11-27 17:30:56.817394] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:25.700 17:30:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:25.700 17:30:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:25.700 17:30:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:25.700 17:30:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:25.700 17:30:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:25.700 17:30:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:25.700 17:30:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:25.700 17:30:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:25.700 17:30:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:25.700 17:30:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:25.700 17:30:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:25.700 17:30:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:25.700 17:30:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:25.700 17:30:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:25.700 17:30:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:25.700 17:30:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:25.700 17:30:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:25.700 "name": "Existed_Raid", 00:09:25.700 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:25.700 "strip_size_kb": 0, 00:09:25.700 "state": "configuring", 00:09:25.700 "raid_level": "raid1", 00:09:25.700 "superblock": false, 00:09:25.700 "num_base_bdevs": 3, 00:09:25.700 "num_base_bdevs_discovered": 0, 00:09:25.700 "num_base_bdevs_operational": 3, 00:09:25.700 "base_bdevs_list": [ 00:09:25.700 { 00:09:25.700 "name": "BaseBdev1", 00:09:25.700 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:25.700 "is_configured": false, 00:09:25.700 "data_offset": 0, 00:09:25.700 "data_size": 0 00:09:25.700 }, 00:09:25.700 { 00:09:25.700 "name": "BaseBdev2", 00:09:25.700 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:25.700 "is_configured": false, 00:09:25.700 "data_offset": 0, 00:09:25.700 "data_size": 0 00:09:25.700 }, 00:09:25.700 { 00:09:25.700 "name": "BaseBdev3", 00:09:25.700 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:25.700 "is_configured": false, 00:09:25.700 "data_offset": 0, 00:09:25.700 "data_size": 0 00:09:25.700 } 00:09:25.700 ] 00:09:25.700 }' 00:09:25.700 17:30:56 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:25.700 17:30:56 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:26.271 17:30:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:26.271 17:30:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:26.271 17:30:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:26.271 [2024-11-27 17:30:57.236369] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:26.271 [2024-11-27 17:30:57.236495] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:09:26.271 17:30:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:26.271 17:30:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:09:26.271 17:30:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:26.271 17:30:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:26.271 [2024-11-27 17:30:57.248342] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:26.271 [2024-11-27 17:30:57.248421] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:26.271 [2024-11-27 17:30:57.248448] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:26.271 [2024-11-27 17:30:57.248470] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:26.271 [2024-11-27 17:30:57.248487] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:26.271 [2024-11-27 17:30:57.248508] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:26.271 17:30:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:26.271 17:30:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:09:26.271 17:30:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:26.271 17:30:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:26.271 [2024-11-27 17:30:57.275362] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:26.271 BaseBdev1 00:09:26.271 17:30:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:26.271 17:30:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:09:26.271 17:30:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:09:26.271 17:30:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:09:26.271 17:30:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:09:26.271 17:30:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:09:26.271 17:30:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:09:26.271 17:30:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:09:26.271 17:30:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:26.271 17:30:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:26.271 17:30:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:26.271 17:30:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:09:26.271 17:30:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:26.271 17:30:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:26.271 [ 00:09:26.271 { 00:09:26.271 "name": "BaseBdev1", 00:09:26.271 "aliases": [ 00:09:26.271 "b88b1c4a-19e2-494b-8bd6-302fc271acde" 00:09:26.271 ], 00:09:26.271 "product_name": "Malloc disk", 00:09:26.271 "block_size": 512, 00:09:26.271 "num_blocks": 65536, 00:09:26.271 "uuid": "b88b1c4a-19e2-494b-8bd6-302fc271acde", 00:09:26.271 "assigned_rate_limits": { 00:09:26.271 "rw_ios_per_sec": 0, 00:09:26.271 "rw_mbytes_per_sec": 0, 00:09:26.271 "r_mbytes_per_sec": 0, 00:09:26.271 "w_mbytes_per_sec": 0 00:09:26.271 }, 00:09:26.271 "claimed": true, 00:09:26.271 "claim_type": "exclusive_write", 00:09:26.271 "zoned": false, 00:09:26.271 "supported_io_types": { 00:09:26.271 "read": true, 00:09:26.271 "write": true, 00:09:26.271 "unmap": true, 00:09:26.271 "flush": true, 00:09:26.271 "reset": true, 00:09:26.271 "nvme_admin": false, 00:09:26.271 "nvme_io": false, 00:09:26.271 "nvme_io_md": false, 00:09:26.271 "write_zeroes": true, 00:09:26.271 "zcopy": true, 00:09:26.271 "get_zone_info": false, 00:09:26.271 "zone_management": false, 00:09:26.271 "zone_append": false, 00:09:26.271 "compare": false, 00:09:26.271 "compare_and_write": false, 00:09:26.271 "abort": true, 00:09:26.271 "seek_hole": false, 00:09:26.271 "seek_data": false, 00:09:26.271 "copy": true, 00:09:26.271 "nvme_iov_md": false 00:09:26.271 }, 00:09:26.271 "memory_domains": [ 00:09:26.271 { 00:09:26.271 "dma_device_id": "system", 00:09:26.271 "dma_device_type": 1 00:09:26.271 }, 00:09:26.271 { 00:09:26.271 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:26.271 "dma_device_type": 2 00:09:26.271 } 00:09:26.271 ], 00:09:26.271 "driver_specific": {} 00:09:26.271 } 00:09:26.271 ] 00:09:26.271 17:30:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:26.271 17:30:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:09:26.271 17:30:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:26.271 17:30:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:26.271 17:30:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:26.271 17:30:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:26.271 17:30:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:26.271 17:30:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:26.271 17:30:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:26.271 17:30:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:26.271 17:30:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:26.271 17:30:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:26.271 17:30:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:26.271 17:30:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:26.271 17:30:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:26.271 17:30:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:26.271 17:30:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:26.271 17:30:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:26.271 "name": "Existed_Raid", 00:09:26.271 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:26.271 "strip_size_kb": 0, 00:09:26.271 "state": "configuring", 00:09:26.271 "raid_level": "raid1", 00:09:26.271 "superblock": false, 00:09:26.271 "num_base_bdevs": 3, 00:09:26.271 "num_base_bdevs_discovered": 1, 00:09:26.271 "num_base_bdevs_operational": 3, 00:09:26.271 "base_bdevs_list": [ 00:09:26.271 { 00:09:26.271 "name": "BaseBdev1", 00:09:26.271 "uuid": "b88b1c4a-19e2-494b-8bd6-302fc271acde", 00:09:26.271 "is_configured": true, 00:09:26.271 "data_offset": 0, 00:09:26.271 "data_size": 65536 00:09:26.271 }, 00:09:26.271 { 00:09:26.271 "name": "BaseBdev2", 00:09:26.271 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:26.271 "is_configured": false, 00:09:26.271 "data_offset": 0, 00:09:26.271 "data_size": 0 00:09:26.271 }, 00:09:26.271 { 00:09:26.271 "name": "BaseBdev3", 00:09:26.271 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:26.271 "is_configured": false, 00:09:26.271 "data_offset": 0, 00:09:26.271 "data_size": 0 00:09:26.271 } 00:09:26.271 ] 00:09:26.271 }' 00:09:26.271 17:30:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:26.271 17:30:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:26.841 17:30:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:26.841 17:30:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:26.841 17:30:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:26.841 [2024-11-27 17:30:57.746597] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:26.841 [2024-11-27 17:30:57.746650] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:09:26.841 17:30:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:26.841 17:30:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:09:26.841 17:30:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:26.841 17:30:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:26.841 [2024-11-27 17:30:57.758631] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:26.841 [2024-11-27 17:30:57.760851] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:26.841 [2024-11-27 17:30:57.760925] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:26.841 [2024-11-27 17:30:57.760969] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:26.841 [2024-11-27 17:30:57.760993] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:26.841 17:30:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:26.841 17:30:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:09:26.841 17:30:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:26.841 17:30:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:26.841 17:30:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:26.841 17:30:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:26.841 17:30:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:26.841 17:30:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:26.841 17:30:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:26.841 17:30:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:26.841 17:30:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:26.841 17:30:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:26.841 17:30:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:26.841 17:30:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:26.841 17:30:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:26.841 17:30:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:26.841 17:30:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:26.841 17:30:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:26.841 17:30:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:26.841 "name": "Existed_Raid", 00:09:26.841 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:26.841 "strip_size_kb": 0, 00:09:26.841 "state": "configuring", 00:09:26.841 "raid_level": "raid1", 00:09:26.841 "superblock": false, 00:09:26.841 "num_base_bdevs": 3, 00:09:26.841 "num_base_bdevs_discovered": 1, 00:09:26.841 "num_base_bdevs_operational": 3, 00:09:26.841 "base_bdevs_list": [ 00:09:26.841 { 00:09:26.841 "name": "BaseBdev1", 00:09:26.841 "uuid": "b88b1c4a-19e2-494b-8bd6-302fc271acde", 00:09:26.841 "is_configured": true, 00:09:26.841 "data_offset": 0, 00:09:26.841 "data_size": 65536 00:09:26.841 }, 00:09:26.841 { 00:09:26.841 "name": "BaseBdev2", 00:09:26.841 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:26.841 "is_configured": false, 00:09:26.841 "data_offset": 0, 00:09:26.841 "data_size": 0 00:09:26.841 }, 00:09:26.841 { 00:09:26.841 "name": "BaseBdev3", 00:09:26.841 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:26.841 "is_configured": false, 00:09:26.841 "data_offset": 0, 00:09:26.841 "data_size": 0 00:09:26.841 } 00:09:26.841 ] 00:09:26.841 }' 00:09:26.841 17:30:57 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:26.841 17:30:57 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:27.102 17:30:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:09:27.102 17:30:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:27.102 17:30:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:27.102 [2024-11-27 17:30:58.234514] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:27.102 BaseBdev2 00:09:27.102 17:30:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:27.102 17:30:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:09:27.102 17:30:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:09:27.102 17:30:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:09:27.102 17:30:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:09:27.102 17:30:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:09:27.102 17:30:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:09:27.102 17:30:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:09:27.102 17:30:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:27.102 17:30:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:27.102 17:30:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:27.102 17:30:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:09:27.102 17:30:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:27.102 17:30:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:27.102 [ 00:09:27.102 { 00:09:27.102 "name": "BaseBdev2", 00:09:27.102 "aliases": [ 00:09:27.102 "1a54bc5d-9f6a-471a-b77a-fa944a9b3203" 00:09:27.102 ], 00:09:27.102 "product_name": "Malloc disk", 00:09:27.102 "block_size": 512, 00:09:27.102 "num_blocks": 65536, 00:09:27.102 "uuid": "1a54bc5d-9f6a-471a-b77a-fa944a9b3203", 00:09:27.102 "assigned_rate_limits": { 00:09:27.102 "rw_ios_per_sec": 0, 00:09:27.102 "rw_mbytes_per_sec": 0, 00:09:27.102 "r_mbytes_per_sec": 0, 00:09:27.102 "w_mbytes_per_sec": 0 00:09:27.102 }, 00:09:27.102 "claimed": true, 00:09:27.102 "claim_type": "exclusive_write", 00:09:27.102 "zoned": false, 00:09:27.102 "supported_io_types": { 00:09:27.102 "read": true, 00:09:27.102 "write": true, 00:09:27.102 "unmap": true, 00:09:27.102 "flush": true, 00:09:27.102 "reset": true, 00:09:27.102 "nvme_admin": false, 00:09:27.102 "nvme_io": false, 00:09:27.102 "nvme_io_md": false, 00:09:27.102 "write_zeroes": true, 00:09:27.102 "zcopy": true, 00:09:27.102 "get_zone_info": false, 00:09:27.102 "zone_management": false, 00:09:27.102 "zone_append": false, 00:09:27.102 "compare": false, 00:09:27.102 "compare_and_write": false, 00:09:27.102 "abort": true, 00:09:27.102 "seek_hole": false, 00:09:27.102 "seek_data": false, 00:09:27.102 "copy": true, 00:09:27.102 "nvme_iov_md": false 00:09:27.102 }, 00:09:27.102 "memory_domains": [ 00:09:27.102 { 00:09:27.102 "dma_device_id": "system", 00:09:27.102 "dma_device_type": 1 00:09:27.102 }, 00:09:27.102 { 00:09:27.102 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:27.102 "dma_device_type": 2 00:09:27.102 } 00:09:27.102 ], 00:09:27.102 "driver_specific": {} 00:09:27.102 } 00:09:27.102 ] 00:09:27.102 17:30:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:27.102 17:30:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:09:27.102 17:30:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:09:27.102 17:30:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:27.102 17:30:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:27.102 17:30:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:27.102 17:30:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:27.102 17:30:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:27.102 17:30:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:27.102 17:30:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:27.102 17:30:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:27.102 17:30:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:27.102 17:30:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:27.102 17:30:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:27.102 17:30:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:27.102 17:30:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:27.102 17:30:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:27.102 17:30:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:27.362 17:30:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:27.362 17:30:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:27.362 "name": "Existed_Raid", 00:09:27.362 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:27.362 "strip_size_kb": 0, 00:09:27.362 "state": "configuring", 00:09:27.362 "raid_level": "raid1", 00:09:27.362 "superblock": false, 00:09:27.362 "num_base_bdevs": 3, 00:09:27.362 "num_base_bdevs_discovered": 2, 00:09:27.362 "num_base_bdevs_operational": 3, 00:09:27.362 "base_bdevs_list": [ 00:09:27.362 { 00:09:27.362 "name": "BaseBdev1", 00:09:27.362 "uuid": "b88b1c4a-19e2-494b-8bd6-302fc271acde", 00:09:27.362 "is_configured": true, 00:09:27.362 "data_offset": 0, 00:09:27.362 "data_size": 65536 00:09:27.362 }, 00:09:27.362 { 00:09:27.362 "name": "BaseBdev2", 00:09:27.362 "uuid": "1a54bc5d-9f6a-471a-b77a-fa944a9b3203", 00:09:27.362 "is_configured": true, 00:09:27.362 "data_offset": 0, 00:09:27.362 "data_size": 65536 00:09:27.362 }, 00:09:27.362 { 00:09:27.362 "name": "BaseBdev3", 00:09:27.362 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:27.362 "is_configured": false, 00:09:27.362 "data_offset": 0, 00:09:27.362 "data_size": 0 00:09:27.362 } 00:09:27.362 ] 00:09:27.362 }' 00:09:27.362 17:30:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:27.362 17:30:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:27.622 17:30:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:09:27.622 17:30:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:27.622 17:30:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:27.622 [2024-11-27 17:30:58.678518] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:27.622 [2024-11-27 17:30:58.678646] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:09:27.622 [2024-11-27 17:30:58.678663] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:09:27.622 [2024-11-27 17:30:58.678992] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:09:27.622 [2024-11-27 17:30:58.679167] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:09:27.622 [2024-11-27 17:30:58.679179] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:09:27.622 [2024-11-27 17:30:58.679405] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:27.622 BaseBdev3 00:09:27.622 17:30:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:27.622 17:30:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:09:27.622 17:30:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:09:27.622 17:30:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:09:27.622 17:30:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:09:27.622 17:30:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:09:27.622 17:30:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:09:27.622 17:30:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:09:27.622 17:30:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:27.622 17:30:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:27.622 17:30:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:27.622 17:30:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:09:27.622 17:30:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:27.622 17:30:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:27.622 [ 00:09:27.622 { 00:09:27.622 "name": "BaseBdev3", 00:09:27.622 "aliases": [ 00:09:27.622 "bade2ed0-85c9-427a-9670-e26b93750480" 00:09:27.622 ], 00:09:27.622 "product_name": "Malloc disk", 00:09:27.622 "block_size": 512, 00:09:27.622 "num_blocks": 65536, 00:09:27.622 "uuid": "bade2ed0-85c9-427a-9670-e26b93750480", 00:09:27.622 "assigned_rate_limits": { 00:09:27.622 "rw_ios_per_sec": 0, 00:09:27.622 "rw_mbytes_per_sec": 0, 00:09:27.622 "r_mbytes_per_sec": 0, 00:09:27.622 "w_mbytes_per_sec": 0 00:09:27.622 }, 00:09:27.622 "claimed": true, 00:09:27.622 "claim_type": "exclusive_write", 00:09:27.622 "zoned": false, 00:09:27.622 "supported_io_types": { 00:09:27.622 "read": true, 00:09:27.622 "write": true, 00:09:27.622 "unmap": true, 00:09:27.622 "flush": true, 00:09:27.622 "reset": true, 00:09:27.622 "nvme_admin": false, 00:09:27.622 "nvme_io": false, 00:09:27.622 "nvme_io_md": false, 00:09:27.622 "write_zeroes": true, 00:09:27.622 "zcopy": true, 00:09:27.622 "get_zone_info": false, 00:09:27.622 "zone_management": false, 00:09:27.622 "zone_append": false, 00:09:27.622 "compare": false, 00:09:27.622 "compare_and_write": false, 00:09:27.622 "abort": true, 00:09:27.622 "seek_hole": false, 00:09:27.622 "seek_data": false, 00:09:27.622 "copy": true, 00:09:27.622 "nvme_iov_md": false 00:09:27.622 }, 00:09:27.622 "memory_domains": [ 00:09:27.622 { 00:09:27.622 "dma_device_id": "system", 00:09:27.622 "dma_device_type": 1 00:09:27.622 }, 00:09:27.622 { 00:09:27.622 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:27.622 "dma_device_type": 2 00:09:27.622 } 00:09:27.622 ], 00:09:27.622 "driver_specific": {} 00:09:27.622 } 00:09:27.622 ] 00:09:27.622 17:30:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:27.622 17:30:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:09:27.622 17:30:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:09:27.622 17:30:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:27.622 17:30:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 3 00:09:27.622 17:30:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:27.622 17:30:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:27.622 17:30:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:27.622 17:30:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:27.622 17:30:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:27.622 17:30:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:27.622 17:30:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:27.622 17:30:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:27.622 17:30:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:27.622 17:30:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:27.622 17:30:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:27.622 17:30:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:27.622 17:30:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:27.622 17:30:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:27.622 17:30:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:27.622 "name": "Existed_Raid", 00:09:27.622 "uuid": "e86d0380-1e57-4718-9ad7-25eeb44274f5", 00:09:27.622 "strip_size_kb": 0, 00:09:27.622 "state": "online", 00:09:27.622 "raid_level": "raid1", 00:09:27.622 "superblock": false, 00:09:27.622 "num_base_bdevs": 3, 00:09:27.622 "num_base_bdevs_discovered": 3, 00:09:27.622 "num_base_bdevs_operational": 3, 00:09:27.622 "base_bdevs_list": [ 00:09:27.622 { 00:09:27.622 "name": "BaseBdev1", 00:09:27.622 "uuid": "b88b1c4a-19e2-494b-8bd6-302fc271acde", 00:09:27.622 "is_configured": true, 00:09:27.622 "data_offset": 0, 00:09:27.622 "data_size": 65536 00:09:27.622 }, 00:09:27.622 { 00:09:27.622 "name": "BaseBdev2", 00:09:27.622 "uuid": "1a54bc5d-9f6a-471a-b77a-fa944a9b3203", 00:09:27.622 "is_configured": true, 00:09:27.622 "data_offset": 0, 00:09:27.622 "data_size": 65536 00:09:27.622 }, 00:09:27.622 { 00:09:27.622 "name": "BaseBdev3", 00:09:27.622 "uuid": "bade2ed0-85c9-427a-9670-e26b93750480", 00:09:27.622 "is_configured": true, 00:09:27.622 "data_offset": 0, 00:09:27.622 "data_size": 65536 00:09:27.623 } 00:09:27.623 ] 00:09:27.623 }' 00:09:27.623 17:30:58 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:27.623 17:30:58 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:28.190 17:30:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:09:28.190 17:30:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:09:28.190 17:30:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:28.190 17:30:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:28.190 17:30:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:09:28.190 17:30:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:28.191 17:30:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:09:28.191 17:30:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:28.191 17:30:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:28.191 17:30:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:28.191 [2024-11-27 17:30:59.142054] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:28.191 17:30:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:28.191 17:30:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:28.191 "name": "Existed_Raid", 00:09:28.191 "aliases": [ 00:09:28.191 "e86d0380-1e57-4718-9ad7-25eeb44274f5" 00:09:28.191 ], 00:09:28.191 "product_name": "Raid Volume", 00:09:28.191 "block_size": 512, 00:09:28.191 "num_blocks": 65536, 00:09:28.191 "uuid": "e86d0380-1e57-4718-9ad7-25eeb44274f5", 00:09:28.191 "assigned_rate_limits": { 00:09:28.191 "rw_ios_per_sec": 0, 00:09:28.191 "rw_mbytes_per_sec": 0, 00:09:28.191 "r_mbytes_per_sec": 0, 00:09:28.191 "w_mbytes_per_sec": 0 00:09:28.191 }, 00:09:28.191 "claimed": false, 00:09:28.191 "zoned": false, 00:09:28.191 "supported_io_types": { 00:09:28.191 "read": true, 00:09:28.191 "write": true, 00:09:28.191 "unmap": false, 00:09:28.191 "flush": false, 00:09:28.191 "reset": true, 00:09:28.191 "nvme_admin": false, 00:09:28.191 "nvme_io": false, 00:09:28.191 "nvme_io_md": false, 00:09:28.191 "write_zeroes": true, 00:09:28.191 "zcopy": false, 00:09:28.191 "get_zone_info": false, 00:09:28.191 "zone_management": false, 00:09:28.191 "zone_append": false, 00:09:28.191 "compare": false, 00:09:28.191 "compare_and_write": false, 00:09:28.191 "abort": false, 00:09:28.191 "seek_hole": false, 00:09:28.191 "seek_data": false, 00:09:28.191 "copy": false, 00:09:28.191 "nvme_iov_md": false 00:09:28.191 }, 00:09:28.191 "memory_domains": [ 00:09:28.191 { 00:09:28.191 "dma_device_id": "system", 00:09:28.191 "dma_device_type": 1 00:09:28.191 }, 00:09:28.191 { 00:09:28.191 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:28.191 "dma_device_type": 2 00:09:28.191 }, 00:09:28.191 { 00:09:28.191 "dma_device_id": "system", 00:09:28.191 "dma_device_type": 1 00:09:28.191 }, 00:09:28.191 { 00:09:28.191 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:28.191 "dma_device_type": 2 00:09:28.191 }, 00:09:28.191 { 00:09:28.191 "dma_device_id": "system", 00:09:28.191 "dma_device_type": 1 00:09:28.191 }, 00:09:28.191 { 00:09:28.191 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:28.191 "dma_device_type": 2 00:09:28.191 } 00:09:28.191 ], 00:09:28.191 "driver_specific": { 00:09:28.191 "raid": { 00:09:28.191 "uuid": "e86d0380-1e57-4718-9ad7-25eeb44274f5", 00:09:28.191 "strip_size_kb": 0, 00:09:28.191 "state": "online", 00:09:28.191 "raid_level": "raid1", 00:09:28.191 "superblock": false, 00:09:28.191 "num_base_bdevs": 3, 00:09:28.191 "num_base_bdevs_discovered": 3, 00:09:28.191 "num_base_bdevs_operational": 3, 00:09:28.191 "base_bdevs_list": [ 00:09:28.191 { 00:09:28.191 "name": "BaseBdev1", 00:09:28.191 "uuid": "b88b1c4a-19e2-494b-8bd6-302fc271acde", 00:09:28.191 "is_configured": true, 00:09:28.191 "data_offset": 0, 00:09:28.191 "data_size": 65536 00:09:28.191 }, 00:09:28.191 { 00:09:28.191 "name": "BaseBdev2", 00:09:28.191 "uuid": "1a54bc5d-9f6a-471a-b77a-fa944a9b3203", 00:09:28.191 "is_configured": true, 00:09:28.191 "data_offset": 0, 00:09:28.191 "data_size": 65536 00:09:28.191 }, 00:09:28.191 { 00:09:28.191 "name": "BaseBdev3", 00:09:28.191 "uuid": "bade2ed0-85c9-427a-9670-e26b93750480", 00:09:28.191 "is_configured": true, 00:09:28.191 "data_offset": 0, 00:09:28.191 "data_size": 65536 00:09:28.191 } 00:09:28.191 ] 00:09:28.191 } 00:09:28.191 } 00:09:28.191 }' 00:09:28.191 17:30:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:28.191 17:30:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:09:28.191 BaseBdev2 00:09:28.191 BaseBdev3' 00:09:28.191 17:30:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:28.191 17:30:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:28.191 17:30:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:28.191 17:30:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:09:28.191 17:30:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:28.191 17:30:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:28.191 17:30:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:28.191 17:30:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:28.191 17:30:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:28.191 17:30:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:28.191 17:30:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:28.191 17:30:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:09:28.191 17:30:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:28.191 17:30:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:28.191 17:30:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:28.191 17:30:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:28.191 17:30:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:28.191 17:30:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:28.191 17:30:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:28.191 17:30:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:28.191 17:30:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:09:28.191 17:30:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:28.191 17:30:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:28.191 17:30:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:28.450 17:30:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:28.450 17:30:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:28.450 17:30:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:09:28.450 17:30:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:28.450 17:30:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:28.450 [2024-11-27 17:30:59.413373] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:09:28.450 17:30:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:28.450 17:30:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:09:28.450 17:30:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:09:28.451 17:30:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:09:28.451 17:30:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@199 -- # return 0 00:09:28.451 17:30:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:09:28.451 17:30:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 2 00:09:28.451 17:30:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:28.451 17:30:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:28.451 17:30:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:28.451 17:30:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:28.451 17:30:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:09:28.451 17:30:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:28.451 17:30:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:28.451 17:30:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:28.451 17:30:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:28.451 17:30:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:28.451 17:30:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:28.451 17:30:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:28.451 17:30:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:28.451 17:30:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:28.451 17:30:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:28.451 "name": "Existed_Raid", 00:09:28.451 "uuid": "e86d0380-1e57-4718-9ad7-25eeb44274f5", 00:09:28.451 "strip_size_kb": 0, 00:09:28.451 "state": "online", 00:09:28.451 "raid_level": "raid1", 00:09:28.451 "superblock": false, 00:09:28.451 "num_base_bdevs": 3, 00:09:28.451 "num_base_bdevs_discovered": 2, 00:09:28.451 "num_base_bdevs_operational": 2, 00:09:28.451 "base_bdevs_list": [ 00:09:28.451 { 00:09:28.451 "name": null, 00:09:28.451 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:28.451 "is_configured": false, 00:09:28.451 "data_offset": 0, 00:09:28.451 "data_size": 65536 00:09:28.451 }, 00:09:28.451 { 00:09:28.451 "name": "BaseBdev2", 00:09:28.451 "uuid": "1a54bc5d-9f6a-471a-b77a-fa944a9b3203", 00:09:28.451 "is_configured": true, 00:09:28.451 "data_offset": 0, 00:09:28.451 "data_size": 65536 00:09:28.451 }, 00:09:28.451 { 00:09:28.451 "name": "BaseBdev3", 00:09:28.451 "uuid": "bade2ed0-85c9-427a-9670-e26b93750480", 00:09:28.451 "is_configured": true, 00:09:28.451 "data_offset": 0, 00:09:28.451 "data_size": 65536 00:09:28.451 } 00:09:28.451 ] 00:09:28.451 }' 00:09:28.451 17:30:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:28.451 17:30:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:28.715 17:30:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:09:28.715 17:30:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:28.715 17:30:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:09:28.715 17:30:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:28.715 17:30:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:28.715 17:30:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:28.993 17:30:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:28.993 17:30:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:09:28.993 17:30:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:09:28.993 17:30:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:09:28.993 17:30:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:28.993 17:30:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:28.993 [2024-11-27 17:30:59.953192] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:09:28.993 17:30:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:28.993 17:30:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:09:28.993 17:30:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:28.993 17:30:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:28.993 17:30:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:28.993 17:30:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:09:28.993 17:30:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:28.993 17:30:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:28.993 17:31:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:09:28.993 17:31:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:09:28.993 17:31:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:09:28.993 17:31:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:28.993 17:31:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:28.993 [2024-11-27 17:31:00.033750] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:09:28.993 [2024-11-27 17:31:00.033855] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:28.993 [2024-11-27 17:31:00.053809] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:28.993 [2024-11-27 17:31:00.053862] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:28.993 [2024-11-27 17:31:00.053878] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:09:28.993 17:31:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:28.993 17:31:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:09:28.993 17:31:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:28.993 17:31:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:28.993 17:31:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:28.993 17:31:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:28.993 17:31:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:09:28.993 17:31:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:28.993 17:31:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:09:28.993 17:31:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:09:28.993 17:31:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 3 -gt 2 ']' 00:09:28.993 17:31:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:09:28.993 17:31:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:28.993 17:31:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:09:28.993 17:31:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:28.993 17:31:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:28.993 BaseBdev2 00:09:28.993 17:31:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:28.993 17:31:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:09:28.993 17:31:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:09:28.993 17:31:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:09:28.993 17:31:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:09:28.993 17:31:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:09:28.993 17:31:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:09:28.993 17:31:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:09:28.993 17:31:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:28.993 17:31:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:28.993 17:31:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:28.993 17:31:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:09:28.993 17:31:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:28.993 17:31:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:28.993 [ 00:09:28.993 { 00:09:28.993 "name": "BaseBdev2", 00:09:28.993 "aliases": [ 00:09:28.993 "fa67d169-8e79-4583-a193-e31661ecdd3b" 00:09:28.993 ], 00:09:28.993 "product_name": "Malloc disk", 00:09:28.993 "block_size": 512, 00:09:28.993 "num_blocks": 65536, 00:09:28.993 "uuid": "fa67d169-8e79-4583-a193-e31661ecdd3b", 00:09:28.993 "assigned_rate_limits": { 00:09:28.993 "rw_ios_per_sec": 0, 00:09:28.993 "rw_mbytes_per_sec": 0, 00:09:28.993 "r_mbytes_per_sec": 0, 00:09:28.993 "w_mbytes_per_sec": 0 00:09:28.993 }, 00:09:28.993 "claimed": false, 00:09:28.993 "zoned": false, 00:09:28.993 "supported_io_types": { 00:09:28.993 "read": true, 00:09:28.993 "write": true, 00:09:28.993 "unmap": true, 00:09:28.993 "flush": true, 00:09:28.993 "reset": true, 00:09:28.993 "nvme_admin": false, 00:09:28.993 "nvme_io": false, 00:09:28.993 "nvme_io_md": false, 00:09:28.993 "write_zeroes": true, 00:09:28.993 "zcopy": true, 00:09:28.993 "get_zone_info": false, 00:09:28.993 "zone_management": false, 00:09:28.993 "zone_append": false, 00:09:28.993 "compare": false, 00:09:28.993 "compare_and_write": false, 00:09:28.993 "abort": true, 00:09:28.993 "seek_hole": false, 00:09:28.993 "seek_data": false, 00:09:28.993 "copy": true, 00:09:28.993 "nvme_iov_md": false 00:09:28.993 }, 00:09:28.993 "memory_domains": [ 00:09:28.993 { 00:09:28.993 "dma_device_id": "system", 00:09:28.993 "dma_device_type": 1 00:09:28.993 }, 00:09:28.993 { 00:09:28.993 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:28.993 "dma_device_type": 2 00:09:28.993 } 00:09:28.993 ], 00:09:28.993 "driver_specific": {} 00:09:28.993 } 00:09:28.993 ] 00:09:28.993 17:31:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:28.993 17:31:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:09:28.993 17:31:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:09:28.993 17:31:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:28.993 17:31:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:09:28.993 17:31:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:28.993 17:31:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:29.269 BaseBdev3 00:09:29.269 17:31:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:29.269 17:31:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:09:29.269 17:31:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:09:29.269 17:31:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:09:29.269 17:31:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:09:29.269 17:31:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:09:29.269 17:31:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:09:29.269 17:31:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:09:29.269 17:31:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:29.269 17:31:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:29.269 17:31:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:29.269 17:31:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:09:29.269 17:31:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:29.269 17:31:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:29.269 [ 00:09:29.269 { 00:09:29.269 "name": "BaseBdev3", 00:09:29.269 "aliases": [ 00:09:29.269 "d2748fc8-f22b-4b66-92db-70d41507870c" 00:09:29.269 ], 00:09:29.269 "product_name": "Malloc disk", 00:09:29.269 "block_size": 512, 00:09:29.269 "num_blocks": 65536, 00:09:29.269 "uuid": "d2748fc8-f22b-4b66-92db-70d41507870c", 00:09:29.269 "assigned_rate_limits": { 00:09:29.269 "rw_ios_per_sec": 0, 00:09:29.269 "rw_mbytes_per_sec": 0, 00:09:29.269 "r_mbytes_per_sec": 0, 00:09:29.269 "w_mbytes_per_sec": 0 00:09:29.269 }, 00:09:29.269 "claimed": false, 00:09:29.269 "zoned": false, 00:09:29.269 "supported_io_types": { 00:09:29.269 "read": true, 00:09:29.269 "write": true, 00:09:29.269 "unmap": true, 00:09:29.269 "flush": true, 00:09:29.269 "reset": true, 00:09:29.269 "nvme_admin": false, 00:09:29.269 "nvme_io": false, 00:09:29.269 "nvme_io_md": false, 00:09:29.269 "write_zeroes": true, 00:09:29.269 "zcopy": true, 00:09:29.269 "get_zone_info": false, 00:09:29.269 "zone_management": false, 00:09:29.269 "zone_append": false, 00:09:29.269 "compare": false, 00:09:29.269 "compare_and_write": false, 00:09:29.269 "abort": true, 00:09:29.269 "seek_hole": false, 00:09:29.269 "seek_data": false, 00:09:29.269 "copy": true, 00:09:29.269 "nvme_iov_md": false 00:09:29.269 }, 00:09:29.269 "memory_domains": [ 00:09:29.269 { 00:09:29.269 "dma_device_id": "system", 00:09:29.269 "dma_device_type": 1 00:09:29.269 }, 00:09:29.269 { 00:09:29.269 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:29.269 "dma_device_type": 2 00:09:29.269 } 00:09:29.269 ], 00:09:29.269 "driver_specific": {} 00:09:29.269 } 00:09:29.269 ] 00:09:29.269 17:31:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:29.269 17:31:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:09:29.269 17:31:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:09:29.269 17:31:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:29.269 17:31:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:09:29.269 17:31:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:29.269 17:31:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:29.269 [2024-11-27 17:31:00.227935] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:29.269 [2024-11-27 17:31:00.228048] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:29.269 [2024-11-27 17:31:00.228089] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:29.269 [2024-11-27 17:31:00.230247] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:29.269 17:31:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:29.269 17:31:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:29.269 17:31:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:29.269 17:31:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:29.269 17:31:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:29.269 17:31:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:29.269 17:31:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:29.269 17:31:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:29.269 17:31:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:29.269 17:31:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:29.269 17:31:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:29.269 17:31:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:29.269 17:31:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:29.269 17:31:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:29.269 17:31:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:29.269 17:31:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:29.269 17:31:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:29.269 "name": "Existed_Raid", 00:09:29.269 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:29.269 "strip_size_kb": 0, 00:09:29.269 "state": "configuring", 00:09:29.269 "raid_level": "raid1", 00:09:29.269 "superblock": false, 00:09:29.269 "num_base_bdevs": 3, 00:09:29.269 "num_base_bdevs_discovered": 2, 00:09:29.269 "num_base_bdevs_operational": 3, 00:09:29.269 "base_bdevs_list": [ 00:09:29.269 { 00:09:29.269 "name": "BaseBdev1", 00:09:29.269 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:29.269 "is_configured": false, 00:09:29.269 "data_offset": 0, 00:09:29.269 "data_size": 0 00:09:29.269 }, 00:09:29.269 { 00:09:29.269 "name": "BaseBdev2", 00:09:29.269 "uuid": "fa67d169-8e79-4583-a193-e31661ecdd3b", 00:09:29.269 "is_configured": true, 00:09:29.269 "data_offset": 0, 00:09:29.269 "data_size": 65536 00:09:29.269 }, 00:09:29.269 { 00:09:29.269 "name": "BaseBdev3", 00:09:29.269 "uuid": "d2748fc8-f22b-4b66-92db-70d41507870c", 00:09:29.269 "is_configured": true, 00:09:29.269 "data_offset": 0, 00:09:29.269 "data_size": 65536 00:09:29.269 } 00:09:29.269 ] 00:09:29.269 }' 00:09:29.269 17:31:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:29.269 17:31:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:29.529 17:31:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:09:29.529 17:31:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:29.529 17:31:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:29.529 [2024-11-27 17:31:00.703123] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:09:29.530 17:31:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:29.530 17:31:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:29.530 17:31:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:29.530 17:31:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:29.530 17:31:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:29.530 17:31:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:29.530 17:31:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:29.530 17:31:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:29.530 17:31:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:29.530 17:31:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:29.530 17:31:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:29.530 17:31:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:29.530 17:31:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:29.530 17:31:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:29.530 17:31:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:29.790 17:31:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:29.790 17:31:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:29.790 "name": "Existed_Raid", 00:09:29.790 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:29.790 "strip_size_kb": 0, 00:09:29.790 "state": "configuring", 00:09:29.790 "raid_level": "raid1", 00:09:29.790 "superblock": false, 00:09:29.790 "num_base_bdevs": 3, 00:09:29.790 "num_base_bdevs_discovered": 1, 00:09:29.790 "num_base_bdevs_operational": 3, 00:09:29.790 "base_bdevs_list": [ 00:09:29.790 { 00:09:29.790 "name": "BaseBdev1", 00:09:29.790 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:29.790 "is_configured": false, 00:09:29.790 "data_offset": 0, 00:09:29.790 "data_size": 0 00:09:29.790 }, 00:09:29.790 { 00:09:29.790 "name": null, 00:09:29.790 "uuid": "fa67d169-8e79-4583-a193-e31661ecdd3b", 00:09:29.790 "is_configured": false, 00:09:29.790 "data_offset": 0, 00:09:29.790 "data_size": 65536 00:09:29.790 }, 00:09:29.790 { 00:09:29.790 "name": "BaseBdev3", 00:09:29.790 "uuid": "d2748fc8-f22b-4b66-92db-70d41507870c", 00:09:29.790 "is_configured": true, 00:09:29.790 "data_offset": 0, 00:09:29.790 "data_size": 65536 00:09:29.790 } 00:09:29.790 ] 00:09:29.790 }' 00:09:29.790 17:31:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:29.790 17:31:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:30.050 17:31:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:30.050 17:31:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:09:30.050 17:31:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:30.050 17:31:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:30.050 17:31:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:30.050 17:31:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:09:30.050 17:31:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:09:30.050 17:31:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:30.050 17:31:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:30.050 [2024-11-27 17:31:01.187064] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:30.050 BaseBdev1 00:09:30.050 17:31:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:30.050 17:31:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:09:30.050 17:31:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:09:30.050 17:31:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:09:30.050 17:31:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:09:30.050 17:31:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:09:30.050 17:31:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:09:30.050 17:31:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:09:30.050 17:31:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:30.050 17:31:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:30.050 17:31:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:30.050 17:31:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:09:30.050 17:31:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:30.050 17:31:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:30.050 [ 00:09:30.050 { 00:09:30.050 "name": "BaseBdev1", 00:09:30.050 "aliases": [ 00:09:30.050 "ab8ba4de-5ae4-4176-ba04-a58482879a12" 00:09:30.050 ], 00:09:30.050 "product_name": "Malloc disk", 00:09:30.050 "block_size": 512, 00:09:30.050 "num_blocks": 65536, 00:09:30.050 "uuid": "ab8ba4de-5ae4-4176-ba04-a58482879a12", 00:09:30.050 "assigned_rate_limits": { 00:09:30.050 "rw_ios_per_sec": 0, 00:09:30.050 "rw_mbytes_per_sec": 0, 00:09:30.050 "r_mbytes_per_sec": 0, 00:09:30.050 "w_mbytes_per_sec": 0 00:09:30.050 }, 00:09:30.050 "claimed": true, 00:09:30.050 "claim_type": "exclusive_write", 00:09:30.050 "zoned": false, 00:09:30.050 "supported_io_types": { 00:09:30.050 "read": true, 00:09:30.050 "write": true, 00:09:30.050 "unmap": true, 00:09:30.050 "flush": true, 00:09:30.050 "reset": true, 00:09:30.050 "nvme_admin": false, 00:09:30.050 "nvme_io": false, 00:09:30.050 "nvme_io_md": false, 00:09:30.050 "write_zeroes": true, 00:09:30.050 "zcopy": true, 00:09:30.050 "get_zone_info": false, 00:09:30.050 "zone_management": false, 00:09:30.050 "zone_append": false, 00:09:30.050 "compare": false, 00:09:30.050 "compare_and_write": false, 00:09:30.050 "abort": true, 00:09:30.050 "seek_hole": false, 00:09:30.050 "seek_data": false, 00:09:30.050 "copy": true, 00:09:30.050 "nvme_iov_md": false 00:09:30.050 }, 00:09:30.050 "memory_domains": [ 00:09:30.050 { 00:09:30.050 "dma_device_id": "system", 00:09:30.050 "dma_device_type": 1 00:09:30.050 }, 00:09:30.050 { 00:09:30.050 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:30.050 "dma_device_type": 2 00:09:30.051 } 00:09:30.051 ], 00:09:30.051 "driver_specific": {} 00:09:30.051 } 00:09:30.051 ] 00:09:30.051 17:31:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:30.051 17:31:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:09:30.051 17:31:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:30.051 17:31:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:30.051 17:31:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:30.051 17:31:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:30.051 17:31:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:30.051 17:31:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:30.051 17:31:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:30.051 17:31:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:30.051 17:31:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:30.051 17:31:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:30.051 17:31:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:30.051 17:31:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:30.051 17:31:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:30.051 17:31:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:30.310 17:31:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:30.310 17:31:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:30.310 "name": "Existed_Raid", 00:09:30.310 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:30.310 "strip_size_kb": 0, 00:09:30.310 "state": "configuring", 00:09:30.310 "raid_level": "raid1", 00:09:30.310 "superblock": false, 00:09:30.310 "num_base_bdevs": 3, 00:09:30.310 "num_base_bdevs_discovered": 2, 00:09:30.310 "num_base_bdevs_operational": 3, 00:09:30.310 "base_bdevs_list": [ 00:09:30.310 { 00:09:30.310 "name": "BaseBdev1", 00:09:30.310 "uuid": "ab8ba4de-5ae4-4176-ba04-a58482879a12", 00:09:30.310 "is_configured": true, 00:09:30.310 "data_offset": 0, 00:09:30.310 "data_size": 65536 00:09:30.310 }, 00:09:30.310 { 00:09:30.310 "name": null, 00:09:30.311 "uuid": "fa67d169-8e79-4583-a193-e31661ecdd3b", 00:09:30.311 "is_configured": false, 00:09:30.311 "data_offset": 0, 00:09:30.311 "data_size": 65536 00:09:30.311 }, 00:09:30.311 { 00:09:30.311 "name": "BaseBdev3", 00:09:30.311 "uuid": "d2748fc8-f22b-4b66-92db-70d41507870c", 00:09:30.311 "is_configured": true, 00:09:30.311 "data_offset": 0, 00:09:30.311 "data_size": 65536 00:09:30.311 } 00:09:30.311 ] 00:09:30.311 }' 00:09:30.311 17:31:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:30.311 17:31:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:30.570 17:31:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:30.570 17:31:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:30.570 17:31:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:09:30.570 17:31:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:30.570 17:31:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:30.570 17:31:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:09:30.570 17:31:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:09:30.570 17:31:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:30.570 17:31:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:30.570 [2024-11-27 17:31:01.690277] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:09:30.570 17:31:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:30.570 17:31:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:30.570 17:31:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:30.570 17:31:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:30.570 17:31:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:30.570 17:31:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:30.570 17:31:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:30.570 17:31:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:30.570 17:31:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:30.570 17:31:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:30.570 17:31:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:30.570 17:31:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:30.571 17:31:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:30.571 17:31:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:30.571 17:31:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:30.571 17:31:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:30.571 17:31:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:30.571 "name": "Existed_Raid", 00:09:30.571 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:30.571 "strip_size_kb": 0, 00:09:30.571 "state": "configuring", 00:09:30.571 "raid_level": "raid1", 00:09:30.571 "superblock": false, 00:09:30.571 "num_base_bdevs": 3, 00:09:30.571 "num_base_bdevs_discovered": 1, 00:09:30.571 "num_base_bdevs_operational": 3, 00:09:30.571 "base_bdevs_list": [ 00:09:30.571 { 00:09:30.571 "name": "BaseBdev1", 00:09:30.571 "uuid": "ab8ba4de-5ae4-4176-ba04-a58482879a12", 00:09:30.571 "is_configured": true, 00:09:30.571 "data_offset": 0, 00:09:30.571 "data_size": 65536 00:09:30.571 }, 00:09:30.571 { 00:09:30.571 "name": null, 00:09:30.571 "uuid": "fa67d169-8e79-4583-a193-e31661ecdd3b", 00:09:30.571 "is_configured": false, 00:09:30.571 "data_offset": 0, 00:09:30.571 "data_size": 65536 00:09:30.571 }, 00:09:30.571 { 00:09:30.571 "name": null, 00:09:30.571 "uuid": "d2748fc8-f22b-4b66-92db-70d41507870c", 00:09:30.571 "is_configured": false, 00:09:30.571 "data_offset": 0, 00:09:30.571 "data_size": 65536 00:09:30.571 } 00:09:30.571 ] 00:09:30.571 }' 00:09:30.571 17:31:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:30.571 17:31:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:31.140 17:31:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:09:31.140 17:31:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:31.140 17:31:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:31.140 17:31:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:31.140 17:31:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:31.140 17:31:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:09:31.140 17:31:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:09:31.140 17:31:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:31.140 17:31:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:31.140 [2024-11-27 17:31:02.209377] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:31.140 17:31:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:31.140 17:31:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:31.140 17:31:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:31.140 17:31:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:31.140 17:31:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:31.140 17:31:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:31.140 17:31:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:31.140 17:31:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:31.140 17:31:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:31.140 17:31:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:31.140 17:31:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:31.140 17:31:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:31.140 17:31:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:31.140 17:31:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:31.140 17:31:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:31.140 17:31:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:31.141 17:31:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:31.141 "name": "Existed_Raid", 00:09:31.141 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:31.141 "strip_size_kb": 0, 00:09:31.141 "state": "configuring", 00:09:31.141 "raid_level": "raid1", 00:09:31.141 "superblock": false, 00:09:31.141 "num_base_bdevs": 3, 00:09:31.141 "num_base_bdevs_discovered": 2, 00:09:31.141 "num_base_bdevs_operational": 3, 00:09:31.141 "base_bdevs_list": [ 00:09:31.141 { 00:09:31.141 "name": "BaseBdev1", 00:09:31.141 "uuid": "ab8ba4de-5ae4-4176-ba04-a58482879a12", 00:09:31.141 "is_configured": true, 00:09:31.141 "data_offset": 0, 00:09:31.141 "data_size": 65536 00:09:31.141 }, 00:09:31.141 { 00:09:31.141 "name": null, 00:09:31.141 "uuid": "fa67d169-8e79-4583-a193-e31661ecdd3b", 00:09:31.141 "is_configured": false, 00:09:31.141 "data_offset": 0, 00:09:31.141 "data_size": 65536 00:09:31.141 }, 00:09:31.141 { 00:09:31.141 "name": "BaseBdev3", 00:09:31.141 "uuid": "d2748fc8-f22b-4b66-92db-70d41507870c", 00:09:31.141 "is_configured": true, 00:09:31.141 "data_offset": 0, 00:09:31.141 "data_size": 65536 00:09:31.141 } 00:09:31.141 ] 00:09:31.141 }' 00:09:31.141 17:31:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:31.141 17:31:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:31.710 17:31:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:31.711 17:31:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:31.711 17:31:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:09:31.711 17:31:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:31.711 17:31:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:31.711 17:31:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:09:31.711 17:31:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:09:31.711 17:31:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:31.711 17:31:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:31.711 [2024-11-27 17:31:02.668637] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:09:31.711 17:31:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:31.711 17:31:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:31.711 17:31:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:31.711 17:31:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:31.711 17:31:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:31.711 17:31:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:31.711 17:31:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:31.711 17:31:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:31.711 17:31:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:31.711 17:31:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:31.711 17:31:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:31.711 17:31:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:31.711 17:31:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:31.711 17:31:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:31.711 17:31:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:31.711 17:31:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:31.711 17:31:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:31.711 "name": "Existed_Raid", 00:09:31.711 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:31.711 "strip_size_kb": 0, 00:09:31.711 "state": "configuring", 00:09:31.711 "raid_level": "raid1", 00:09:31.711 "superblock": false, 00:09:31.711 "num_base_bdevs": 3, 00:09:31.711 "num_base_bdevs_discovered": 1, 00:09:31.711 "num_base_bdevs_operational": 3, 00:09:31.711 "base_bdevs_list": [ 00:09:31.711 { 00:09:31.711 "name": null, 00:09:31.711 "uuid": "ab8ba4de-5ae4-4176-ba04-a58482879a12", 00:09:31.711 "is_configured": false, 00:09:31.711 "data_offset": 0, 00:09:31.711 "data_size": 65536 00:09:31.711 }, 00:09:31.711 { 00:09:31.711 "name": null, 00:09:31.711 "uuid": "fa67d169-8e79-4583-a193-e31661ecdd3b", 00:09:31.711 "is_configured": false, 00:09:31.711 "data_offset": 0, 00:09:31.711 "data_size": 65536 00:09:31.711 }, 00:09:31.711 { 00:09:31.711 "name": "BaseBdev3", 00:09:31.711 "uuid": "d2748fc8-f22b-4b66-92db-70d41507870c", 00:09:31.711 "is_configured": true, 00:09:31.711 "data_offset": 0, 00:09:31.711 "data_size": 65536 00:09:31.711 } 00:09:31.711 ] 00:09:31.711 }' 00:09:31.711 17:31:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:31.711 17:31:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:31.971 17:31:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:31.971 17:31:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:09:31.971 17:31:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:31.971 17:31:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:31.971 17:31:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:31.971 17:31:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:09:31.971 17:31:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:09:31.971 17:31:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:31.971 17:31:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:31.971 [2024-11-27 17:31:03.159710] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:32.230 17:31:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:32.230 17:31:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:32.230 17:31:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:32.230 17:31:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:32.230 17:31:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:32.230 17:31:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:32.230 17:31:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:32.230 17:31:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:32.230 17:31:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:32.230 17:31:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:32.230 17:31:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:32.230 17:31:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:32.230 17:31:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:32.230 17:31:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:32.230 17:31:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:32.230 17:31:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:32.230 17:31:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:32.230 "name": "Existed_Raid", 00:09:32.230 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:32.230 "strip_size_kb": 0, 00:09:32.230 "state": "configuring", 00:09:32.230 "raid_level": "raid1", 00:09:32.230 "superblock": false, 00:09:32.230 "num_base_bdevs": 3, 00:09:32.230 "num_base_bdevs_discovered": 2, 00:09:32.230 "num_base_bdevs_operational": 3, 00:09:32.230 "base_bdevs_list": [ 00:09:32.230 { 00:09:32.230 "name": null, 00:09:32.230 "uuid": "ab8ba4de-5ae4-4176-ba04-a58482879a12", 00:09:32.230 "is_configured": false, 00:09:32.230 "data_offset": 0, 00:09:32.230 "data_size": 65536 00:09:32.230 }, 00:09:32.230 { 00:09:32.230 "name": "BaseBdev2", 00:09:32.230 "uuid": "fa67d169-8e79-4583-a193-e31661ecdd3b", 00:09:32.230 "is_configured": true, 00:09:32.230 "data_offset": 0, 00:09:32.230 "data_size": 65536 00:09:32.230 }, 00:09:32.230 { 00:09:32.230 "name": "BaseBdev3", 00:09:32.230 "uuid": "d2748fc8-f22b-4b66-92db-70d41507870c", 00:09:32.230 "is_configured": true, 00:09:32.230 "data_offset": 0, 00:09:32.230 "data_size": 65536 00:09:32.230 } 00:09:32.230 ] 00:09:32.230 }' 00:09:32.230 17:31:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:32.230 17:31:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:32.490 17:31:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:32.490 17:31:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:09:32.490 17:31:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:32.490 17:31:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:32.490 17:31:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:32.490 17:31:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:09:32.490 17:31:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:32.490 17:31:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:09:32.490 17:31:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:32.490 17:31:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:32.490 17:31:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:32.490 17:31:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u ab8ba4de-5ae4-4176-ba04-a58482879a12 00:09:32.490 17:31:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:32.490 17:31:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:32.490 [2024-11-27 17:31:03.675562] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:09:32.490 [2024-11-27 17:31:03.675680] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:09:32.490 [2024-11-27 17:31:03.675693] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:09:32.490 [2024-11-27 17:31:03.676028] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002870 00:09:32.490 [2024-11-27 17:31:03.676187] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:09:32.490 [2024-11-27 17:31:03.676224] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001c80 00:09:32.490 [2024-11-27 17:31:03.676428] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:32.490 NewBaseBdev 00:09:32.750 17:31:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:32.750 17:31:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:09:32.750 17:31:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=NewBaseBdev 00:09:32.750 17:31:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:09:32.750 17:31:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:09:32.750 17:31:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:09:32.750 17:31:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:09:32.750 17:31:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:09:32.750 17:31:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:32.750 17:31:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:32.750 17:31:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:32.750 17:31:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:09:32.750 17:31:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:32.750 17:31:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:32.750 [ 00:09:32.750 { 00:09:32.750 "name": "NewBaseBdev", 00:09:32.750 "aliases": [ 00:09:32.750 "ab8ba4de-5ae4-4176-ba04-a58482879a12" 00:09:32.750 ], 00:09:32.750 "product_name": "Malloc disk", 00:09:32.750 "block_size": 512, 00:09:32.750 "num_blocks": 65536, 00:09:32.750 "uuid": "ab8ba4de-5ae4-4176-ba04-a58482879a12", 00:09:32.750 "assigned_rate_limits": { 00:09:32.750 "rw_ios_per_sec": 0, 00:09:32.750 "rw_mbytes_per_sec": 0, 00:09:32.750 "r_mbytes_per_sec": 0, 00:09:32.750 "w_mbytes_per_sec": 0 00:09:32.750 }, 00:09:32.750 "claimed": true, 00:09:32.750 "claim_type": "exclusive_write", 00:09:32.750 "zoned": false, 00:09:32.750 "supported_io_types": { 00:09:32.750 "read": true, 00:09:32.750 "write": true, 00:09:32.750 "unmap": true, 00:09:32.750 "flush": true, 00:09:32.750 "reset": true, 00:09:32.750 "nvme_admin": false, 00:09:32.750 "nvme_io": false, 00:09:32.750 "nvme_io_md": false, 00:09:32.750 "write_zeroes": true, 00:09:32.750 "zcopy": true, 00:09:32.750 "get_zone_info": false, 00:09:32.750 "zone_management": false, 00:09:32.750 "zone_append": false, 00:09:32.750 "compare": false, 00:09:32.750 "compare_and_write": false, 00:09:32.750 "abort": true, 00:09:32.750 "seek_hole": false, 00:09:32.750 "seek_data": false, 00:09:32.750 "copy": true, 00:09:32.750 "nvme_iov_md": false 00:09:32.750 }, 00:09:32.750 "memory_domains": [ 00:09:32.750 { 00:09:32.750 "dma_device_id": "system", 00:09:32.750 "dma_device_type": 1 00:09:32.750 }, 00:09:32.750 { 00:09:32.750 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:32.750 "dma_device_type": 2 00:09:32.750 } 00:09:32.750 ], 00:09:32.750 "driver_specific": {} 00:09:32.750 } 00:09:32.750 ] 00:09:32.750 17:31:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:32.750 17:31:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:09:32.750 17:31:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid1 0 3 00:09:32.750 17:31:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:32.750 17:31:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:32.750 17:31:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:32.750 17:31:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:32.750 17:31:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:32.750 17:31:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:32.750 17:31:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:32.750 17:31:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:32.750 17:31:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:32.750 17:31:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:32.750 17:31:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:32.750 17:31:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:32.750 17:31:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:32.750 17:31:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:32.750 17:31:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:32.750 "name": "Existed_Raid", 00:09:32.750 "uuid": "f686576f-b542-42aa-b400-8ebe39c2fa12", 00:09:32.750 "strip_size_kb": 0, 00:09:32.750 "state": "online", 00:09:32.750 "raid_level": "raid1", 00:09:32.750 "superblock": false, 00:09:32.750 "num_base_bdevs": 3, 00:09:32.750 "num_base_bdevs_discovered": 3, 00:09:32.750 "num_base_bdevs_operational": 3, 00:09:32.750 "base_bdevs_list": [ 00:09:32.750 { 00:09:32.750 "name": "NewBaseBdev", 00:09:32.750 "uuid": "ab8ba4de-5ae4-4176-ba04-a58482879a12", 00:09:32.750 "is_configured": true, 00:09:32.750 "data_offset": 0, 00:09:32.750 "data_size": 65536 00:09:32.750 }, 00:09:32.750 { 00:09:32.750 "name": "BaseBdev2", 00:09:32.750 "uuid": "fa67d169-8e79-4583-a193-e31661ecdd3b", 00:09:32.750 "is_configured": true, 00:09:32.750 "data_offset": 0, 00:09:32.750 "data_size": 65536 00:09:32.750 }, 00:09:32.750 { 00:09:32.750 "name": "BaseBdev3", 00:09:32.750 "uuid": "d2748fc8-f22b-4b66-92db-70d41507870c", 00:09:32.750 "is_configured": true, 00:09:32.750 "data_offset": 0, 00:09:32.750 "data_size": 65536 00:09:32.750 } 00:09:32.750 ] 00:09:32.750 }' 00:09:32.750 17:31:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:32.750 17:31:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:33.010 17:31:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:09:33.010 17:31:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:09:33.010 17:31:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:33.010 17:31:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:33.010 17:31:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:09:33.010 17:31:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:33.010 17:31:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:09:33.010 17:31:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:33.010 17:31:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:33.010 17:31:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:33.010 [2024-11-27 17:31:04.143142] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:33.010 17:31:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:33.010 17:31:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:33.010 "name": "Existed_Raid", 00:09:33.010 "aliases": [ 00:09:33.010 "f686576f-b542-42aa-b400-8ebe39c2fa12" 00:09:33.010 ], 00:09:33.010 "product_name": "Raid Volume", 00:09:33.010 "block_size": 512, 00:09:33.010 "num_blocks": 65536, 00:09:33.010 "uuid": "f686576f-b542-42aa-b400-8ebe39c2fa12", 00:09:33.010 "assigned_rate_limits": { 00:09:33.010 "rw_ios_per_sec": 0, 00:09:33.010 "rw_mbytes_per_sec": 0, 00:09:33.010 "r_mbytes_per_sec": 0, 00:09:33.010 "w_mbytes_per_sec": 0 00:09:33.010 }, 00:09:33.010 "claimed": false, 00:09:33.010 "zoned": false, 00:09:33.010 "supported_io_types": { 00:09:33.010 "read": true, 00:09:33.010 "write": true, 00:09:33.010 "unmap": false, 00:09:33.010 "flush": false, 00:09:33.010 "reset": true, 00:09:33.010 "nvme_admin": false, 00:09:33.011 "nvme_io": false, 00:09:33.011 "nvme_io_md": false, 00:09:33.011 "write_zeroes": true, 00:09:33.011 "zcopy": false, 00:09:33.011 "get_zone_info": false, 00:09:33.011 "zone_management": false, 00:09:33.011 "zone_append": false, 00:09:33.011 "compare": false, 00:09:33.011 "compare_and_write": false, 00:09:33.011 "abort": false, 00:09:33.011 "seek_hole": false, 00:09:33.011 "seek_data": false, 00:09:33.011 "copy": false, 00:09:33.011 "nvme_iov_md": false 00:09:33.011 }, 00:09:33.011 "memory_domains": [ 00:09:33.011 { 00:09:33.011 "dma_device_id": "system", 00:09:33.011 "dma_device_type": 1 00:09:33.011 }, 00:09:33.011 { 00:09:33.011 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:33.011 "dma_device_type": 2 00:09:33.011 }, 00:09:33.011 { 00:09:33.011 "dma_device_id": "system", 00:09:33.011 "dma_device_type": 1 00:09:33.011 }, 00:09:33.011 { 00:09:33.011 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:33.011 "dma_device_type": 2 00:09:33.011 }, 00:09:33.011 { 00:09:33.011 "dma_device_id": "system", 00:09:33.011 "dma_device_type": 1 00:09:33.011 }, 00:09:33.011 { 00:09:33.011 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:33.011 "dma_device_type": 2 00:09:33.011 } 00:09:33.011 ], 00:09:33.011 "driver_specific": { 00:09:33.011 "raid": { 00:09:33.011 "uuid": "f686576f-b542-42aa-b400-8ebe39c2fa12", 00:09:33.011 "strip_size_kb": 0, 00:09:33.011 "state": "online", 00:09:33.011 "raid_level": "raid1", 00:09:33.011 "superblock": false, 00:09:33.011 "num_base_bdevs": 3, 00:09:33.011 "num_base_bdevs_discovered": 3, 00:09:33.011 "num_base_bdevs_operational": 3, 00:09:33.011 "base_bdevs_list": [ 00:09:33.011 { 00:09:33.011 "name": "NewBaseBdev", 00:09:33.011 "uuid": "ab8ba4de-5ae4-4176-ba04-a58482879a12", 00:09:33.011 "is_configured": true, 00:09:33.011 "data_offset": 0, 00:09:33.011 "data_size": 65536 00:09:33.011 }, 00:09:33.011 { 00:09:33.011 "name": "BaseBdev2", 00:09:33.011 "uuid": "fa67d169-8e79-4583-a193-e31661ecdd3b", 00:09:33.011 "is_configured": true, 00:09:33.011 "data_offset": 0, 00:09:33.011 "data_size": 65536 00:09:33.011 }, 00:09:33.011 { 00:09:33.011 "name": "BaseBdev3", 00:09:33.011 "uuid": "d2748fc8-f22b-4b66-92db-70d41507870c", 00:09:33.011 "is_configured": true, 00:09:33.011 "data_offset": 0, 00:09:33.011 "data_size": 65536 00:09:33.011 } 00:09:33.011 ] 00:09:33.011 } 00:09:33.011 } 00:09:33.011 }' 00:09:33.011 17:31:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:33.271 17:31:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:09:33.272 BaseBdev2 00:09:33.272 BaseBdev3' 00:09:33.272 17:31:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:33.272 17:31:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:33.272 17:31:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:33.272 17:31:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:09:33.272 17:31:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:33.272 17:31:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:33.272 17:31:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:33.272 17:31:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:33.272 17:31:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:33.272 17:31:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:33.272 17:31:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:33.272 17:31:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:33.272 17:31:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:09:33.272 17:31:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:33.272 17:31:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:33.272 17:31:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:33.272 17:31:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:33.272 17:31:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:33.272 17:31:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:33.272 17:31:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:09:33.272 17:31:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:33.272 17:31:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:33.272 17:31:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:33.272 17:31:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:33.272 17:31:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:33.272 17:31:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:33.272 17:31:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:33.272 17:31:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:33.272 17:31:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:33.272 [2024-11-27 17:31:04.442317] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:33.272 [2024-11-27 17:31:04.442346] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:33.272 [2024-11-27 17:31:04.442428] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:33.272 [2024-11-27 17:31:04.442689] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:33.272 [2024-11-27 17:31:04.442698] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name Existed_Raid, state offline 00:09:33.272 17:31:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:33.272 17:31:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 78285 00:09:33.272 17:31:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@950 -- # '[' -z 78285 ']' 00:09:33.272 17:31:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # kill -0 78285 00:09:33.272 17:31:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@955 -- # uname 00:09:33.272 17:31:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:09:33.272 17:31:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 78285 00:09:33.532 killing process with pid 78285 00:09:33.532 17:31:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:09:33.532 17:31:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:09:33.532 17:31:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 78285' 00:09:33.532 17:31:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@969 -- # kill 78285 00:09:33.532 [2024-11-27 17:31:04.491518] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:09:33.532 17:31:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@974 -- # wait 78285 00:09:33.533 [2024-11-27 17:31:04.549726] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:09:33.793 17:31:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:09:33.793 00:09:33.793 real 0m9.026s 00:09:33.793 user 0m15.138s 00:09:33.793 sys 0m1.897s 00:09:33.793 17:31:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:09:33.793 ************************************ 00:09:33.793 END TEST raid_state_function_test 00:09:33.793 ************************************ 00:09:33.793 17:31:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:33.793 17:31:04 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test raid1 3 true 00:09:33.793 17:31:04 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:09:33.793 17:31:04 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:09:33.793 17:31:04 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:09:34.062 ************************************ 00:09:34.062 START TEST raid_state_function_test_sb 00:09:34.062 ************************************ 00:09:34.062 17:31:04 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1125 -- # raid_state_function_test raid1 3 true 00:09:34.062 17:31:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:09:34.062 17:31:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=3 00:09:34.062 17:31:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:09:34.062 17:31:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:09:34.062 17:31:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:09:34.062 17:31:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:34.062 17:31:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:09:34.062 17:31:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:34.062 17:31:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:34.062 17:31:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:09:34.062 17:31:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:34.062 17:31:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:34.062 17:31:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:09:34.062 17:31:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:34.062 17:31:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:34.062 17:31:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:09:34.062 17:31:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:09:34.062 17:31:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:09:34.062 17:31:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:09:34.062 17:31:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:09:34.062 17:31:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:09:34.062 17:31:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:09:34.062 17:31:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:09:34.062 17:31:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:09:34.062 17:31:04 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:09:34.062 Process raid pid: 78895 00:09:34.063 17:31:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=78895 00:09:34.063 17:31:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:09:34.063 17:31:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 78895' 00:09:34.063 17:31:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 78895 00:09:34.063 17:31:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@831 -- # '[' -z 78895 ']' 00:09:34.063 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:34.063 17:31:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:34.063 17:31:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@836 -- # local max_retries=100 00:09:34.063 17:31:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:34.063 17:31:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # xtrace_disable 00:09:34.063 17:31:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:34.063 [2024-11-27 17:31:05.089097] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:09:34.063 [2024-11-27 17:31:05.089231] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:09:34.063 [2024-11-27 17:31:05.237572] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:34.322 [2024-11-27 17:31:05.306846] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:09:34.322 [2024-11-27 17:31:05.382845] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:34.322 [2024-11-27 17:31:05.382885] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:34.892 17:31:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:09:34.892 17:31:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@864 -- # return 0 00:09:34.892 17:31:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:09:34.892 17:31:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:34.892 17:31:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:34.892 [2024-11-27 17:31:05.913930] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:34.892 [2024-11-27 17:31:05.913988] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:34.892 [2024-11-27 17:31:05.914016] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:34.892 [2024-11-27 17:31:05.914026] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:34.892 [2024-11-27 17:31:05.914032] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:34.892 [2024-11-27 17:31:05.914046] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:34.892 17:31:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:34.892 17:31:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:34.892 17:31:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:34.892 17:31:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:34.892 17:31:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:34.892 17:31:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:34.892 17:31:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:34.892 17:31:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:34.892 17:31:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:34.892 17:31:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:34.892 17:31:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:34.892 17:31:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:34.892 17:31:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:34.892 17:31:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:34.892 17:31:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:34.892 17:31:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:34.892 17:31:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:34.892 "name": "Existed_Raid", 00:09:34.892 "uuid": "3da12f27-7fbb-4ca7-8bcf-09bb92d091ff", 00:09:34.892 "strip_size_kb": 0, 00:09:34.892 "state": "configuring", 00:09:34.892 "raid_level": "raid1", 00:09:34.892 "superblock": true, 00:09:34.892 "num_base_bdevs": 3, 00:09:34.892 "num_base_bdevs_discovered": 0, 00:09:34.892 "num_base_bdevs_operational": 3, 00:09:34.892 "base_bdevs_list": [ 00:09:34.892 { 00:09:34.892 "name": "BaseBdev1", 00:09:34.892 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:34.892 "is_configured": false, 00:09:34.892 "data_offset": 0, 00:09:34.892 "data_size": 0 00:09:34.892 }, 00:09:34.892 { 00:09:34.892 "name": "BaseBdev2", 00:09:34.892 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:34.892 "is_configured": false, 00:09:34.892 "data_offset": 0, 00:09:34.892 "data_size": 0 00:09:34.892 }, 00:09:34.892 { 00:09:34.892 "name": "BaseBdev3", 00:09:34.892 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:34.892 "is_configured": false, 00:09:34.892 "data_offset": 0, 00:09:34.892 "data_size": 0 00:09:34.892 } 00:09:34.892 ] 00:09:34.892 }' 00:09:34.892 17:31:05 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:34.892 17:31:05 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:35.462 17:31:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:35.462 17:31:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:35.462 17:31:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:35.462 [2024-11-27 17:31:06.361021] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:35.462 [2024-11-27 17:31:06.361118] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:09:35.462 17:31:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:35.462 17:31:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:09:35.462 17:31:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:35.462 17:31:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:35.462 [2024-11-27 17:31:06.373028] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:35.462 [2024-11-27 17:31:06.373107] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:35.462 [2024-11-27 17:31:06.373165] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:35.462 [2024-11-27 17:31:06.373190] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:35.462 [2024-11-27 17:31:06.373215] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:35.462 [2024-11-27 17:31:06.373244] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:35.462 17:31:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:35.462 17:31:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:09:35.462 17:31:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:35.462 17:31:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:35.462 [2024-11-27 17:31:06.400052] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:35.462 BaseBdev1 00:09:35.462 17:31:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:35.462 17:31:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:09:35.462 17:31:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:09:35.462 17:31:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:09:35.462 17:31:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:09:35.462 17:31:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:09:35.462 17:31:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:09:35.462 17:31:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:09:35.462 17:31:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:35.462 17:31:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:35.462 17:31:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:35.462 17:31:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:09:35.462 17:31:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:35.462 17:31:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:35.462 [ 00:09:35.462 { 00:09:35.462 "name": "BaseBdev1", 00:09:35.462 "aliases": [ 00:09:35.463 "1dd7a0ad-c398-4810-a5fa-86103c5068c0" 00:09:35.463 ], 00:09:35.463 "product_name": "Malloc disk", 00:09:35.463 "block_size": 512, 00:09:35.463 "num_blocks": 65536, 00:09:35.463 "uuid": "1dd7a0ad-c398-4810-a5fa-86103c5068c0", 00:09:35.463 "assigned_rate_limits": { 00:09:35.463 "rw_ios_per_sec": 0, 00:09:35.463 "rw_mbytes_per_sec": 0, 00:09:35.463 "r_mbytes_per_sec": 0, 00:09:35.463 "w_mbytes_per_sec": 0 00:09:35.463 }, 00:09:35.463 "claimed": true, 00:09:35.463 "claim_type": "exclusive_write", 00:09:35.463 "zoned": false, 00:09:35.463 "supported_io_types": { 00:09:35.463 "read": true, 00:09:35.463 "write": true, 00:09:35.463 "unmap": true, 00:09:35.463 "flush": true, 00:09:35.463 "reset": true, 00:09:35.463 "nvme_admin": false, 00:09:35.463 "nvme_io": false, 00:09:35.463 "nvme_io_md": false, 00:09:35.463 "write_zeroes": true, 00:09:35.463 "zcopy": true, 00:09:35.463 "get_zone_info": false, 00:09:35.463 "zone_management": false, 00:09:35.463 "zone_append": false, 00:09:35.463 "compare": false, 00:09:35.463 "compare_and_write": false, 00:09:35.463 "abort": true, 00:09:35.463 "seek_hole": false, 00:09:35.463 "seek_data": false, 00:09:35.463 "copy": true, 00:09:35.463 "nvme_iov_md": false 00:09:35.463 }, 00:09:35.463 "memory_domains": [ 00:09:35.463 { 00:09:35.463 "dma_device_id": "system", 00:09:35.463 "dma_device_type": 1 00:09:35.463 }, 00:09:35.463 { 00:09:35.463 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:35.463 "dma_device_type": 2 00:09:35.463 } 00:09:35.463 ], 00:09:35.463 "driver_specific": {} 00:09:35.463 } 00:09:35.463 ] 00:09:35.463 17:31:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:35.463 17:31:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:09:35.463 17:31:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:35.463 17:31:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:35.463 17:31:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:35.463 17:31:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:35.463 17:31:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:35.463 17:31:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:35.463 17:31:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:35.463 17:31:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:35.463 17:31:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:35.463 17:31:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:35.463 17:31:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:35.463 17:31:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:35.463 17:31:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:35.463 17:31:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:35.463 17:31:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:35.463 17:31:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:35.463 "name": "Existed_Raid", 00:09:35.463 "uuid": "7e394e7e-f014-485a-b0d7-48908d3b1324", 00:09:35.463 "strip_size_kb": 0, 00:09:35.463 "state": "configuring", 00:09:35.463 "raid_level": "raid1", 00:09:35.463 "superblock": true, 00:09:35.463 "num_base_bdevs": 3, 00:09:35.463 "num_base_bdevs_discovered": 1, 00:09:35.463 "num_base_bdevs_operational": 3, 00:09:35.463 "base_bdevs_list": [ 00:09:35.463 { 00:09:35.463 "name": "BaseBdev1", 00:09:35.463 "uuid": "1dd7a0ad-c398-4810-a5fa-86103c5068c0", 00:09:35.463 "is_configured": true, 00:09:35.463 "data_offset": 2048, 00:09:35.463 "data_size": 63488 00:09:35.463 }, 00:09:35.463 { 00:09:35.463 "name": "BaseBdev2", 00:09:35.463 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:35.463 "is_configured": false, 00:09:35.463 "data_offset": 0, 00:09:35.463 "data_size": 0 00:09:35.463 }, 00:09:35.463 { 00:09:35.463 "name": "BaseBdev3", 00:09:35.463 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:35.463 "is_configured": false, 00:09:35.463 "data_offset": 0, 00:09:35.463 "data_size": 0 00:09:35.463 } 00:09:35.463 ] 00:09:35.463 }' 00:09:35.463 17:31:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:35.463 17:31:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:36.034 17:31:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:36.034 17:31:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:36.034 17:31:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:36.034 [2024-11-27 17:31:06.923184] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:36.034 [2024-11-27 17:31:06.923233] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:09:36.034 17:31:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:36.034 17:31:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:09:36.034 17:31:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:36.034 17:31:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:36.034 [2024-11-27 17:31:06.931225] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:36.034 [2024-11-27 17:31:06.933351] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:36.034 [2024-11-27 17:31:06.933435] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:36.034 [2024-11-27 17:31:06.933449] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:36.034 [2024-11-27 17:31:06.933460] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:36.034 17:31:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:36.034 17:31:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:09:36.034 17:31:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:36.034 17:31:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:36.034 17:31:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:36.034 17:31:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:36.034 17:31:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:36.034 17:31:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:36.034 17:31:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:36.034 17:31:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:36.034 17:31:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:36.034 17:31:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:36.034 17:31:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:36.034 17:31:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:36.034 17:31:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:36.034 17:31:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:36.034 17:31:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:36.034 17:31:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:36.034 17:31:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:36.034 "name": "Existed_Raid", 00:09:36.034 "uuid": "b930a285-a867-406b-9cb3-7c03032f3fd4", 00:09:36.034 "strip_size_kb": 0, 00:09:36.034 "state": "configuring", 00:09:36.034 "raid_level": "raid1", 00:09:36.034 "superblock": true, 00:09:36.034 "num_base_bdevs": 3, 00:09:36.034 "num_base_bdevs_discovered": 1, 00:09:36.034 "num_base_bdevs_operational": 3, 00:09:36.034 "base_bdevs_list": [ 00:09:36.034 { 00:09:36.034 "name": "BaseBdev1", 00:09:36.034 "uuid": "1dd7a0ad-c398-4810-a5fa-86103c5068c0", 00:09:36.034 "is_configured": true, 00:09:36.034 "data_offset": 2048, 00:09:36.034 "data_size": 63488 00:09:36.034 }, 00:09:36.034 { 00:09:36.034 "name": "BaseBdev2", 00:09:36.034 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:36.034 "is_configured": false, 00:09:36.034 "data_offset": 0, 00:09:36.034 "data_size": 0 00:09:36.034 }, 00:09:36.034 { 00:09:36.034 "name": "BaseBdev3", 00:09:36.034 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:36.034 "is_configured": false, 00:09:36.034 "data_offset": 0, 00:09:36.034 "data_size": 0 00:09:36.034 } 00:09:36.034 ] 00:09:36.034 }' 00:09:36.034 17:31:06 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:36.034 17:31:06 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:36.294 17:31:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:09:36.294 17:31:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:36.294 17:31:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:36.294 [2024-11-27 17:31:07.328946] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:36.294 BaseBdev2 00:09:36.294 17:31:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:36.294 17:31:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:09:36.294 17:31:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:09:36.294 17:31:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:09:36.294 17:31:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:09:36.294 17:31:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:09:36.294 17:31:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:09:36.294 17:31:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:09:36.294 17:31:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:36.294 17:31:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:36.294 17:31:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:36.294 17:31:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:09:36.294 17:31:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:36.294 17:31:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:36.294 [ 00:09:36.294 { 00:09:36.294 "name": "BaseBdev2", 00:09:36.294 "aliases": [ 00:09:36.294 "94244386-eef5-4952-a87c-3a8da0da35c0" 00:09:36.294 ], 00:09:36.294 "product_name": "Malloc disk", 00:09:36.294 "block_size": 512, 00:09:36.294 "num_blocks": 65536, 00:09:36.294 "uuid": "94244386-eef5-4952-a87c-3a8da0da35c0", 00:09:36.294 "assigned_rate_limits": { 00:09:36.294 "rw_ios_per_sec": 0, 00:09:36.294 "rw_mbytes_per_sec": 0, 00:09:36.294 "r_mbytes_per_sec": 0, 00:09:36.294 "w_mbytes_per_sec": 0 00:09:36.294 }, 00:09:36.294 "claimed": true, 00:09:36.294 "claim_type": "exclusive_write", 00:09:36.294 "zoned": false, 00:09:36.294 "supported_io_types": { 00:09:36.294 "read": true, 00:09:36.294 "write": true, 00:09:36.294 "unmap": true, 00:09:36.294 "flush": true, 00:09:36.294 "reset": true, 00:09:36.294 "nvme_admin": false, 00:09:36.294 "nvme_io": false, 00:09:36.294 "nvme_io_md": false, 00:09:36.294 "write_zeroes": true, 00:09:36.294 "zcopy": true, 00:09:36.294 "get_zone_info": false, 00:09:36.294 "zone_management": false, 00:09:36.294 "zone_append": false, 00:09:36.294 "compare": false, 00:09:36.294 "compare_and_write": false, 00:09:36.295 "abort": true, 00:09:36.295 "seek_hole": false, 00:09:36.295 "seek_data": false, 00:09:36.295 "copy": true, 00:09:36.295 "nvme_iov_md": false 00:09:36.295 }, 00:09:36.295 "memory_domains": [ 00:09:36.295 { 00:09:36.295 "dma_device_id": "system", 00:09:36.295 "dma_device_type": 1 00:09:36.295 }, 00:09:36.295 { 00:09:36.295 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:36.295 "dma_device_type": 2 00:09:36.295 } 00:09:36.295 ], 00:09:36.295 "driver_specific": {} 00:09:36.295 } 00:09:36.295 ] 00:09:36.295 17:31:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:36.295 17:31:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:09:36.295 17:31:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:09:36.295 17:31:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:36.295 17:31:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:36.295 17:31:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:36.295 17:31:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:36.295 17:31:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:36.295 17:31:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:36.295 17:31:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:36.295 17:31:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:36.295 17:31:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:36.295 17:31:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:36.295 17:31:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:36.295 17:31:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:36.295 17:31:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:36.295 17:31:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:36.295 17:31:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:36.295 17:31:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:36.295 17:31:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:36.295 "name": "Existed_Raid", 00:09:36.295 "uuid": "b930a285-a867-406b-9cb3-7c03032f3fd4", 00:09:36.295 "strip_size_kb": 0, 00:09:36.295 "state": "configuring", 00:09:36.295 "raid_level": "raid1", 00:09:36.295 "superblock": true, 00:09:36.295 "num_base_bdevs": 3, 00:09:36.295 "num_base_bdevs_discovered": 2, 00:09:36.295 "num_base_bdevs_operational": 3, 00:09:36.295 "base_bdevs_list": [ 00:09:36.295 { 00:09:36.295 "name": "BaseBdev1", 00:09:36.295 "uuid": "1dd7a0ad-c398-4810-a5fa-86103c5068c0", 00:09:36.295 "is_configured": true, 00:09:36.295 "data_offset": 2048, 00:09:36.295 "data_size": 63488 00:09:36.295 }, 00:09:36.295 { 00:09:36.295 "name": "BaseBdev2", 00:09:36.295 "uuid": "94244386-eef5-4952-a87c-3a8da0da35c0", 00:09:36.295 "is_configured": true, 00:09:36.295 "data_offset": 2048, 00:09:36.295 "data_size": 63488 00:09:36.295 }, 00:09:36.295 { 00:09:36.295 "name": "BaseBdev3", 00:09:36.295 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:36.295 "is_configured": false, 00:09:36.295 "data_offset": 0, 00:09:36.295 "data_size": 0 00:09:36.295 } 00:09:36.295 ] 00:09:36.295 }' 00:09:36.295 17:31:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:36.295 17:31:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:36.866 17:31:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:09:36.866 17:31:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:36.866 17:31:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:36.866 [2024-11-27 17:31:07.808914] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:36.866 [2024-11-27 17:31:07.809252] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:09:36.866 [2024-11-27 17:31:07.809328] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:09:36.866 BaseBdev3 00:09:36.866 [2024-11-27 17:31:07.809641] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:09:36.866 [2024-11-27 17:31:07.809851] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:09:36.866 [2024-11-27 17:31:07.809894] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:09:36.866 [2024-11-27 17:31:07.810057] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:36.866 17:31:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:36.866 17:31:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:09:36.866 17:31:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:09:36.866 17:31:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:09:36.866 17:31:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:09:36.866 17:31:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:09:36.866 17:31:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:09:36.866 17:31:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:09:36.866 17:31:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:36.866 17:31:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:36.866 17:31:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:36.866 17:31:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:09:36.866 17:31:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:36.866 17:31:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:36.866 [ 00:09:36.866 { 00:09:36.866 "name": "BaseBdev3", 00:09:36.866 "aliases": [ 00:09:36.866 "bc09e990-a64f-4a3a-ac0a-2e3ddb1f2ad7" 00:09:36.866 ], 00:09:36.866 "product_name": "Malloc disk", 00:09:36.866 "block_size": 512, 00:09:36.866 "num_blocks": 65536, 00:09:36.866 "uuid": "bc09e990-a64f-4a3a-ac0a-2e3ddb1f2ad7", 00:09:36.866 "assigned_rate_limits": { 00:09:36.866 "rw_ios_per_sec": 0, 00:09:36.866 "rw_mbytes_per_sec": 0, 00:09:36.866 "r_mbytes_per_sec": 0, 00:09:36.866 "w_mbytes_per_sec": 0 00:09:36.866 }, 00:09:36.866 "claimed": true, 00:09:36.866 "claim_type": "exclusive_write", 00:09:36.866 "zoned": false, 00:09:36.866 "supported_io_types": { 00:09:36.866 "read": true, 00:09:36.866 "write": true, 00:09:36.866 "unmap": true, 00:09:36.866 "flush": true, 00:09:36.866 "reset": true, 00:09:36.866 "nvme_admin": false, 00:09:36.866 "nvme_io": false, 00:09:36.866 "nvme_io_md": false, 00:09:36.866 "write_zeroes": true, 00:09:36.866 "zcopy": true, 00:09:36.866 "get_zone_info": false, 00:09:36.866 "zone_management": false, 00:09:36.867 "zone_append": false, 00:09:36.867 "compare": false, 00:09:36.867 "compare_and_write": false, 00:09:36.867 "abort": true, 00:09:36.867 "seek_hole": false, 00:09:36.867 "seek_data": false, 00:09:36.867 "copy": true, 00:09:36.867 "nvme_iov_md": false 00:09:36.867 }, 00:09:36.867 "memory_domains": [ 00:09:36.867 { 00:09:36.867 "dma_device_id": "system", 00:09:36.867 "dma_device_type": 1 00:09:36.867 }, 00:09:36.867 { 00:09:36.867 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:36.867 "dma_device_type": 2 00:09:36.867 } 00:09:36.867 ], 00:09:36.867 "driver_specific": {} 00:09:36.867 } 00:09:36.867 ] 00:09:36.867 17:31:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:36.867 17:31:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:09:36.867 17:31:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:09:36.867 17:31:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:36.867 17:31:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 3 00:09:36.867 17:31:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:36.867 17:31:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:36.867 17:31:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:36.867 17:31:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:36.867 17:31:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:36.867 17:31:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:36.867 17:31:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:36.867 17:31:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:36.867 17:31:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:36.867 17:31:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:36.867 17:31:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:36.867 17:31:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:36.867 17:31:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:36.867 17:31:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:36.867 17:31:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:36.867 "name": "Existed_Raid", 00:09:36.867 "uuid": "b930a285-a867-406b-9cb3-7c03032f3fd4", 00:09:36.867 "strip_size_kb": 0, 00:09:36.867 "state": "online", 00:09:36.867 "raid_level": "raid1", 00:09:36.867 "superblock": true, 00:09:36.867 "num_base_bdevs": 3, 00:09:36.867 "num_base_bdevs_discovered": 3, 00:09:36.867 "num_base_bdevs_operational": 3, 00:09:36.867 "base_bdevs_list": [ 00:09:36.867 { 00:09:36.867 "name": "BaseBdev1", 00:09:36.867 "uuid": "1dd7a0ad-c398-4810-a5fa-86103c5068c0", 00:09:36.867 "is_configured": true, 00:09:36.867 "data_offset": 2048, 00:09:36.867 "data_size": 63488 00:09:36.867 }, 00:09:36.867 { 00:09:36.867 "name": "BaseBdev2", 00:09:36.867 "uuid": "94244386-eef5-4952-a87c-3a8da0da35c0", 00:09:36.867 "is_configured": true, 00:09:36.867 "data_offset": 2048, 00:09:36.867 "data_size": 63488 00:09:36.867 }, 00:09:36.867 { 00:09:36.867 "name": "BaseBdev3", 00:09:36.867 "uuid": "bc09e990-a64f-4a3a-ac0a-2e3ddb1f2ad7", 00:09:36.867 "is_configured": true, 00:09:36.867 "data_offset": 2048, 00:09:36.867 "data_size": 63488 00:09:36.867 } 00:09:36.867 ] 00:09:36.867 }' 00:09:36.867 17:31:07 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:36.867 17:31:07 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:37.128 17:31:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:09:37.128 17:31:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:09:37.128 17:31:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:37.128 17:31:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:37.128 17:31:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:09:37.128 17:31:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:37.128 17:31:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:09:37.128 17:31:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:37.128 17:31:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:37.128 17:31:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:37.128 [2024-11-27 17:31:08.264457] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:37.128 17:31:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:37.128 17:31:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:37.128 "name": "Existed_Raid", 00:09:37.128 "aliases": [ 00:09:37.128 "b930a285-a867-406b-9cb3-7c03032f3fd4" 00:09:37.128 ], 00:09:37.128 "product_name": "Raid Volume", 00:09:37.128 "block_size": 512, 00:09:37.128 "num_blocks": 63488, 00:09:37.128 "uuid": "b930a285-a867-406b-9cb3-7c03032f3fd4", 00:09:37.128 "assigned_rate_limits": { 00:09:37.128 "rw_ios_per_sec": 0, 00:09:37.128 "rw_mbytes_per_sec": 0, 00:09:37.128 "r_mbytes_per_sec": 0, 00:09:37.128 "w_mbytes_per_sec": 0 00:09:37.128 }, 00:09:37.128 "claimed": false, 00:09:37.128 "zoned": false, 00:09:37.128 "supported_io_types": { 00:09:37.128 "read": true, 00:09:37.128 "write": true, 00:09:37.128 "unmap": false, 00:09:37.128 "flush": false, 00:09:37.128 "reset": true, 00:09:37.128 "nvme_admin": false, 00:09:37.128 "nvme_io": false, 00:09:37.128 "nvme_io_md": false, 00:09:37.128 "write_zeroes": true, 00:09:37.128 "zcopy": false, 00:09:37.128 "get_zone_info": false, 00:09:37.128 "zone_management": false, 00:09:37.128 "zone_append": false, 00:09:37.128 "compare": false, 00:09:37.128 "compare_and_write": false, 00:09:37.128 "abort": false, 00:09:37.128 "seek_hole": false, 00:09:37.128 "seek_data": false, 00:09:37.128 "copy": false, 00:09:37.128 "nvme_iov_md": false 00:09:37.128 }, 00:09:37.128 "memory_domains": [ 00:09:37.128 { 00:09:37.128 "dma_device_id": "system", 00:09:37.128 "dma_device_type": 1 00:09:37.128 }, 00:09:37.128 { 00:09:37.128 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:37.128 "dma_device_type": 2 00:09:37.128 }, 00:09:37.128 { 00:09:37.128 "dma_device_id": "system", 00:09:37.128 "dma_device_type": 1 00:09:37.128 }, 00:09:37.128 { 00:09:37.128 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:37.128 "dma_device_type": 2 00:09:37.128 }, 00:09:37.128 { 00:09:37.128 "dma_device_id": "system", 00:09:37.128 "dma_device_type": 1 00:09:37.128 }, 00:09:37.128 { 00:09:37.128 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:37.128 "dma_device_type": 2 00:09:37.128 } 00:09:37.128 ], 00:09:37.128 "driver_specific": { 00:09:37.128 "raid": { 00:09:37.128 "uuid": "b930a285-a867-406b-9cb3-7c03032f3fd4", 00:09:37.128 "strip_size_kb": 0, 00:09:37.128 "state": "online", 00:09:37.128 "raid_level": "raid1", 00:09:37.128 "superblock": true, 00:09:37.128 "num_base_bdevs": 3, 00:09:37.128 "num_base_bdevs_discovered": 3, 00:09:37.128 "num_base_bdevs_operational": 3, 00:09:37.128 "base_bdevs_list": [ 00:09:37.128 { 00:09:37.128 "name": "BaseBdev1", 00:09:37.128 "uuid": "1dd7a0ad-c398-4810-a5fa-86103c5068c0", 00:09:37.128 "is_configured": true, 00:09:37.128 "data_offset": 2048, 00:09:37.128 "data_size": 63488 00:09:37.128 }, 00:09:37.128 { 00:09:37.128 "name": "BaseBdev2", 00:09:37.128 "uuid": "94244386-eef5-4952-a87c-3a8da0da35c0", 00:09:37.128 "is_configured": true, 00:09:37.128 "data_offset": 2048, 00:09:37.128 "data_size": 63488 00:09:37.128 }, 00:09:37.128 { 00:09:37.128 "name": "BaseBdev3", 00:09:37.128 "uuid": "bc09e990-a64f-4a3a-ac0a-2e3ddb1f2ad7", 00:09:37.128 "is_configured": true, 00:09:37.128 "data_offset": 2048, 00:09:37.128 "data_size": 63488 00:09:37.128 } 00:09:37.128 ] 00:09:37.128 } 00:09:37.128 } 00:09:37.128 }' 00:09:37.128 17:31:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:37.389 17:31:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:09:37.389 BaseBdev2 00:09:37.389 BaseBdev3' 00:09:37.389 17:31:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:37.389 17:31:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:37.389 17:31:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:37.389 17:31:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:09:37.389 17:31:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:37.389 17:31:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:37.389 17:31:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:37.389 17:31:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:37.389 17:31:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:37.389 17:31:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:37.389 17:31:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:37.389 17:31:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:09:37.389 17:31:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:37.389 17:31:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:37.389 17:31:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:37.389 17:31:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:37.389 17:31:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:37.389 17:31:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:37.389 17:31:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:37.389 17:31:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:09:37.389 17:31:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:37.389 17:31:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:37.389 17:31:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:37.389 17:31:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:37.389 17:31:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:37.389 17:31:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:37.389 17:31:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:09:37.389 17:31:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:37.389 17:31:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:37.389 [2024-11-27 17:31:08.535762] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:09:37.389 17:31:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:37.389 17:31:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:09:37.389 17:31:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:09:37.389 17:31:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:09:37.389 17:31:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@199 -- # return 0 00:09:37.389 17:31:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:09:37.389 17:31:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 2 00:09:37.390 17:31:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:37.390 17:31:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:37.390 17:31:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:37.390 17:31:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:37.390 17:31:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:09:37.390 17:31:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:37.390 17:31:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:37.390 17:31:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:37.390 17:31:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:37.390 17:31:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:37.390 17:31:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:37.390 17:31:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:37.390 17:31:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:37.650 17:31:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:37.650 17:31:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:37.650 "name": "Existed_Raid", 00:09:37.650 "uuid": "b930a285-a867-406b-9cb3-7c03032f3fd4", 00:09:37.650 "strip_size_kb": 0, 00:09:37.650 "state": "online", 00:09:37.650 "raid_level": "raid1", 00:09:37.650 "superblock": true, 00:09:37.650 "num_base_bdevs": 3, 00:09:37.650 "num_base_bdevs_discovered": 2, 00:09:37.650 "num_base_bdevs_operational": 2, 00:09:37.650 "base_bdevs_list": [ 00:09:37.650 { 00:09:37.650 "name": null, 00:09:37.650 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:37.650 "is_configured": false, 00:09:37.650 "data_offset": 0, 00:09:37.650 "data_size": 63488 00:09:37.650 }, 00:09:37.650 { 00:09:37.650 "name": "BaseBdev2", 00:09:37.650 "uuid": "94244386-eef5-4952-a87c-3a8da0da35c0", 00:09:37.650 "is_configured": true, 00:09:37.650 "data_offset": 2048, 00:09:37.650 "data_size": 63488 00:09:37.650 }, 00:09:37.650 { 00:09:37.650 "name": "BaseBdev3", 00:09:37.650 "uuid": "bc09e990-a64f-4a3a-ac0a-2e3ddb1f2ad7", 00:09:37.650 "is_configured": true, 00:09:37.650 "data_offset": 2048, 00:09:37.650 "data_size": 63488 00:09:37.650 } 00:09:37.650 ] 00:09:37.650 }' 00:09:37.650 17:31:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:37.650 17:31:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:37.911 17:31:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:09:37.911 17:31:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:37.911 17:31:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:37.911 17:31:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:37.911 17:31:08 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:37.911 17:31:08 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:09:37.911 17:31:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:37.911 17:31:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:09:37.911 17:31:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:09:37.911 17:31:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:09:37.911 17:31:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:37.911 17:31:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:37.911 [2024-11-27 17:31:09.023098] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:09:37.911 17:31:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:37.911 17:31:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:09:37.911 17:31:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:37.911 17:31:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:37.911 17:31:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:37.911 17:31:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:37.911 17:31:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:09:37.911 17:31:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:37.911 17:31:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:09:37.911 17:31:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:09:37.911 17:31:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:09:37.911 17:31:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:37.911 17:31:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:38.174 [2024-11-27 17:31:09.103571] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:09:38.174 [2024-11-27 17:31:09.103731] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:38.174 [2024-11-27 17:31:09.124810] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:38.174 [2024-11-27 17:31:09.124865] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:38.174 [2024-11-27 17:31:09.124893] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:09:38.174 17:31:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:38.174 17:31:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:09:38.174 17:31:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:09:38.174 17:31:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:38.174 17:31:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:38.174 17:31:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:09:38.174 17:31:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:38.174 17:31:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:38.174 17:31:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:09:38.174 17:31:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:09:38.174 17:31:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 3 -gt 2 ']' 00:09:38.174 17:31:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:09:38.174 17:31:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:38.174 17:31:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:09:38.174 17:31:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:38.174 17:31:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:38.174 BaseBdev2 00:09:38.175 17:31:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:38.175 17:31:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:09:38.175 17:31:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:09:38.175 17:31:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:09:38.175 17:31:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:09:38.175 17:31:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:09:38.175 17:31:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:09:38.175 17:31:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:09:38.175 17:31:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:38.175 17:31:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:38.175 17:31:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:38.175 17:31:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:09:38.175 17:31:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:38.175 17:31:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:38.175 [ 00:09:38.175 { 00:09:38.175 "name": "BaseBdev2", 00:09:38.175 "aliases": [ 00:09:38.175 "4be23e28-10a5-4d20-a69a-28c41b6f6559" 00:09:38.175 ], 00:09:38.175 "product_name": "Malloc disk", 00:09:38.175 "block_size": 512, 00:09:38.175 "num_blocks": 65536, 00:09:38.175 "uuid": "4be23e28-10a5-4d20-a69a-28c41b6f6559", 00:09:38.175 "assigned_rate_limits": { 00:09:38.175 "rw_ios_per_sec": 0, 00:09:38.175 "rw_mbytes_per_sec": 0, 00:09:38.175 "r_mbytes_per_sec": 0, 00:09:38.175 "w_mbytes_per_sec": 0 00:09:38.175 }, 00:09:38.175 "claimed": false, 00:09:38.175 "zoned": false, 00:09:38.175 "supported_io_types": { 00:09:38.175 "read": true, 00:09:38.175 "write": true, 00:09:38.175 "unmap": true, 00:09:38.175 "flush": true, 00:09:38.175 "reset": true, 00:09:38.175 "nvme_admin": false, 00:09:38.175 "nvme_io": false, 00:09:38.175 "nvme_io_md": false, 00:09:38.175 "write_zeroes": true, 00:09:38.175 "zcopy": true, 00:09:38.175 "get_zone_info": false, 00:09:38.175 "zone_management": false, 00:09:38.175 "zone_append": false, 00:09:38.175 "compare": false, 00:09:38.175 "compare_and_write": false, 00:09:38.175 "abort": true, 00:09:38.175 "seek_hole": false, 00:09:38.175 "seek_data": false, 00:09:38.175 "copy": true, 00:09:38.175 "nvme_iov_md": false 00:09:38.175 }, 00:09:38.175 "memory_domains": [ 00:09:38.175 { 00:09:38.175 "dma_device_id": "system", 00:09:38.175 "dma_device_type": 1 00:09:38.175 }, 00:09:38.175 { 00:09:38.175 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:38.175 "dma_device_type": 2 00:09:38.175 } 00:09:38.175 ], 00:09:38.175 "driver_specific": {} 00:09:38.175 } 00:09:38.175 ] 00:09:38.175 17:31:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:38.175 17:31:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:09:38.175 17:31:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:09:38.175 17:31:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:38.175 17:31:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:09:38.175 17:31:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:38.175 17:31:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:38.175 BaseBdev3 00:09:38.175 17:31:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:38.175 17:31:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:09:38.175 17:31:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:09:38.175 17:31:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:09:38.175 17:31:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:09:38.175 17:31:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:09:38.175 17:31:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:09:38.175 17:31:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:09:38.175 17:31:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:38.175 17:31:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:38.175 17:31:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:38.175 17:31:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:09:38.175 17:31:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:38.175 17:31:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:38.175 [ 00:09:38.175 { 00:09:38.175 "name": "BaseBdev3", 00:09:38.175 "aliases": [ 00:09:38.175 "cd354491-def0-478d-be8e-0225ce81133d" 00:09:38.175 ], 00:09:38.175 "product_name": "Malloc disk", 00:09:38.175 "block_size": 512, 00:09:38.175 "num_blocks": 65536, 00:09:38.175 "uuid": "cd354491-def0-478d-be8e-0225ce81133d", 00:09:38.175 "assigned_rate_limits": { 00:09:38.175 "rw_ios_per_sec": 0, 00:09:38.175 "rw_mbytes_per_sec": 0, 00:09:38.175 "r_mbytes_per_sec": 0, 00:09:38.175 "w_mbytes_per_sec": 0 00:09:38.175 }, 00:09:38.175 "claimed": false, 00:09:38.175 "zoned": false, 00:09:38.175 "supported_io_types": { 00:09:38.175 "read": true, 00:09:38.175 "write": true, 00:09:38.175 "unmap": true, 00:09:38.175 "flush": true, 00:09:38.175 "reset": true, 00:09:38.175 "nvme_admin": false, 00:09:38.175 "nvme_io": false, 00:09:38.175 "nvme_io_md": false, 00:09:38.175 "write_zeroes": true, 00:09:38.175 "zcopy": true, 00:09:38.175 "get_zone_info": false, 00:09:38.175 "zone_management": false, 00:09:38.175 "zone_append": false, 00:09:38.175 "compare": false, 00:09:38.175 "compare_and_write": false, 00:09:38.175 "abort": true, 00:09:38.175 "seek_hole": false, 00:09:38.175 "seek_data": false, 00:09:38.175 "copy": true, 00:09:38.175 "nvme_iov_md": false 00:09:38.175 }, 00:09:38.175 "memory_domains": [ 00:09:38.175 { 00:09:38.175 "dma_device_id": "system", 00:09:38.175 "dma_device_type": 1 00:09:38.175 }, 00:09:38.175 { 00:09:38.175 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:38.175 "dma_device_type": 2 00:09:38.175 } 00:09:38.175 ], 00:09:38.175 "driver_specific": {} 00:09:38.175 } 00:09:38.175 ] 00:09:38.175 17:31:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:38.175 17:31:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:09:38.175 17:31:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:09:38.175 17:31:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:09:38.175 17:31:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:09:38.175 17:31:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:38.175 17:31:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:38.175 [2024-11-27 17:31:09.297445] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:38.175 [2024-11-27 17:31:09.297580] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:38.175 [2024-11-27 17:31:09.297622] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:38.175 [2024-11-27 17:31:09.299690] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:38.175 17:31:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:38.175 17:31:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:38.175 17:31:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:38.175 17:31:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:38.175 17:31:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:38.175 17:31:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:38.175 17:31:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:38.175 17:31:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:38.175 17:31:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:38.175 17:31:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:38.175 17:31:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:38.175 17:31:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:38.175 17:31:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:38.175 17:31:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:38.175 17:31:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:38.175 17:31:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:38.175 17:31:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:38.175 "name": "Existed_Raid", 00:09:38.175 "uuid": "b739e458-8a58-4ee4-aeb8-e3a5d9fe10eb", 00:09:38.175 "strip_size_kb": 0, 00:09:38.175 "state": "configuring", 00:09:38.175 "raid_level": "raid1", 00:09:38.175 "superblock": true, 00:09:38.175 "num_base_bdevs": 3, 00:09:38.175 "num_base_bdevs_discovered": 2, 00:09:38.175 "num_base_bdevs_operational": 3, 00:09:38.175 "base_bdevs_list": [ 00:09:38.176 { 00:09:38.176 "name": "BaseBdev1", 00:09:38.176 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:38.176 "is_configured": false, 00:09:38.176 "data_offset": 0, 00:09:38.176 "data_size": 0 00:09:38.176 }, 00:09:38.176 { 00:09:38.176 "name": "BaseBdev2", 00:09:38.176 "uuid": "4be23e28-10a5-4d20-a69a-28c41b6f6559", 00:09:38.176 "is_configured": true, 00:09:38.176 "data_offset": 2048, 00:09:38.176 "data_size": 63488 00:09:38.176 }, 00:09:38.176 { 00:09:38.176 "name": "BaseBdev3", 00:09:38.176 "uuid": "cd354491-def0-478d-be8e-0225ce81133d", 00:09:38.176 "is_configured": true, 00:09:38.176 "data_offset": 2048, 00:09:38.176 "data_size": 63488 00:09:38.176 } 00:09:38.176 ] 00:09:38.176 }' 00:09:38.176 17:31:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:38.176 17:31:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:38.745 17:31:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:09:38.745 17:31:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:38.745 17:31:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:38.745 [2024-11-27 17:31:09.764624] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:09:38.745 17:31:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:38.745 17:31:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:38.745 17:31:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:38.745 17:31:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:38.745 17:31:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:38.745 17:31:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:38.745 17:31:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:38.745 17:31:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:38.745 17:31:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:38.745 17:31:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:38.745 17:31:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:38.745 17:31:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:38.745 17:31:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:38.745 17:31:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:38.745 17:31:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:38.745 17:31:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:38.745 17:31:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:38.745 "name": "Existed_Raid", 00:09:38.745 "uuid": "b739e458-8a58-4ee4-aeb8-e3a5d9fe10eb", 00:09:38.745 "strip_size_kb": 0, 00:09:38.745 "state": "configuring", 00:09:38.746 "raid_level": "raid1", 00:09:38.746 "superblock": true, 00:09:38.746 "num_base_bdevs": 3, 00:09:38.746 "num_base_bdevs_discovered": 1, 00:09:38.746 "num_base_bdevs_operational": 3, 00:09:38.746 "base_bdevs_list": [ 00:09:38.746 { 00:09:38.746 "name": "BaseBdev1", 00:09:38.746 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:38.746 "is_configured": false, 00:09:38.746 "data_offset": 0, 00:09:38.746 "data_size": 0 00:09:38.746 }, 00:09:38.746 { 00:09:38.746 "name": null, 00:09:38.746 "uuid": "4be23e28-10a5-4d20-a69a-28c41b6f6559", 00:09:38.746 "is_configured": false, 00:09:38.746 "data_offset": 0, 00:09:38.746 "data_size": 63488 00:09:38.746 }, 00:09:38.746 { 00:09:38.746 "name": "BaseBdev3", 00:09:38.746 "uuid": "cd354491-def0-478d-be8e-0225ce81133d", 00:09:38.746 "is_configured": true, 00:09:38.746 "data_offset": 2048, 00:09:38.746 "data_size": 63488 00:09:38.746 } 00:09:38.746 ] 00:09:38.746 }' 00:09:38.746 17:31:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:38.746 17:31:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:39.316 17:31:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:09:39.317 17:31:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:39.317 17:31:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:39.317 17:31:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:39.317 17:31:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:39.317 17:31:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:09:39.317 17:31:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:09:39.317 17:31:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:39.317 17:31:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:39.317 [2024-11-27 17:31:10.248602] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:39.317 BaseBdev1 00:09:39.317 17:31:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:39.317 17:31:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:09:39.317 17:31:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:09:39.317 17:31:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:09:39.317 17:31:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:09:39.317 17:31:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:09:39.317 17:31:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:09:39.317 17:31:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:09:39.317 17:31:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:39.317 17:31:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:39.317 17:31:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:39.317 17:31:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:09:39.317 17:31:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:39.317 17:31:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:39.317 [ 00:09:39.317 { 00:09:39.317 "name": "BaseBdev1", 00:09:39.317 "aliases": [ 00:09:39.317 "f88684cb-076e-41f9-854d-931023cf205e" 00:09:39.317 ], 00:09:39.317 "product_name": "Malloc disk", 00:09:39.317 "block_size": 512, 00:09:39.317 "num_blocks": 65536, 00:09:39.317 "uuid": "f88684cb-076e-41f9-854d-931023cf205e", 00:09:39.317 "assigned_rate_limits": { 00:09:39.317 "rw_ios_per_sec": 0, 00:09:39.317 "rw_mbytes_per_sec": 0, 00:09:39.317 "r_mbytes_per_sec": 0, 00:09:39.317 "w_mbytes_per_sec": 0 00:09:39.317 }, 00:09:39.317 "claimed": true, 00:09:39.317 "claim_type": "exclusive_write", 00:09:39.317 "zoned": false, 00:09:39.317 "supported_io_types": { 00:09:39.317 "read": true, 00:09:39.317 "write": true, 00:09:39.317 "unmap": true, 00:09:39.317 "flush": true, 00:09:39.317 "reset": true, 00:09:39.317 "nvme_admin": false, 00:09:39.317 "nvme_io": false, 00:09:39.317 "nvme_io_md": false, 00:09:39.317 "write_zeroes": true, 00:09:39.317 "zcopy": true, 00:09:39.317 "get_zone_info": false, 00:09:39.317 "zone_management": false, 00:09:39.317 "zone_append": false, 00:09:39.317 "compare": false, 00:09:39.317 "compare_and_write": false, 00:09:39.317 "abort": true, 00:09:39.317 "seek_hole": false, 00:09:39.317 "seek_data": false, 00:09:39.317 "copy": true, 00:09:39.317 "nvme_iov_md": false 00:09:39.317 }, 00:09:39.317 "memory_domains": [ 00:09:39.317 { 00:09:39.317 "dma_device_id": "system", 00:09:39.317 "dma_device_type": 1 00:09:39.317 }, 00:09:39.317 { 00:09:39.317 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:39.317 "dma_device_type": 2 00:09:39.317 } 00:09:39.317 ], 00:09:39.317 "driver_specific": {} 00:09:39.317 } 00:09:39.317 ] 00:09:39.317 17:31:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:39.317 17:31:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:09:39.317 17:31:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:39.317 17:31:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:39.317 17:31:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:39.317 17:31:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:39.317 17:31:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:39.317 17:31:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:39.317 17:31:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:39.317 17:31:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:39.317 17:31:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:39.317 17:31:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:39.317 17:31:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:39.317 17:31:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:39.317 17:31:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:39.317 17:31:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:39.317 17:31:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:39.317 17:31:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:39.317 "name": "Existed_Raid", 00:09:39.317 "uuid": "b739e458-8a58-4ee4-aeb8-e3a5d9fe10eb", 00:09:39.317 "strip_size_kb": 0, 00:09:39.317 "state": "configuring", 00:09:39.317 "raid_level": "raid1", 00:09:39.317 "superblock": true, 00:09:39.317 "num_base_bdevs": 3, 00:09:39.317 "num_base_bdevs_discovered": 2, 00:09:39.317 "num_base_bdevs_operational": 3, 00:09:39.317 "base_bdevs_list": [ 00:09:39.317 { 00:09:39.317 "name": "BaseBdev1", 00:09:39.317 "uuid": "f88684cb-076e-41f9-854d-931023cf205e", 00:09:39.317 "is_configured": true, 00:09:39.317 "data_offset": 2048, 00:09:39.317 "data_size": 63488 00:09:39.317 }, 00:09:39.317 { 00:09:39.317 "name": null, 00:09:39.317 "uuid": "4be23e28-10a5-4d20-a69a-28c41b6f6559", 00:09:39.317 "is_configured": false, 00:09:39.317 "data_offset": 0, 00:09:39.317 "data_size": 63488 00:09:39.317 }, 00:09:39.317 { 00:09:39.317 "name": "BaseBdev3", 00:09:39.317 "uuid": "cd354491-def0-478d-be8e-0225ce81133d", 00:09:39.317 "is_configured": true, 00:09:39.317 "data_offset": 2048, 00:09:39.317 "data_size": 63488 00:09:39.317 } 00:09:39.317 ] 00:09:39.317 }' 00:09:39.317 17:31:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:39.317 17:31:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:39.577 17:31:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:39.577 17:31:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:09:39.577 17:31:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:39.577 17:31:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:39.577 17:31:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:39.837 17:31:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:09:39.837 17:31:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:09:39.837 17:31:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:39.837 17:31:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:39.837 [2024-11-27 17:31:10.783734] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:09:39.837 17:31:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:39.837 17:31:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:39.837 17:31:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:39.837 17:31:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:39.837 17:31:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:39.837 17:31:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:39.837 17:31:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:39.837 17:31:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:39.837 17:31:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:39.837 17:31:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:39.837 17:31:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:39.837 17:31:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:39.837 17:31:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:39.837 17:31:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:39.837 17:31:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:39.837 17:31:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:39.837 17:31:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:39.837 "name": "Existed_Raid", 00:09:39.837 "uuid": "b739e458-8a58-4ee4-aeb8-e3a5d9fe10eb", 00:09:39.837 "strip_size_kb": 0, 00:09:39.837 "state": "configuring", 00:09:39.837 "raid_level": "raid1", 00:09:39.837 "superblock": true, 00:09:39.837 "num_base_bdevs": 3, 00:09:39.837 "num_base_bdevs_discovered": 1, 00:09:39.837 "num_base_bdevs_operational": 3, 00:09:39.837 "base_bdevs_list": [ 00:09:39.837 { 00:09:39.837 "name": "BaseBdev1", 00:09:39.837 "uuid": "f88684cb-076e-41f9-854d-931023cf205e", 00:09:39.837 "is_configured": true, 00:09:39.837 "data_offset": 2048, 00:09:39.837 "data_size": 63488 00:09:39.837 }, 00:09:39.837 { 00:09:39.837 "name": null, 00:09:39.837 "uuid": "4be23e28-10a5-4d20-a69a-28c41b6f6559", 00:09:39.837 "is_configured": false, 00:09:39.837 "data_offset": 0, 00:09:39.837 "data_size": 63488 00:09:39.837 }, 00:09:39.837 { 00:09:39.837 "name": null, 00:09:39.837 "uuid": "cd354491-def0-478d-be8e-0225ce81133d", 00:09:39.837 "is_configured": false, 00:09:39.837 "data_offset": 0, 00:09:39.837 "data_size": 63488 00:09:39.837 } 00:09:39.837 ] 00:09:39.837 }' 00:09:39.837 17:31:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:39.837 17:31:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:40.098 17:31:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:09:40.098 17:31:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:40.098 17:31:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:40.098 17:31:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:40.098 17:31:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:40.365 17:31:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:09:40.365 17:31:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:09:40.365 17:31:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:40.365 17:31:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:40.365 [2024-11-27 17:31:11.306864] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:40.365 17:31:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:40.365 17:31:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:40.365 17:31:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:40.365 17:31:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:40.365 17:31:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:40.365 17:31:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:40.365 17:31:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:40.365 17:31:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:40.365 17:31:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:40.365 17:31:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:40.365 17:31:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:40.365 17:31:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:40.365 17:31:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:40.365 17:31:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:40.365 17:31:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:40.365 17:31:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:40.365 17:31:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:40.365 "name": "Existed_Raid", 00:09:40.365 "uuid": "b739e458-8a58-4ee4-aeb8-e3a5d9fe10eb", 00:09:40.365 "strip_size_kb": 0, 00:09:40.365 "state": "configuring", 00:09:40.365 "raid_level": "raid1", 00:09:40.365 "superblock": true, 00:09:40.365 "num_base_bdevs": 3, 00:09:40.365 "num_base_bdevs_discovered": 2, 00:09:40.365 "num_base_bdevs_operational": 3, 00:09:40.365 "base_bdevs_list": [ 00:09:40.365 { 00:09:40.365 "name": "BaseBdev1", 00:09:40.365 "uuid": "f88684cb-076e-41f9-854d-931023cf205e", 00:09:40.365 "is_configured": true, 00:09:40.365 "data_offset": 2048, 00:09:40.365 "data_size": 63488 00:09:40.365 }, 00:09:40.365 { 00:09:40.365 "name": null, 00:09:40.365 "uuid": "4be23e28-10a5-4d20-a69a-28c41b6f6559", 00:09:40.365 "is_configured": false, 00:09:40.365 "data_offset": 0, 00:09:40.365 "data_size": 63488 00:09:40.365 }, 00:09:40.365 { 00:09:40.365 "name": "BaseBdev3", 00:09:40.365 "uuid": "cd354491-def0-478d-be8e-0225ce81133d", 00:09:40.365 "is_configured": true, 00:09:40.365 "data_offset": 2048, 00:09:40.365 "data_size": 63488 00:09:40.365 } 00:09:40.365 ] 00:09:40.365 }' 00:09:40.365 17:31:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:40.365 17:31:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:40.643 17:31:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:40.643 17:31:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:40.643 17:31:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:40.643 17:31:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:09:40.643 17:31:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:40.643 17:31:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:09:40.643 17:31:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:09:40.643 17:31:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:40.643 17:31:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:40.643 [2024-11-27 17:31:11.806075] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:09:40.643 17:31:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:40.920 17:31:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:40.920 17:31:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:40.920 17:31:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:40.920 17:31:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:40.920 17:31:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:40.920 17:31:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:40.920 17:31:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:40.920 17:31:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:40.920 17:31:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:40.920 17:31:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:40.920 17:31:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:40.920 17:31:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:40.920 17:31:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:40.920 17:31:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:40.920 17:31:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:40.920 17:31:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:40.920 "name": "Existed_Raid", 00:09:40.920 "uuid": "b739e458-8a58-4ee4-aeb8-e3a5d9fe10eb", 00:09:40.920 "strip_size_kb": 0, 00:09:40.920 "state": "configuring", 00:09:40.920 "raid_level": "raid1", 00:09:40.920 "superblock": true, 00:09:40.920 "num_base_bdevs": 3, 00:09:40.920 "num_base_bdevs_discovered": 1, 00:09:40.920 "num_base_bdevs_operational": 3, 00:09:40.920 "base_bdevs_list": [ 00:09:40.920 { 00:09:40.920 "name": null, 00:09:40.920 "uuid": "f88684cb-076e-41f9-854d-931023cf205e", 00:09:40.920 "is_configured": false, 00:09:40.920 "data_offset": 0, 00:09:40.920 "data_size": 63488 00:09:40.920 }, 00:09:40.920 { 00:09:40.920 "name": null, 00:09:40.920 "uuid": "4be23e28-10a5-4d20-a69a-28c41b6f6559", 00:09:40.920 "is_configured": false, 00:09:40.920 "data_offset": 0, 00:09:40.920 "data_size": 63488 00:09:40.920 }, 00:09:40.920 { 00:09:40.920 "name": "BaseBdev3", 00:09:40.920 "uuid": "cd354491-def0-478d-be8e-0225ce81133d", 00:09:40.920 "is_configured": true, 00:09:40.920 "data_offset": 2048, 00:09:40.920 "data_size": 63488 00:09:40.920 } 00:09:40.920 ] 00:09:40.920 }' 00:09:40.920 17:31:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:40.920 17:31:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:41.181 17:31:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:41.181 17:31:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:41.181 17:31:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:41.181 17:31:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:09:41.181 17:31:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:41.181 17:31:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:09:41.181 17:31:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:09:41.181 17:31:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:41.181 17:31:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:41.181 [2024-11-27 17:31:12.308857] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:41.181 17:31:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:41.181 17:31:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 3 00:09:41.181 17:31:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:41.181 17:31:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:41.181 17:31:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:41.181 17:31:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:41.181 17:31:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:41.181 17:31:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:41.181 17:31:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:41.181 17:31:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:41.181 17:31:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:41.181 17:31:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:41.181 17:31:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:41.181 17:31:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:41.181 17:31:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:41.181 17:31:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:41.181 17:31:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:41.181 "name": "Existed_Raid", 00:09:41.181 "uuid": "b739e458-8a58-4ee4-aeb8-e3a5d9fe10eb", 00:09:41.181 "strip_size_kb": 0, 00:09:41.181 "state": "configuring", 00:09:41.181 "raid_level": "raid1", 00:09:41.181 "superblock": true, 00:09:41.181 "num_base_bdevs": 3, 00:09:41.181 "num_base_bdevs_discovered": 2, 00:09:41.181 "num_base_bdevs_operational": 3, 00:09:41.181 "base_bdevs_list": [ 00:09:41.181 { 00:09:41.181 "name": null, 00:09:41.181 "uuid": "f88684cb-076e-41f9-854d-931023cf205e", 00:09:41.181 "is_configured": false, 00:09:41.181 "data_offset": 0, 00:09:41.181 "data_size": 63488 00:09:41.181 }, 00:09:41.181 { 00:09:41.181 "name": "BaseBdev2", 00:09:41.181 "uuid": "4be23e28-10a5-4d20-a69a-28c41b6f6559", 00:09:41.181 "is_configured": true, 00:09:41.181 "data_offset": 2048, 00:09:41.181 "data_size": 63488 00:09:41.181 }, 00:09:41.181 { 00:09:41.181 "name": "BaseBdev3", 00:09:41.181 "uuid": "cd354491-def0-478d-be8e-0225ce81133d", 00:09:41.181 "is_configured": true, 00:09:41.181 "data_offset": 2048, 00:09:41.181 "data_size": 63488 00:09:41.181 } 00:09:41.181 ] 00:09:41.181 }' 00:09:41.181 17:31:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:41.181 17:31:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:41.752 17:31:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:41.752 17:31:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:41.752 17:31:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:41.752 17:31:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:09:41.752 17:31:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:41.752 17:31:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:09:41.752 17:31:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:41.752 17:31:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:41.752 17:31:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:41.752 17:31:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:09:41.752 17:31:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:41.752 17:31:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u f88684cb-076e-41f9-854d-931023cf205e 00:09:41.752 17:31:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:41.752 17:31:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:41.752 [2024-11-27 17:31:12.872710] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:09:41.752 [2024-11-27 17:31:12.872891] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:09:41.752 [2024-11-27 17:31:12.872904] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:09:41.752 [2024-11-27 17:31:12.873200] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002870 00:09:41.752 [2024-11-27 17:31:12.873346] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:09:41.752 [2024-11-27 17:31:12.873362] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001c80 00:09:41.752 [2024-11-27 17:31:12.873472] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:41.752 NewBaseBdev 00:09:41.752 17:31:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:41.752 17:31:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:09:41.752 17:31:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=NewBaseBdev 00:09:41.752 17:31:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:09:41.752 17:31:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:09:41.752 17:31:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:09:41.752 17:31:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:09:41.752 17:31:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:09:41.752 17:31:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:41.752 17:31:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:41.752 17:31:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:41.752 17:31:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:09:41.752 17:31:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:41.752 17:31:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:41.752 [ 00:09:41.752 { 00:09:41.752 "name": "NewBaseBdev", 00:09:41.752 "aliases": [ 00:09:41.752 "f88684cb-076e-41f9-854d-931023cf205e" 00:09:41.752 ], 00:09:41.752 "product_name": "Malloc disk", 00:09:41.752 "block_size": 512, 00:09:41.752 "num_blocks": 65536, 00:09:41.752 "uuid": "f88684cb-076e-41f9-854d-931023cf205e", 00:09:41.752 "assigned_rate_limits": { 00:09:41.752 "rw_ios_per_sec": 0, 00:09:41.752 "rw_mbytes_per_sec": 0, 00:09:41.752 "r_mbytes_per_sec": 0, 00:09:41.752 "w_mbytes_per_sec": 0 00:09:41.752 }, 00:09:41.752 "claimed": true, 00:09:41.752 "claim_type": "exclusive_write", 00:09:41.752 "zoned": false, 00:09:41.752 "supported_io_types": { 00:09:41.752 "read": true, 00:09:41.752 "write": true, 00:09:41.752 "unmap": true, 00:09:41.752 "flush": true, 00:09:41.752 "reset": true, 00:09:41.752 "nvme_admin": false, 00:09:41.752 "nvme_io": false, 00:09:41.752 "nvme_io_md": false, 00:09:41.752 "write_zeroes": true, 00:09:41.752 "zcopy": true, 00:09:41.752 "get_zone_info": false, 00:09:41.752 "zone_management": false, 00:09:41.752 "zone_append": false, 00:09:41.752 "compare": false, 00:09:41.752 "compare_and_write": false, 00:09:41.752 "abort": true, 00:09:41.752 "seek_hole": false, 00:09:41.752 "seek_data": false, 00:09:41.752 "copy": true, 00:09:41.752 "nvme_iov_md": false 00:09:41.752 }, 00:09:41.752 "memory_domains": [ 00:09:41.752 { 00:09:41.752 "dma_device_id": "system", 00:09:41.752 "dma_device_type": 1 00:09:41.752 }, 00:09:41.752 { 00:09:41.752 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:41.752 "dma_device_type": 2 00:09:41.752 } 00:09:41.752 ], 00:09:41.752 "driver_specific": {} 00:09:41.752 } 00:09:41.752 ] 00:09:41.752 17:31:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:41.752 17:31:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:09:41.752 17:31:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid1 0 3 00:09:41.752 17:31:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:41.752 17:31:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:41.752 17:31:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:41.752 17:31:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:41.752 17:31:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:41.752 17:31:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:41.752 17:31:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:41.752 17:31:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:41.752 17:31:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:41.752 17:31:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:41.752 17:31:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:41.752 17:31:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:41.752 17:31:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:41.752 17:31:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:42.012 17:31:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:42.012 "name": "Existed_Raid", 00:09:42.012 "uuid": "b739e458-8a58-4ee4-aeb8-e3a5d9fe10eb", 00:09:42.012 "strip_size_kb": 0, 00:09:42.012 "state": "online", 00:09:42.012 "raid_level": "raid1", 00:09:42.012 "superblock": true, 00:09:42.012 "num_base_bdevs": 3, 00:09:42.012 "num_base_bdevs_discovered": 3, 00:09:42.012 "num_base_bdevs_operational": 3, 00:09:42.012 "base_bdevs_list": [ 00:09:42.012 { 00:09:42.012 "name": "NewBaseBdev", 00:09:42.012 "uuid": "f88684cb-076e-41f9-854d-931023cf205e", 00:09:42.012 "is_configured": true, 00:09:42.012 "data_offset": 2048, 00:09:42.012 "data_size": 63488 00:09:42.012 }, 00:09:42.012 { 00:09:42.012 "name": "BaseBdev2", 00:09:42.012 "uuid": "4be23e28-10a5-4d20-a69a-28c41b6f6559", 00:09:42.012 "is_configured": true, 00:09:42.012 "data_offset": 2048, 00:09:42.012 "data_size": 63488 00:09:42.012 }, 00:09:42.012 { 00:09:42.012 "name": "BaseBdev3", 00:09:42.012 "uuid": "cd354491-def0-478d-be8e-0225ce81133d", 00:09:42.012 "is_configured": true, 00:09:42.012 "data_offset": 2048, 00:09:42.012 "data_size": 63488 00:09:42.012 } 00:09:42.012 ] 00:09:42.012 }' 00:09:42.012 17:31:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:42.012 17:31:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:42.272 17:31:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:09:42.273 17:31:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:09:42.273 17:31:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:42.273 17:31:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:42.273 17:31:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:09:42.273 17:31:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:42.273 17:31:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:09:42.273 17:31:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:42.273 17:31:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:42.273 17:31:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:42.273 [2024-11-27 17:31:13.380184] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:42.273 17:31:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:42.273 17:31:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:42.273 "name": "Existed_Raid", 00:09:42.273 "aliases": [ 00:09:42.273 "b739e458-8a58-4ee4-aeb8-e3a5d9fe10eb" 00:09:42.273 ], 00:09:42.273 "product_name": "Raid Volume", 00:09:42.273 "block_size": 512, 00:09:42.273 "num_blocks": 63488, 00:09:42.273 "uuid": "b739e458-8a58-4ee4-aeb8-e3a5d9fe10eb", 00:09:42.273 "assigned_rate_limits": { 00:09:42.273 "rw_ios_per_sec": 0, 00:09:42.273 "rw_mbytes_per_sec": 0, 00:09:42.273 "r_mbytes_per_sec": 0, 00:09:42.273 "w_mbytes_per_sec": 0 00:09:42.273 }, 00:09:42.273 "claimed": false, 00:09:42.273 "zoned": false, 00:09:42.273 "supported_io_types": { 00:09:42.273 "read": true, 00:09:42.273 "write": true, 00:09:42.273 "unmap": false, 00:09:42.273 "flush": false, 00:09:42.273 "reset": true, 00:09:42.273 "nvme_admin": false, 00:09:42.273 "nvme_io": false, 00:09:42.273 "nvme_io_md": false, 00:09:42.273 "write_zeroes": true, 00:09:42.273 "zcopy": false, 00:09:42.273 "get_zone_info": false, 00:09:42.273 "zone_management": false, 00:09:42.273 "zone_append": false, 00:09:42.273 "compare": false, 00:09:42.273 "compare_and_write": false, 00:09:42.273 "abort": false, 00:09:42.273 "seek_hole": false, 00:09:42.273 "seek_data": false, 00:09:42.273 "copy": false, 00:09:42.273 "nvme_iov_md": false 00:09:42.273 }, 00:09:42.273 "memory_domains": [ 00:09:42.273 { 00:09:42.273 "dma_device_id": "system", 00:09:42.273 "dma_device_type": 1 00:09:42.273 }, 00:09:42.273 { 00:09:42.273 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:42.273 "dma_device_type": 2 00:09:42.273 }, 00:09:42.273 { 00:09:42.273 "dma_device_id": "system", 00:09:42.273 "dma_device_type": 1 00:09:42.273 }, 00:09:42.273 { 00:09:42.273 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:42.273 "dma_device_type": 2 00:09:42.273 }, 00:09:42.273 { 00:09:42.273 "dma_device_id": "system", 00:09:42.273 "dma_device_type": 1 00:09:42.273 }, 00:09:42.273 { 00:09:42.273 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:42.273 "dma_device_type": 2 00:09:42.273 } 00:09:42.273 ], 00:09:42.273 "driver_specific": { 00:09:42.273 "raid": { 00:09:42.273 "uuid": "b739e458-8a58-4ee4-aeb8-e3a5d9fe10eb", 00:09:42.273 "strip_size_kb": 0, 00:09:42.273 "state": "online", 00:09:42.273 "raid_level": "raid1", 00:09:42.273 "superblock": true, 00:09:42.273 "num_base_bdevs": 3, 00:09:42.273 "num_base_bdevs_discovered": 3, 00:09:42.273 "num_base_bdevs_operational": 3, 00:09:42.273 "base_bdevs_list": [ 00:09:42.273 { 00:09:42.273 "name": "NewBaseBdev", 00:09:42.273 "uuid": "f88684cb-076e-41f9-854d-931023cf205e", 00:09:42.273 "is_configured": true, 00:09:42.273 "data_offset": 2048, 00:09:42.273 "data_size": 63488 00:09:42.273 }, 00:09:42.273 { 00:09:42.273 "name": "BaseBdev2", 00:09:42.273 "uuid": "4be23e28-10a5-4d20-a69a-28c41b6f6559", 00:09:42.273 "is_configured": true, 00:09:42.273 "data_offset": 2048, 00:09:42.273 "data_size": 63488 00:09:42.273 }, 00:09:42.273 { 00:09:42.273 "name": "BaseBdev3", 00:09:42.273 "uuid": "cd354491-def0-478d-be8e-0225ce81133d", 00:09:42.273 "is_configured": true, 00:09:42.273 "data_offset": 2048, 00:09:42.273 "data_size": 63488 00:09:42.273 } 00:09:42.273 ] 00:09:42.273 } 00:09:42.273 } 00:09:42.273 }' 00:09:42.273 17:31:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:42.533 17:31:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:09:42.534 BaseBdev2 00:09:42.534 BaseBdev3' 00:09:42.534 17:31:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:42.534 17:31:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:42.534 17:31:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:42.534 17:31:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:09:42.534 17:31:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:42.534 17:31:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:42.534 17:31:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:42.534 17:31:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:42.534 17:31:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:42.534 17:31:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:42.534 17:31:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:42.534 17:31:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:09:42.534 17:31:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:42.534 17:31:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:42.534 17:31:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:42.534 17:31:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:42.534 17:31:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:42.534 17:31:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:42.534 17:31:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:42.534 17:31:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:42.534 17:31:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:09:42.534 17:31:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:42.534 17:31:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:42.534 17:31:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:42.534 17:31:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:42.534 17:31:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:42.534 17:31:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:42.534 17:31:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:42.534 17:31:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:42.534 [2024-11-27 17:31:13.651381] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:42.534 [2024-11-27 17:31:13.651456] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:42.534 [2024-11-27 17:31:13.651551] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:42.534 [2024-11-27 17:31:13.651837] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:42.534 [2024-11-27 17:31:13.651884] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name Existed_Raid, state offline 00:09:42.534 17:31:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:42.534 17:31:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 78895 00:09:42.534 17:31:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@950 -- # '[' -z 78895 ']' 00:09:42.534 17:31:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # kill -0 78895 00:09:42.534 17:31:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@955 -- # uname 00:09:42.534 17:31:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:09:42.534 17:31:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 78895 00:09:42.534 killing process with pid 78895 00:09:42.534 17:31:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:09:42.534 17:31:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:09:42.534 17:31:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@968 -- # echo 'killing process with pid 78895' 00:09:42.534 17:31:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@969 -- # kill 78895 00:09:42.534 [2024-11-27 17:31:13.700778] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:09:42.534 17:31:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@974 -- # wait 78895 00:09:42.794 [2024-11-27 17:31:13.760538] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:09:43.054 17:31:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:09:43.054 00:09:43.054 real 0m9.136s 00:09:43.054 user 0m15.346s 00:09:43.054 sys 0m1.941s 00:09:43.054 17:31:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1126 -- # xtrace_disable 00:09:43.054 17:31:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:09:43.054 ************************************ 00:09:43.054 END TEST raid_state_function_test_sb 00:09:43.054 ************************************ 00:09:43.054 17:31:14 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test raid1 3 00:09:43.054 17:31:14 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:09:43.054 17:31:14 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:09:43.054 17:31:14 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:09:43.054 ************************************ 00:09:43.054 START TEST raid_superblock_test 00:09:43.054 ************************************ 00:09:43.054 17:31:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1125 -- # raid_superblock_test raid1 3 00:09:43.054 17:31:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=raid1 00:09:43.054 17:31:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=3 00:09:43.054 17:31:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:09:43.054 17:31:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:09:43.054 17:31:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:09:43.054 17:31:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:09:43.054 17:31:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:09:43.054 17:31:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:09:43.054 17:31:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:09:43.054 17:31:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:09:43.055 17:31:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:09:43.055 17:31:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:09:43.055 17:31:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:09:43.055 17:31:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' raid1 '!=' raid1 ']' 00:09:43.055 17:31:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@408 -- # strip_size=0 00:09:43.055 17:31:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=79499 00:09:43.055 17:31:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:09:43.055 17:31:14 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 79499 00:09:43.055 17:31:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@831 -- # '[' -z 79499 ']' 00:09:43.055 17:31:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:43.055 17:31:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:09:43.055 17:31:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:43.055 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:43.055 17:31:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:09:43.055 17:31:14 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:43.315 [2024-11-27 17:31:14.290429] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:09:43.315 [2024-11-27 17:31:14.291189] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid79499 ] 00:09:43.315 [2024-11-27 17:31:14.436782] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:43.315 [2024-11-27 17:31:14.504762] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:09:43.574 [2024-11-27 17:31:14.581394] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:43.574 [2024-11-27 17:31:14.581545] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:44.143 17:31:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:09:44.144 17:31:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@864 -- # return 0 00:09:44.144 17:31:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:09:44.144 17:31:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:09:44.144 17:31:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:09:44.144 17:31:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:09:44.144 17:31:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:09:44.144 17:31:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:09:44.144 17:31:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:09:44.144 17:31:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:09:44.144 17:31:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:09:44.144 17:31:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:44.144 17:31:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:44.144 malloc1 00:09:44.144 17:31:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:44.144 17:31:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:09:44.144 17:31:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:44.144 17:31:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:44.144 [2024-11-27 17:31:15.159527] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:09:44.144 [2024-11-27 17:31:15.159673] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:44.144 [2024-11-27 17:31:15.159717] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:09:44.144 [2024-11-27 17:31:15.159753] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:44.144 [2024-11-27 17:31:15.162141] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:44.144 [2024-11-27 17:31:15.162249] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:09:44.144 pt1 00:09:44.144 17:31:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:44.144 17:31:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:09:44.144 17:31:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:09:44.144 17:31:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:09:44.144 17:31:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:09:44.144 17:31:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:09:44.144 17:31:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:09:44.144 17:31:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:09:44.144 17:31:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:09:44.144 17:31:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:09:44.144 17:31:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:44.144 17:31:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:44.144 malloc2 00:09:44.144 17:31:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:44.144 17:31:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:09:44.144 17:31:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:44.144 17:31:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:44.144 [2024-11-27 17:31:15.212901] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:09:44.144 [2024-11-27 17:31:15.213013] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:44.144 [2024-11-27 17:31:15.213051] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:09:44.144 [2024-11-27 17:31:15.213078] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:44.144 [2024-11-27 17:31:15.218240] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:44.144 [2024-11-27 17:31:15.218405] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:09:44.144 pt2 00:09:44.144 17:31:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:44.144 17:31:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:09:44.144 17:31:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:09:44.144 17:31:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc3 00:09:44.144 17:31:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt3 00:09:44.144 17:31:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000003 00:09:44.144 17:31:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:09:44.144 17:31:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:09:44.144 17:31:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:09:44.144 17:31:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc3 00:09:44.144 17:31:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:44.144 17:31:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:44.144 malloc3 00:09:44.144 17:31:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:44.144 17:31:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:09:44.144 17:31:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:44.144 17:31:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:44.144 [2024-11-27 17:31:15.249824] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:09:44.144 [2024-11-27 17:31:15.249943] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:44.144 [2024-11-27 17:31:15.249979] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:09:44.144 [2024-11-27 17:31:15.250011] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:44.144 [2024-11-27 17:31:15.252443] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:44.144 [2024-11-27 17:31:15.252514] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:09:44.144 pt3 00:09:44.144 17:31:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:44.144 17:31:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:09:44.144 17:31:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:09:44.144 17:31:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''pt1 pt2 pt3'\''' -n raid_bdev1 -s 00:09:44.144 17:31:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:44.144 17:31:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:44.144 [2024-11-27 17:31:15.261882] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:09:44.144 [2024-11-27 17:31:15.264060] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:09:44.144 [2024-11-27 17:31:15.264168] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:09:44.144 [2024-11-27 17:31:15.264367] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:09:44.144 [2024-11-27 17:31:15.264411] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:09:44.144 [2024-11-27 17:31:15.264703] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:09:44.144 [2024-11-27 17:31:15.264896] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:09:44.144 [2024-11-27 17:31:15.264953] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:09:44.144 [2024-11-27 17:31:15.265121] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:44.144 17:31:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:44.144 17:31:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:09:44.144 17:31:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:44.144 17:31:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:44.144 17:31:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:44.144 17:31:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:44.144 17:31:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:44.144 17:31:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:44.144 17:31:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:44.144 17:31:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:44.144 17:31:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:44.144 17:31:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:44.144 17:31:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:44.144 17:31:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:44.144 17:31:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:44.144 17:31:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:44.144 17:31:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:44.144 "name": "raid_bdev1", 00:09:44.144 "uuid": "3e232a7c-5ce5-4ecf-b9d3-e21b0cb2a7c0", 00:09:44.144 "strip_size_kb": 0, 00:09:44.144 "state": "online", 00:09:44.144 "raid_level": "raid1", 00:09:44.144 "superblock": true, 00:09:44.144 "num_base_bdevs": 3, 00:09:44.144 "num_base_bdevs_discovered": 3, 00:09:44.144 "num_base_bdevs_operational": 3, 00:09:44.144 "base_bdevs_list": [ 00:09:44.144 { 00:09:44.144 "name": "pt1", 00:09:44.144 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:44.144 "is_configured": true, 00:09:44.144 "data_offset": 2048, 00:09:44.144 "data_size": 63488 00:09:44.144 }, 00:09:44.144 { 00:09:44.144 "name": "pt2", 00:09:44.144 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:44.144 "is_configured": true, 00:09:44.144 "data_offset": 2048, 00:09:44.144 "data_size": 63488 00:09:44.144 }, 00:09:44.144 { 00:09:44.144 "name": "pt3", 00:09:44.145 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:44.145 "is_configured": true, 00:09:44.145 "data_offset": 2048, 00:09:44.145 "data_size": 63488 00:09:44.145 } 00:09:44.145 ] 00:09:44.145 }' 00:09:44.145 17:31:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:44.145 17:31:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:44.713 17:31:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:09:44.713 17:31:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:09:44.713 17:31:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:44.713 17:31:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:44.713 17:31:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:09:44.713 17:31:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:44.713 17:31:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:09:44.713 17:31:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:44.713 17:31:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:44.713 17:31:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:44.713 [2024-11-27 17:31:15.713400] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:44.713 17:31:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:44.713 17:31:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:44.713 "name": "raid_bdev1", 00:09:44.713 "aliases": [ 00:09:44.713 "3e232a7c-5ce5-4ecf-b9d3-e21b0cb2a7c0" 00:09:44.713 ], 00:09:44.713 "product_name": "Raid Volume", 00:09:44.713 "block_size": 512, 00:09:44.713 "num_blocks": 63488, 00:09:44.713 "uuid": "3e232a7c-5ce5-4ecf-b9d3-e21b0cb2a7c0", 00:09:44.713 "assigned_rate_limits": { 00:09:44.713 "rw_ios_per_sec": 0, 00:09:44.713 "rw_mbytes_per_sec": 0, 00:09:44.713 "r_mbytes_per_sec": 0, 00:09:44.713 "w_mbytes_per_sec": 0 00:09:44.713 }, 00:09:44.713 "claimed": false, 00:09:44.713 "zoned": false, 00:09:44.713 "supported_io_types": { 00:09:44.713 "read": true, 00:09:44.713 "write": true, 00:09:44.713 "unmap": false, 00:09:44.713 "flush": false, 00:09:44.713 "reset": true, 00:09:44.713 "nvme_admin": false, 00:09:44.713 "nvme_io": false, 00:09:44.713 "nvme_io_md": false, 00:09:44.713 "write_zeroes": true, 00:09:44.713 "zcopy": false, 00:09:44.713 "get_zone_info": false, 00:09:44.713 "zone_management": false, 00:09:44.713 "zone_append": false, 00:09:44.713 "compare": false, 00:09:44.713 "compare_and_write": false, 00:09:44.713 "abort": false, 00:09:44.713 "seek_hole": false, 00:09:44.713 "seek_data": false, 00:09:44.713 "copy": false, 00:09:44.713 "nvme_iov_md": false 00:09:44.713 }, 00:09:44.713 "memory_domains": [ 00:09:44.713 { 00:09:44.713 "dma_device_id": "system", 00:09:44.713 "dma_device_type": 1 00:09:44.713 }, 00:09:44.713 { 00:09:44.713 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:44.713 "dma_device_type": 2 00:09:44.713 }, 00:09:44.713 { 00:09:44.713 "dma_device_id": "system", 00:09:44.713 "dma_device_type": 1 00:09:44.713 }, 00:09:44.713 { 00:09:44.713 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:44.713 "dma_device_type": 2 00:09:44.713 }, 00:09:44.713 { 00:09:44.713 "dma_device_id": "system", 00:09:44.713 "dma_device_type": 1 00:09:44.713 }, 00:09:44.713 { 00:09:44.713 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:44.713 "dma_device_type": 2 00:09:44.713 } 00:09:44.713 ], 00:09:44.713 "driver_specific": { 00:09:44.713 "raid": { 00:09:44.713 "uuid": "3e232a7c-5ce5-4ecf-b9d3-e21b0cb2a7c0", 00:09:44.713 "strip_size_kb": 0, 00:09:44.713 "state": "online", 00:09:44.713 "raid_level": "raid1", 00:09:44.713 "superblock": true, 00:09:44.713 "num_base_bdevs": 3, 00:09:44.713 "num_base_bdevs_discovered": 3, 00:09:44.713 "num_base_bdevs_operational": 3, 00:09:44.713 "base_bdevs_list": [ 00:09:44.713 { 00:09:44.713 "name": "pt1", 00:09:44.713 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:44.713 "is_configured": true, 00:09:44.713 "data_offset": 2048, 00:09:44.713 "data_size": 63488 00:09:44.713 }, 00:09:44.713 { 00:09:44.713 "name": "pt2", 00:09:44.713 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:44.713 "is_configured": true, 00:09:44.713 "data_offset": 2048, 00:09:44.713 "data_size": 63488 00:09:44.713 }, 00:09:44.713 { 00:09:44.713 "name": "pt3", 00:09:44.713 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:44.713 "is_configured": true, 00:09:44.713 "data_offset": 2048, 00:09:44.713 "data_size": 63488 00:09:44.713 } 00:09:44.713 ] 00:09:44.713 } 00:09:44.713 } 00:09:44.713 }' 00:09:44.713 17:31:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:44.713 17:31:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:09:44.713 pt2 00:09:44.713 pt3' 00:09:44.713 17:31:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:44.713 17:31:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:44.713 17:31:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:44.713 17:31:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:44.714 17:31:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:09:44.714 17:31:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:44.714 17:31:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:44.714 17:31:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:44.714 17:31:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:44.714 17:31:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:44.714 17:31:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:44.714 17:31:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:44.714 17:31:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:09:44.714 17:31:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:44.714 17:31:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:44.973 17:31:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:44.973 17:31:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:44.973 17:31:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:44.973 17:31:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:44.973 17:31:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:09:44.973 17:31:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:44.973 17:31:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:44.973 17:31:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:44.973 17:31:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:44.973 17:31:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:44.973 17:31:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:44.973 17:31:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:09:44.973 17:31:15 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:09:44.973 17:31:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:44.973 17:31:15 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:44.973 [2024-11-27 17:31:15.980816] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:44.973 17:31:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:44.973 17:31:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=3e232a7c-5ce5-4ecf-b9d3-e21b0cb2a7c0 00:09:44.973 17:31:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z 3e232a7c-5ce5-4ecf-b9d3-e21b0cb2a7c0 ']' 00:09:44.973 17:31:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:09:44.973 17:31:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:44.973 17:31:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:44.973 [2024-11-27 17:31:16.028497] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:44.973 [2024-11-27 17:31:16.028557] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:44.973 [2024-11-27 17:31:16.028679] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:44.973 [2024-11-27 17:31:16.028774] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:44.973 [2024-11-27 17:31:16.028822] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:09:44.973 17:31:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:44.973 17:31:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:44.973 17:31:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:44.973 17:31:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:44.973 17:31:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:09:44.973 17:31:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:44.973 17:31:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:09:44.973 17:31:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:09:44.973 17:31:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:09:44.973 17:31:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:09:44.973 17:31:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:44.973 17:31:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:44.973 17:31:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:44.973 17:31:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:09:44.973 17:31:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:09:44.973 17:31:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:44.973 17:31:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:44.973 17:31:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:44.973 17:31:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:09:44.973 17:31:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt3 00:09:44.973 17:31:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:44.973 17:31:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:44.973 17:31:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:44.973 17:31:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:09:44.973 17:31:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:44.973 17:31:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:09:44.973 17:31:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:44.973 17:31:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:45.233 17:31:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:09:45.233 17:31:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:09:45.233 17:31:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@650 -- # local es=0 00:09:45.233 17:31:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:09:45.233 17:31:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:09:45.233 17:31:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:09:45.233 17:31:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:09:45.233 17:31:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:09:45.233 17:31:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:09:45.233 17:31:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:45.233 17:31:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:45.233 [2024-11-27 17:31:16.176285] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:09:45.233 [2024-11-27 17:31:16.178463] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:09:45.233 [2024-11-27 17:31:16.178514] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc3 is claimed 00:09:45.233 [2024-11-27 17:31:16.178567] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:09:45.233 [2024-11-27 17:31:16.178616] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:09:45.233 [2024-11-27 17:31:16.178638] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc3 00:09:45.233 [2024-11-27 17:31:16.178650] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:45.233 [2024-11-27 17:31:16.178661] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state configuring 00:09:45.233 request: 00:09:45.233 { 00:09:45.233 "name": "raid_bdev1", 00:09:45.233 "raid_level": "raid1", 00:09:45.233 "base_bdevs": [ 00:09:45.233 "malloc1", 00:09:45.233 "malloc2", 00:09:45.233 "malloc3" 00:09:45.233 ], 00:09:45.233 "superblock": false, 00:09:45.233 "method": "bdev_raid_create", 00:09:45.233 "req_id": 1 00:09:45.233 } 00:09:45.233 Got JSON-RPC error response 00:09:45.233 response: 00:09:45.233 { 00:09:45.233 "code": -17, 00:09:45.233 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:09:45.233 } 00:09:45.233 17:31:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:09:45.233 17:31:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@653 -- # es=1 00:09:45.233 17:31:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:09:45.233 17:31:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:09:45.233 17:31:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:09:45.233 17:31:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:09:45.233 17:31:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:45.233 17:31:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:45.233 17:31:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:45.233 17:31:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:45.233 17:31:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:09:45.233 17:31:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:09:45.233 17:31:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:09:45.233 17:31:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:45.233 17:31:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:45.233 [2024-11-27 17:31:16.228180] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:09:45.233 [2024-11-27 17:31:16.228268] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:45.233 [2024-11-27 17:31:16.228301] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:09:45.233 [2024-11-27 17:31:16.228351] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:45.233 [2024-11-27 17:31:16.230804] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:45.233 [2024-11-27 17:31:16.230873] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:09:45.233 [2024-11-27 17:31:16.230980] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:09:45.234 [2024-11-27 17:31:16.231054] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:09:45.234 pt1 00:09:45.234 17:31:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:45.234 17:31:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 3 00:09:45.234 17:31:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:45.234 17:31:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:45.234 17:31:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:45.234 17:31:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:45.234 17:31:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:45.234 17:31:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:45.234 17:31:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:45.234 17:31:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:45.234 17:31:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:45.234 17:31:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:45.234 17:31:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:45.234 17:31:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:45.234 17:31:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:45.234 17:31:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:45.234 17:31:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:45.234 "name": "raid_bdev1", 00:09:45.234 "uuid": "3e232a7c-5ce5-4ecf-b9d3-e21b0cb2a7c0", 00:09:45.234 "strip_size_kb": 0, 00:09:45.234 "state": "configuring", 00:09:45.234 "raid_level": "raid1", 00:09:45.234 "superblock": true, 00:09:45.234 "num_base_bdevs": 3, 00:09:45.234 "num_base_bdevs_discovered": 1, 00:09:45.234 "num_base_bdevs_operational": 3, 00:09:45.234 "base_bdevs_list": [ 00:09:45.234 { 00:09:45.234 "name": "pt1", 00:09:45.234 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:45.234 "is_configured": true, 00:09:45.234 "data_offset": 2048, 00:09:45.234 "data_size": 63488 00:09:45.234 }, 00:09:45.234 { 00:09:45.234 "name": null, 00:09:45.234 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:45.234 "is_configured": false, 00:09:45.234 "data_offset": 2048, 00:09:45.234 "data_size": 63488 00:09:45.234 }, 00:09:45.234 { 00:09:45.234 "name": null, 00:09:45.234 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:45.234 "is_configured": false, 00:09:45.234 "data_offset": 2048, 00:09:45.234 "data_size": 63488 00:09:45.234 } 00:09:45.234 ] 00:09:45.234 }' 00:09:45.234 17:31:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:45.234 17:31:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:45.493 17:31:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 3 -gt 2 ']' 00:09:45.493 17:31:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@472 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:09:45.493 17:31:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:45.493 17:31:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:45.493 [2024-11-27 17:31:16.671401] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:09:45.493 [2024-11-27 17:31:16.671526] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:45.493 [2024-11-27 17:31:16.671551] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:09:45.493 [2024-11-27 17:31:16.671565] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:45.493 [2024-11-27 17:31:16.671979] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:45.493 [2024-11-27 17:31:16.671998] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:09:45.493 [2024-11-27 17:31:16.672063] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:09:45.493 [2024-11-27 17:31:16.672086] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:09:45.493 pt2 00:09:45.493 17:31:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:45.493 17:31:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@473 -- # rpc_cmd bdev_passthru_delete pt2 00:09:45.493 17:31:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:45.493 17:31:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:45.753 [2024-11-27 17:31:16.683391] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: pt2 00:09:45.753 17:31:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:45.753 17:31:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@474 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 3 00:09:45.753 17:31:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:45.753 17:31:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:45.753 17:31:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:45.753 17:31:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:45.753 17:31:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:45.753 17:31:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:45.753 17:31:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:45.753 17:31:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:45.753 17:31:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:45.753 17:31:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:45.753 17:31:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:45.753 17:31:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:45.753 17:31:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:45.753 17:31:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:45.753 17:31:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:45.753 "name": "raid_bdev1", 00:09:45.753 "uuid": "3e232a7c-5ce5-4ecf-b9d3-e21b0cb2a7c0", 00:09:45.753 "strip_size_kb": 0, 00:09:45.753 "state": "configuring", 00:09:45.753 "raid_level": "raid1", 00:09:45.753 "superblock": true, 00:09:45.753 "num_base_bdevs": 3, 00:09:45.753 "num_base_bdevs_discovered": 1, 00:09:45.753 "num_base_bdevs_operational": 3, 00:09:45.753 "base_bdevs_list": [ 00:09:45.753 { 00:09:45.753 "name": "pt1", 00:09:45.753 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:45.753 "is_configured": true, 00:09:45.753 "data_offset": 2048, 00:09:45.753 "data_size": 63488 00:09:45.753 }, 00:09:45.753 { 00:09:45.753 "name": null, 00:09:45.753 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:45.753 "is_configured": false, 00:09:45.753 "data_offset": 0, 00:09:45.753 "data_size": 63488 00:09:45.753 }, 00:09:45.753 { 00:09:45.753 "name": null, 00:09:45.753 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:45.753 "is_configured": false, 00:09:45.753 "data_offset": 2048, 00:09:45.753 "data_size": 63488 00:09:45.753 } 00:09:45.753 ] 00:09:45.753 }' 00:09:45.753 17:31:16 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:45.753 17:31:16 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:46.013 17:31:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:09:46.013 17:31:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:09:46.013 17:31:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:09:46.013 17:31:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:46.013 17:31:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:46.013 [2024-11-27 17:31:17.146604] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:09:46.013 [2024-11-27 17:31:17.146720] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:46.013 [2024-11-27 17:31:17.146757] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:09:46.013 [2024-11-27 17:31:17.146804] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:46.013 [2024-11-27 17:31:17.147238] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:46.013 [2024-11-27 17:31:17.147292] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:09:46.013 [2024-11-27 17:31:17.147388] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:09:46.013 [2024-11-27 17:31:17.147436] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:09:46.013 pt2 00:09:46.013 17:31:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:46.013 17:31:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:09:46.013 17:31:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:09:46.013 17:31:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:09:46.013 17:31:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:46.013 17:31:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:46.013 [2024-11-27 17:31:17.158585] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:09:46.013 [2024-11-27 17:31:17.158660] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:46.013 [2024-11-27 17:31:17.158715] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:09:46.013 [2024-11-27 17:31:17.158749] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:46.013 [2024-11-27 17:31:17.159091] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:46.013 [2024-11-27 17:31:17.159148] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:09:46.013 [2024-11-27 17:31:17.159230] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:09:46.013 [2024-11-27 17:31:17.159286] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:09:46.013 [2024-11-27 17:31:17.159417] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:09:46.013 [2024-11-27 17:31:17.159451] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:09:46.013 [2024-11-27 17:31:17.159715] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:09:46.014 [2024-11-27 17:31:17.159862] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:09:46.014 [2024-11-27 17:31:17.159902] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:09:46.014 [2024-11-27 17:31:17.160039] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:46.014 pt3 00:09:46.014 17:31:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:46.014 17:31:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:09:46.014 17:31:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:09:46.014 17:31:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:09:46.014 17:31:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:46.014 17:31:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:46.014 17:31:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:46.014 17:31:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:46.014 17:31:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:46.014 17:31:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:46.014 17:31:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:46.014 17:31:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:46.014 17:31:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:46.014 17:31:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:46.014 17:31:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:46.014 17:31:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:46.014 17:31:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:46.014 17:31:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:46.273 17:31:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:46.273 "name": "raid_bdev1", 00:09:46.273 "uuid": "3e232a7c-5ce5-4ecf-b9d3-e21b0cb2a7c0", 00:09:46.273 "strip_size_kb": 0, 00:09:46.273 "state": "online", 00:09:46.273 "raid_level": "raid1", 00:09:46.273 "superblock": true, 00:09:46.273 "num_base_bdevs": 3, 00:09:46.273 "num_base_bdevs_discovered": 3, 00:09:46.273 "num_base_bdevs_operational": 3, 00:09:46.273 "base_bdevs_list": [ 00:09:46.273 { 00:09:46.273 "name": "pt1", 00:09:46.273 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:46.273 "is_configured": true, 00:09:46.273 "data_offset": 2048, 00:09:46.273 "data_size": 63488 00:09:46.273 }, 00:09:46.273 { 00:09:46.273 "name": "pt2", 00:09:46.273 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:46.273 "is_configured": true, 00:09:46.273 "data_offset": 2048, 00:09:46.273 "data_size": 63488 00:09:46.273 }, 00:09:46.273 { 00:09:46.273 "name": "pt3", 00:09:46.273 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:46.273 "is_configured": true, 00:09:46.273 "data_offset": 2048, 00:09:46.273 "data_size": 63488 00:09:46.273 } 00:09:46.273 ] 00:09:46.273 }' 00:09:46.273 17:31:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:46.273 17:31:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:46.533 17:31:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:09:46.533 17:31:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:09:46.533 17:31:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:09:46.533 17:31:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:09:46.533 17:31:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:09:46.533 17:31:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:09:46.533 17:31:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:09:46.533 17:31:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:46.533 17:31:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:46.533 17:31:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:09:46.533 [2024-11-27 17:31:17.638065] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:46.533 17:31:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:46.533 17:31:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:09:46.533 "name": "raid_bdev1", 00:09:46.533 "aliases": [ 00:09:46.533 "3e232a7c-5ce5-4ecf-b9d3-e21b0cb2a7c0" 00:09:46.533 ], 00:09:46.533 "product_name": "Raid Volume", 00:09:46.533 "block_size": 512, 00:09:46.533 "num_blocks": 63488, 00:09:46.533 "uuid": "3e232a7c-5ce5-4ecf-b9d3-e21b0cb2a7c0", 00:09:46.533 "assigned_rate_limits": { 00:09:46.533 "rw_ios_per_sec": 0, 00:09:46.533 "rw_mbytes_per_sec": 0, 00:09:46.533 "r_mbytes_per_sec": 0, 00:09:46.533 "w_mbytes_per_sec": 0 00:09:46.533 }, 00:09:46.533 "claimed": false, 00:09:46.533 "zoned": false, 00:09:46.533 "supported_io_types": { 00:09:46.533 "read": true, 00:09:46.533 "write": true, 00:09:46.533 "unmap": false, 00:09:46.533 "flush": false, 00:09:46.533 "reset": true, 00:09:46.533 "nvme_admin": false, 00:09:46.533 "nvme_io": false, 00:09:46.533 "nvme_io_md": false, 00:09:46.533 "write_zeroes": true, 00:09:46.533 "zcopy": false, 00:09:46.533 "get_zone_info": false, 00:09:46.533 "zone_management": false, 00:09:46.533 "zone_append": false, 00:09:46.533 "compare": false, 00:09:46.533 "compare_and_write": false, 00:09:46.533 "abort": false, 00:09:46.533 "seek_hole": false, 00:09:46.533 "seek_data": false, 00:09:46.533 "copy": false, 00:09:46.533 "nvme_iov_md": false 00:09:46.533 }, 00:09:46.533 "memory_domains": [ 00:09:46.533 { 00:09:46.533 "dma_device_id": "system", 00:09:46.533 "dma_device_type": 1 00:09:46.533 }, 00:09:46.533 { 00:09:46.533 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:46.533 "dma_device_type": 2 00:09:46.533 }, 00:09:46.533 { 00:09:46.533 "dma_device_id": "system", 00:09:46.533 "dma_device_type": 1 00:09:46.533 }, 00:09:46.533 { 00:09:46.533 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:46.533 "dma_device_type": 2 00:09:46.533 }, 00:09:46.533 { 00:09:46.533 "dma_device_id": "system", 00:09:46.533 "dma_device_type": 1 00:09:46.533 }, 00:09:46.533 { 00:09:46.533 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:46.533 "dma_device_type": 2 00:09:46.533 } 00:09:46.533 ], 00:09:46.533 "driver_specific": { 00:09:46.533 "raid": { 00:09:46.533 "uuid": "3e232a7c-5ce5-4ecf-b9d3-e21b0cb2a7c0", 00:09:46.533 "strip_size_kb": 0, 00:09:46.533 "state": "online", 00:09:46.533 "raid_level": "raid1", 00:09:46.533 "superblock": true, 00:09:46.533 "num_base_bdevs": 3, 00:09:46.533 "num_base_bdevs_discovered": 3, 00:09:46.533 "num_base_bdevs_operational": 3, 00:09:46.533 "base_bdevs_list": [ 00:09:46.533 { 00:09:46.533 "name": "pt1", 00:09:46.533 "uuid": "00000000-0000-0000-0000-000000000001", 00:09:46.533 "is_configured": true, 00:09:46.533 "data_offset": 2048, 00:09:46.533 "data_size": 63488 00:09:46.533 }, 00:09:46.533 { 00:09:46.533 "name": "pt2", 00:09:46.533 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:46.533 "is_configured": true, 00:09:46.533 "data_offset": 2048, 00:09:46.533 "data_size": 63488 00:09:46.533 }, 00:09:46.533 { 00:09:46.533 "name": "pt3", 00:09:46.533 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:46.533 "is_configured": true, 00:09:46.533 "data_offset": 2048, 00:09:46.533 "data_size": 63488 00:09:46.533 } 00:09:46.533 ] 00:09:46.533 } 00:09:46.533 } 00:09:46.533 }' 00:09:46.533 17:31:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:09:46.793 17:31:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:09:46.793 pt2 00:09:46.793 pt3' 00:09:46.793 17:31:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:46.793 17:31:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:09:46.793 17:31:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:46.793 17:31:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:09:46.793 17:31:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:46.793 17:31:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:46.793 17:31:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:46.793 17:31:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:46.793 17:31:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:46.793 17:31:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:46.793 17:31:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:46.793 17:31:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:09:46.793 17:31:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:46.793 17:31:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:46.793 17:31:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:46.793 17:31:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:46.793 17:31:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:46.793 17:31:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:46.793 17:31:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:09:46.793 17:31:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:09:46.793 17:31:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:09:46.793 17:31:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:46.793 17:31:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:46.793 17:31:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:46.793 17:31:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:09:46.793 17:31:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:09:46.793 17:31:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:09:46.793 17:31:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:46.793 17:31:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:46.793 17:31:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:09:46.793 [2024-11-27 17:31:17.921525] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:46.793 17:31:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:46.793 17:31:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' 3e232a7c-5ce5-4ecf-b9d3-e21b0cb2a7c0 '!=' 3e232a7c-5ce5-4ecf-b9d3-e21b0cb2a7c0 ']' 00:09:46.793 17:31:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy raid1 00:09:46.793 17:31:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:09:46.793 17:31:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@199 -- # return 0 00:09:46.793 17:31:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@493 -- # rpc_cmd bdev_passthru_delete pt1 00:09:46.793 17:31:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:46.793 17:31:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:46.793 [2024-11-27 17:31:17.969250] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: pt1 00:09:46.793 17:31:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:46.793 17:31:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@496 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:09:46.793 17:31:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:46.793 17:31:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:46.794 17:31:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:46.794 17:31:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:46.794 17:31:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:09:46.794 17:31:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:46.794 17:31:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:46.794 17:31:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:46.794 17:31:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:46.794 17:31:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:46.794 17:31:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:46.794 17:31:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:47.053 17:31:17 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:47.053 17:31:17 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:47.053 17:31:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:47.053 "name": "raid_bdev1", 00:09:47.053 "uuid": "3e232a7c-5ce5-4ecf-b9d3-e21b0cb2a7c0", 00:09:47.053 "strip_size_kb": 0, 00:09:47.053 "state": "online", 00:09:47.053 "raid_level": "raid1", 00:09:47.053 "superblock": true, 00:09:47.053 "num_base_bdevs": 3, 00:09:47.053 "num_base_bdevs_discovered": 2, 00:09:47.053 "num_base_bdevs_operational": 2, 00:09:47.053 "base_bdevs_list": [ 00:09:47.053 { 00:09:47.053 "name": null, 00:09:47.053 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:47.053 "is_configured": false, 00:09:47.053 "data_offset": 0, 00:09:47.053 "data_size": 63488 00:09:47.053 }, 00:09:47.053 { 00:09:47.053 "name": "pt2", 00:09:47.053 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:47.053 "is_configured": true, 00:09:47.053 "data_offset": 2048, 00:09:47.053 "data_size": 63488 00:09:47.053 }, 00:09:47.053 { 00:09:47.053 "name": "pt3", 00:09:47.053 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:47.053 "is_configured": true, 00:09:47.053 "data_offset": 2048, 00:09:47.053 "data_size": 63488 00:09:47.053 } 00:09:47.053 ] 00:09:47.053 }' 00:09:47.053 17:31:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:47.053 17:31:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:47.311 17:31:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@499 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:09:47.311 17:31:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:47.311 17:31:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:47.311 [2024-11-27 17:31:18.464326] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:47.311 [2024-11-27 17:31:18.464403] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:47.312 [2024-11-27 17:31:18.464467] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:47.312 [2024-11-27 17:31:18.464543] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:47.312 [2024-11-27 17:31:18.464552] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:09:47.312 17:31:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:47.312 17:31:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:47.312 17:31:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:47.312 17:31:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:47.312 17:31:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # jq -r '.[]' 00:09:47.312 17:31:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:47.571 17:31:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # raid_bdev= 00:09:47.571 17:31:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@501 -- # '[' -n '' ']' 00:09:47.571 17:31:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i = 1 )) 00:09:47.571 17:31:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:09:47.571 17:31:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt2 00:09:47.571 17:31:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:47.571 17:31:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:47.571 17:31:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:47.571 17:31:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:09:47.571 17:31:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:09:47.571 17:31:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt3 00:09:47.571 17:31:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:47.571 17:31:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:47.571 17:31:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:47.571 17:31:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:09:47.571 17:31:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:09:47.571 17:31:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i = 1 )) 00:09:47.571 17:31:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:09:47.571 17:31:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@512 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:09:47.571 17:31:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:47.571 17:31:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:47.571 [2024-11-27 17:31:18.548191] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:09:47.571 [2024-11-27 17:31:18.548241] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:47.571 [2024-11-27 17:31:18.548263] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009980 00:09:47.571 [2024-11-27 17:31:18.548271] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:47.571 [2024-11-27 17:31:18.550747] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:47.571 [2024-11-27 17:31:18.550823] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:09:47.571 [2024-11-27 17:31:18.550897] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:09:47.571 [2024-11-27 17:31:18.550948] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:09:47.571 pt2 00:09:47.571 17:31:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:47.571 17:31:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@515 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 2 00:09:47.571 17:31:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:47.571 17:31:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:47.571 17:31:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:47.571 17:31:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:47.571 17:31:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:09:47.571 17:31:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:47.571 17:31:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:47.571 17:31:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:47.571 17:31:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:47.571 17:31:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:47.571 17:31:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:47.571 17:31:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:47.571 17:31:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:47.571 17:31:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:47.571 17:31:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:47.571 "name": "raid_bdev1", 00:09:47.571 "uuid": "3e232a7c-5ce5-4ecf-b9d3-e21b0cb2a7c0", 00:09:47.571 "strip_size_kb": 0, 00:09:47.571 "state": "configuring", 00:09:47.571 "raid_level": "raid1", 00:09:47.571 "superblock": true, 00:09:47.571 "num_base_bdevs": 3, 00:09:47.571 "num_base_bdevs_discovered": 1, 00:09:47.571 "num_base_bdevs_operational": 2, 00:09:47.571 "base_bdevs_list": [ 00:09:47.571 { 00:09:47.571 "name": null, 00:09:47.571 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:47.571 "is_configured": false, 00:09:47.571 "data_offset": 2048, 00:09:47.571 "data_size": 63488 00:09:47.571 }, 00:09:47.571 { 00:09:47.571 "name": "pt2", 00:09:47.571 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:47.571 "is_configured": true, 00:09:47.571 "data_offset": 2048, 00:09:47.571 "data_size": 63488 00:09:47.571 }, 00:09:47.571 { 00:09:47.571 "name": null, 00:09:47.571 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:47.571 "is_configured": false, 00:09:47.572 "data_offset": 2048, 00:09:47.572 "data_size": 63488 00:09:47.572 } 00:09:47.572 ] 00:09:47.572 }' 00:09:47.572 17:31:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:47.572 17:31:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:47.831 17:31:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i++ )) 00:09:47.831 17:31:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:09:47.831 17:31:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@519 -- # i=2 00:09:47.831 17:31:18 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@520 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:09:47.831 17:31:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:47.831 17:31:18 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:47.831 [2024-11-27 17:31:19.007432] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:09:47.831 [2024-11-27 17:31:19.007491] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:47.831 [2024-11-27 17:31:19.007514] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009f80 00:09:47.831 [2024-11-27 17:31:19.007523] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:47.831 [2024-11-27 17:31:19.007919] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:47.831 [2024-11-27 17:31:19.007935] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:09:47.831 [2024-11-27 17:31:19.008007] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:09:47.831 [2024-11-27 17:31:19.008033] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:09:47.831 [2024-11-27 17:31:19.008129] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:09:47.831 [2024-11-27 17:31:19.008150] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:09:47.831 [2024-11-27 17:31:19.008409] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:09:47.831 [2024-11-27 17:31:19.008528] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:09:47.831 [2024-11-27 17:31:19.008540] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001c80 00:09:47.831 [2024-11-27 17:31:19.008646] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:47.831 pt3 00:09:47.831 17:31:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:47.831 17:31:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@523 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:09:47.831 17:31:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:47.831 17:31:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:47.831 17:31:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:47.831 17:31:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:47.831 17:31:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:09:47.832 17:31:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:47.832 17:31:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:47.832 17:31:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:47.832 17:31:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:47.832 17:31:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:47.832 17:31:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:47.832 17:31:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:47.832 17:31:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.091 17:31:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:48.091 17:31:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:48.091 "name": "raid_bdev1", 00:09:48.091 "uuid": "3e232a7c-5ce5-4ecf-b9d3-e21b0cb2a7c0", 00:09:48.091 "strip_size_kb": 0, 00:09:48.091 "state": "online", 00:09:48.091 "raid_level": "raid1", 00:09:48.091 "superblock": true, 00:09:48.091 "num_base_bdevs": 3, 00:09:48.091 "num_base_bdevs_discovered": 2, 00:09:48.091 "num_base_bdevs_operational": 2, 00:09:48.091 "base_bdevs_list": [ 00:09:48.091 { 00:09:48.091 "name": null, 00:09:48.091 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:48.091 "is_configured": false, 00:09:48.091 "data_offset": 2048, 00:09:48.091 "data_size": 63488 00:09:48.091 }, 00:09:48.091 { 00:09:48.091 "name": "pt2", 00:09:48.091 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:48.091 "is_configured": true, 00:09:48.091 "data_offset": 2048, 00:09:48.091 "data_size": 63488 00:09:48.091 }, 00:09:48.091 { 00:09:48.091 "name": "pt3", 00:09:48.091 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:48.091 "is_configured": true, 00:09:48.091 "data_offset": 2048, 00:09:48.091 "data_size": 63488 00:09:48.091 } 00:09:48.091 ] 00:09:48.091 }' 00:09:48.091 17:31:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:48.091 17:31:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.351 17:31:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@526 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:09:48.351 17:31:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:48.351 17:31:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.351 [2024-11-27 17:31:19.462623] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:48.351 [2024-11-27 17:31:19.462695] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:48.351 [2024-11-27 17:31:19.462801] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:48.351 [2024-11-27 17:31:19.462875] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:48.351 [2024-11-27 17:31:19.462921] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name raid_bdev1, state offline 00:09:48.351 17:31:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:48.351 17:31:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:48.351 17:31:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:48.351 17:31:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.351 17:31:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # jq -r '.[]' 00:09:48.351 17:31:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:48.351 17:31:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # raid_bdev= 00:09:48.351 17:31:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@528 -- # '[' -n '' ']' 00:09:48.351 17:31:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@532 -- # '[' 3 -gt 2 ']' 00:09:48.351 17:31:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@534 -- # i=2 00:09:48.351 17:31:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@535 -- # rpc_cmd bdev_passthru_delete pt3 00:09:48.351 17:31:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:48.351 17:31:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.351 17:31:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:48.351 17:31:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@540 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:09:48.351 17:31:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:48.351 17:31:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.351 [2024-11-27 17:31:19.530508] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:09:48.351 [2024-11-27 17:31:19.530613] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:48.351 [2024-11-27 17:31:19.530644] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:09:48.351 [2024-11-27 17:31:19.530692] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:48.351 [2024-11-27 17:31:19.533096] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:48.351 [2024-11-27 17:31:19.533194] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:09:48.351 [2024-11-27 17:31:19.533302] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:09:48.351 [2024-11-27 17:31:19.533371] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:09:48.351 [2024-11-27 17:31:19.533508] bdev_raid.c:3675:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev pt2 (4) greater than existing raid bdev raid_bdev1 (2) 00:09:48.352 [2024-11-27 17:31:19.533565] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:48.352 [2024-11-27 17:31:19.533602] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002000 name raid_bdev1, state configuring 00:09:48.352 [2024-11-27 17:31:19.533684] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:09:48.352 pt1 00:09:48.352 17:31:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:48.352 17:31:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@542 -- # '[' 3 -gt 2 ']' 00:09:48.352 17:31:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@545 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 2 00:09:48.352 17:31:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:48.352 17:31:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:48.352 17:31:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:48.352 17:31:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:48.352 17:31:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:09:48.352 17:31:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:48.352 17:31:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:48.352 17:31:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:48.352 17:31:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:48.611 17:31:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:48.611 17:31:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:48.611 17:31:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:48.611 17:31:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.611 17:31:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:48.611 17:31:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:48.611 "name": "raid_bdev1", 00:09:48.611 "uuid": "3e232a7c-5ce5-4ecf-b9d3-e21b0cb2a7c0", 00:09:48.611 "strip_size_kb": 0, 00:09:48.611 "state": "configuring", 00:09:48.611 "raid_level": "raid1", 00:09:48.611 "superblock": true, 00:09:48.611 "num_base_bdevs": 3, 00:09:48.611 "num_base_bdevs_discovered": 1, 00:09:48.611 "num_base_bdevs_operational": 2, 00:09:48.611 "base_bdevs_list": [ 00:09:48.611 { 00:09:48.611 "name": null, 00:09:48.611 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:48.611 "is_configured": false, 00:09:48.611 "data_offset": 2048, 00:09:48.612 "data_size": 63488 00:09:48.612 }, 00:09:48.612 { 00:09:48.612 "name": "pt2", 00:09:48.612 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:48.612 "is_configured": true, 00:09:48.612 "data_offset": 2048, 00:09:48.612 "data_size": 63488 00:09:48.612 }, 00:09:48.612 { 00:09:48.612 "name": null, 00:09:48.612 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:48.612 "is_configured": false, 00:09:48.612 "data_offset": 2048, 00:09:48.612 "data_size": 63488 00:09:48.612 } 00:09:48.612 ] 00:09:48.612 }' 00:09:48.612 17:31:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:48.612 17:31:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.871 17:31:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@546 -- # rpc_cmd bdev_raid_get_bdevs configuring 00:09:48.872 17:31:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@546 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:09:48.872 17:31:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:48.872 17:31:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.872 17:31:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:48.872 17:31:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@546 -- # [[ false == \f\a\l\s\e ]] 00:09:48.872 17:31:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@549 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:09:48.872 17:31:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:48.872 17:31:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:48.872 [2024-11-27 17:31:20.041622] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:09:48.872 [2024-11-27 17:31:20.041738] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:48.872 [2024-11-27 17:31:20.041759] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a880 00:09:48.872 [2024-11-27 17:31:20.041770] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:48.872 [2024-11-27 17:31:20.042197] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:48.872 [2024-11-27 17:31:20.042222] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:09:48.872 [2024-11-27 17:31:20.042293] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:09:48.872 [2024-11-27 17:31:20.042316] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:09:48.872 [2024-11-27 17:31:20.042408] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000002380 00:09:48.872 [2024-11-27 17:31:20.042419] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:09:48.872 [2024-11-27 17:31:20.042654] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000026d0 00:09:48.872 [2024-11-27 17:31:20.042799] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000002380 00:09:48.872 [2024-11-27 17:31:20.042809] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000002380 00:09:48.872 [2024-11-27 17:31:20.042916] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:48.872 pt3 00:09:48.872 17:31:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:48.872 17:31:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@554 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:09:48.872 17:31:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:48.872 17:31:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:48.872 17:31:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:48.872 17:31:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:48.872 17:31:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:09:48.872 17:31:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:48.872 17:31:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:48.872 17:31:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:48.872 17:31:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:48.872 17:31:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:48.872 17:31:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:48.872 17:31:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:48.872 17:31:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:49.131 17:31:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:49.131 17:31:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:49.131 "name": "raid_bdev1", 00:09:49.131 "uuid": "3e232a7c-5ce5-4ecf-b9d3-e21b0cb2a7c0", 00:09:49.131 "strip_size_kb": 0, 00:09:49.131 "state": "online", 00:09:49.132 "raid_level": "raid1", 00:09:49.132 "superblock": true, 00:09:49.132 "num_base_bdevs": 3, 00:09:49.132 "num_base_bdevs_discovered": 2, 00:09:49.132 "num_base_bdevs_operational": 2, 00:09:49.132 "base_bdevs_list": [ 00:09:49.132 { 00:09:49.132 "name": null, 00:09:49.132 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:49.132 "is_configured": false, 00:09:49.132 "data_offset": 2048, 00:09:49.132 "data_size": 63488 00:09:49.132 }, 00:09:49.132 { 00:09:49.132 "name": "pt2", 00:09:49.132 "uuid": "00000000-0000-0000-0000-000000000002", 00:09:49.132 "is_configured": true, 00:09:49.132 "data_offset": 2048, 00:09:49.132 "data_size": 63488 00:09:49.132 }, 00:09:49.132 { 00:09:49.132 "name": "pt3", 00:09:49.132 "uuid": "00000000-0000-0000-0000-000000000003", 00:09:49.132 "is_configured": true, 00:09:49.132 "data_offset": 2048, 00:09:49.132 "data_size": 63488 00:09:49.132 } 00:09:49.132 ] 00:09:49.132 }' 00:09:49.132 17:31:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:49.132 17:31:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:49.391 17:31:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:09:49.391 17:31:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # rpc_cmd bdev_raid_get_bdevs online 00:09:49.391 17:31:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:49.391 17:31:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:49.391 17:31:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:49.391 17:31:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # [[ false == \f\a\l\s\e ]] 00:09:49.391 17:31:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # jq -r '.[] | .uuid' 00:09:49.391 17:31:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:09:49.391 17:31:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:49.391 17:31:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:49.391 [2024-11-27 17:31:20.552991] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:09:49.391 17:31:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:49.651 17:31:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # '[' 3e232a7c-5ce5-4ecf-b9d3-e21b0cb2a7c0 '!=' 3e232a7c-5ce5-4ecf-b9d3-e21b0cb2a7c0 ']' 00:09:49.651 17:31:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 79499 00:09:49.651 17:31:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@950 -- # '[' -z 79499 ']' 00:09:49.651 17:31:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # kill -0 79499 00:09:49.651 17:31:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@955 -- # uname 00:09:49.651 17:31:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:09:49.651 17:31:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 79499 00:09:49.651 17:31:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:09:49.651 killing process with pid 79499 00:09:49.651 17:31:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:09:49.651 17:31:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 79499' 00:09:49.651 17:31:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@969 -- # kill 79499 00:09:49.651 [2024-11-27 17:31:20.631739] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:09:49.651 [2024-11-27 17:31:20.631815] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:49.651 [2024-11-27 17:31:20.631876] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:49.651 [2024-11-27 17:31:20.631885] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002380 name raid_bdev1, state offline 00:09:49.651 17:31:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@974 -- # wait 79499 00:09:49.651 [2024-11-27 17:31:20.694026] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:09:49.911 17:31:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:09:49.911 00:09:49.911 real 0m6.862s 00:09:49.911 user 0m11.314s 00:09:49.911 sys 0m1.435s 00:09:49.911 17:31:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:09:49.911 ************************************ 00:09:49.911 END TEST raid_superblock_test 00:09:49.911 ************************************ 00:09:49.911 17:31:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:09:50.171 17:31:21 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test raid1 3 read 00:09:50.171 17:31:21 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:09:50.171 17:31:21 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:09:50.171 17:31:21 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:09:50.171 ************************************ 00:09:50.171 START TEST raid_read_error_test 00:09:50.171 ************************************ 00:09:50.171 17:31:21 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1125 -- # raid_io_error_test raid1 3 read 00:09:50.171 17:31:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid1 00:09:50.171 17:31:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=3 00:09:50.171 17:31:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:09:50.171 17:31:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:09:50.171 17:31:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:50.171 17:31:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:09:50.171 17:31:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:50.171 17:31:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:50.171 17:31:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:09:50.171 17:31:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:50.171 17:31:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:50.171 17:31:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:09:50.171 17:31:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:50.171 17:31:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:50.171 17:31:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:09:50.171 17:31:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:09:50.171 17:31:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:09:50.171 17:31:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:09:50.171 17:31:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:09:50.171 17:31:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:09:50.171 17:31:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:09:50.171 17:31:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid1 '!=' raid1 ']' 00:09:50.171 17:31:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@804 -- # strip_size=0 00:09:50.171 17:31:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:09:50.171 17:31:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.dVrmP5oyRq 00:09:50.172 17:31:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=79939 00:09:50.172 17:31:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:09:50.172 17:31:21 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 79939 00:09:50.172 17:31:21 bdev_raid.raid_read_error_test -- common/autotest_common.sh@831 -- # '[' -z 79939 ']' 00:09:50.172 17:31:21 bdev_raid.raid_read_error_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:50.172 17:31:21 bdev_raid.raid_read_error_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:09:50.172 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:50.172 17:31:21 bdev_raid.raid_read_error_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:50.172 17:31:21 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:09:50.172 17:31:21 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:50.172 [2024-11-27 17:31:21.241490] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:09:50.172 [2024-11-27 17:31:21.241639] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid79939 ] 00:09:50.431 [2024-11-27 17:31:21.388354] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:50.431 [2024-11-27 17:31:21.456207] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:09:50.431 [2024-11-27 17:31:21.531845] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:50.431 [2024-11-27 17:31:21.531901] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:51.001 17:31:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:09:51.001 17:31:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@864 -- # return 0 00:09:51.001 17:31:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:51.001 17:31:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:09:51.001 17:31:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:51.001 17:31:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:51.001 BaseBdev1_malloc 00:09:51.001 17:31:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:51.001 17:31:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:09:51.002 17:31:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:51.002 17:31:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:51.002 true 00:09:51.002 17:31:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:51.002 17:31:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:09:51.002 17:31:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:51.002 17:31:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:51.002 [2024-11-27 17:31:22.098112] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:09:51.002 [2024-11-27 17:31:22.098193] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:51.002 [2024-11-27 17:31:22.098214] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:09:51.002 [2024-11-27 17:31:22.098223] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:51.002 [2024-11-27 17:31:22.100672] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:51.002 [2024-11-27 17:31:22.100762] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:09:51.002 BaseBdev1 00:09:51.002 17:31:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:51.002 17:31:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:51.002 17:31:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:09:51.002 17:31:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:51.002 17:31:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:51.002 BaseBdev2_malloc 00:09:51.002 17:31:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:51.002 17:31:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:09:51.002 17:31:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:51.002 17:31:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:51.002 true 00:09:51.002 17:31:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:51.002 17:31:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:09:51.002 17:31:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:51.002 17:31:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:51.002 [2024-11-27 17:31:22.160216] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:09:51.002 [2024-11-27 17:31:22.160370] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:51.002 [2024-11-27 17:31:22.160408] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:09:51.002 [2024-11-27 17:31:22.160424] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:51.002 [2024-11-27 17:31:22.163321] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:51.002 [2024-11-27 17:31:22.163357] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:09:51.002 BaseBdev2 00:09:51.002 17:31:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:51.002 17:31:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:51.002 17:31:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:09:51.002 17:31:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:51.002 17:31:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:51.002 BaseBdev3_malloc 00:09:51.002 17:31:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:51.002 17:31:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:09:51.002 17:31:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:51.002 17:31:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:51.261 true 00:09:51.261 17:31:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:51.261 17:31:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:09:51.261 17:31:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:51.261 17:31:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:51.261 [2024-11-27 17:31:22.206839] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:09:51.261 [2024-11-27 17:31:22.206890] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:51.261 [2024-11-27 17:31:22.206927] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008780 00:09:51.261 [2024-11-27 17:31:22.206936] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:51.261 [2024-11-27 17:31:22.209236] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:51.261 [2024-11-27 17:31:22.209318] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:09:51.261 BaseBdev3 00:09:51.261 17:31:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:51.261 17:31:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n raid_bdev1 -s 00:09:51.261 17:31:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:51.261 17:31:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:51.261 [2024-11-27 17:31:22.218916] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:51.261 [2024-11-27 17:31:22.220956] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:51.261 [2024-11-27 17:31:22.221069] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:51.261 [2024-11-27 17:31:22.221272] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:09:51.261 [2024-11-27 17:31:22.221289] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:09:51.261 [2024-11-27 17:31:22.221537] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:09:51.261 [2024-11-27 17:31:22.221681] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:09:51.261 [2024-11-27 17:31:22.221691] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001c80 00:09:51.261 [2024-11-27 17:31:22.221824] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:51.261 17:31:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:51.261 17:31:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:09:51.261 17:31:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:51.261 17:31:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:51.261 17:31:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:51.261 17:31:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:51.261 17:31:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:51.261 17:31:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:51.261 17:31:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:51.261 17:31:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:51.261 17:31:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:51.262 17:31:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:51.262 17:31:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:51.262 17:31:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:51.262 17:31:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:51.262 17:31:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:51.262 17:31:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:51.262 "name": "raid_bdev1", 00:09:51.262 "uuid": "e425a927-ab59-423f-b1c0-30167f3d2760", 00:09:51.262 "strip_size_kb": 0, 00:09:51.262 "state": "online", 00:09:51.262 "raid_level": "raid1", 00:09:51.262 "superblock": true, 00:09:51.262 "num_base_bdevs": 3, 00:09:51.262 "num_base_bdevs_discovered": 3, 00:09:51.262 "num_base_bdevs_operational": 3, 00:09:51.262 "base_bdevs_list": [ 00:09:51.262 { 00:09:51.262 "name": "BaseBdev1", 00:09:51.262 "uuid": "1b2eff38-ab02-51ef-9267-a73c36d55003", 00:09:51.262 "is_configured": true, 00:09:51.262 "data_offset": 2048, 00:09:51.262 "data_size": 63488 00:09:51.262 }, 00:09:51.262 { 00:09:51.262 "name": "BaseBdev2", 00:09:51.262 "uuid": "c4f459b9-6b11-5890-a5ae-fffa3547e859", 00:09:51.262 "is_configured": true, 00:09:51.262 "data_offset": 2048, 00:09:51.262 "data_size": 63488 00:09:51.262 }, 00:09:51.262 { 00:09:51.262 "name": "BaseBdev3", 00:09:51.262 "uuid": "9ab44a9f-23fa-53ca-822a-5f3045560192", 00:09:51.262 "is_configured": true, 00:09:51.262 "data_offset": 2048, 00:09:51.262 "data_size": 63488 00:09:51.262 } 00:09:51.262 ] 00:09:51.262 }' 00:09:51.262 17:31:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:51.262 17:31:22 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:51.521 17:31:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:09:51.521 17:31:22 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:09:51.781 [2024-11-27 17:31:22.722567] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:09:52.745 17:31:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:09:52.746 17:31:23 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:52.746 17:31:23 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:52.746 17:31:23 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:52.746 17:31:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:09:52.746 17:31:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid1 = \r\a\i\d\1 ]] 00:09:52.746 17:31:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ read = \w\r\i\t\e ]] 00:09:52.746 17:31:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=3 00:09:52.746 17:31:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:09:52.746 17:31:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:52.746 17:31:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:52.746 17:31:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:52.746 17:31:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:52.746 17:31:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:52.746 17:31:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:52.746 17:31:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:52.746 17:31:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:52.746 17:31:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:52.746 17:31:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:52.746 17:31:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:52.746 17:31:23 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:52.746 17:31:23 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:52.746 17:31:23 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:52.746 17:31:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:52.746 "name": "raid_bdev1", 00:09:52.746 "uuid": "e425a927-ab59-423f-b1c0-30167f3d2760", 00:09:52.746 "strip_size_kb": 0, 00:09:52.746 "state": "online", 00:09:52.746 "raid_level": "raid1", 00:09:52.746 "superblock": true, 00:09:52.746 "num_base_bdevs": 3, 00:09:52.746 "num_base_bdevs_discovered": 3, 00:09:52.746 "num_base_bdevs_operational": 3, 00:09:52.746 "base_bdevs_list": [ 00:09:52.746 { 00:09:52.746 "name": "BaseBdev1", 00:09:52.746 "uuid": "1b2eff38-ab02-51ef-9267-a73c36d55003", 00:09:52.746 "is_configured": true, 00:09:52.746 "data_offset": 2048, 00:09:52.746 "data_size": 63488 00:09:52.746 }, 00:09:52.746 { 00:09:52.746 "name": "BaseBdev2", 00:09:52.746 "uuid": "c4f459b9-6b11-5890-a5ae-fffa3547e859", 00:09:52.746 "is_configured": true, 00:09:52.746 "data_offset": 2048, 00:09:52.746 "data_size": 63488 00:09:52.746 }, 00:09:52.746 { 00:09:52.746 "name": "BaseBdev3", 00:09:52.746 "uuid": "9ab44a9f-23fa-53ca-822a-5f3045560192", 00:09:52.746 "is_configured": true, 00:09:52.746 "data_offset": 2048, 00:09:52.746 "data_size": 63488 00:09:52.746 } 00:09:52.746 ] 00:09:52.746 }' 00:09:52.746 17:31:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:52.746 17:31:23 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:53.006 17:31:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:09:53.006 17:31:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:53.006 17:31:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:53.006 [2024-11-27 17:31:24.123839] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:53.006 [2024-11-27 17:31:24.123884] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:53.006 [2024-11-27 17:31:24.126364] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:53.006 [2024-11-27 17:31:24.126490] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:53.006 [2024-11-27 17:31:24.126608] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:53.006 [2024-11-27 17:31:24.126630] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name raid_bdev1, state offline 00:09:53.006 { 00:09:53.006 "results": [ 00:09:53.006 { 00:09:53.006 "job": "raid_bdev1", 00:09:53.006 "core_mask": "0x1", 00:09:53.006 "workload": "randrw", 00:09:53.006 "percentage": 50, 00:09:53.006 "status": "finished", 00:09:53.006 "queue_depth": 1, 00:09:53.007 "io_size": 131072, 00:09:53.007 "runtime": 1.401893, 00:09:53.007 "iops": 11336.100544050081, 00:09:53.007 "mibps": 1417.0125680062602, 00:09:53.007 "io_failed": 0, 00:09:53.007 "io_timeout": 0, 00:09:53.007 "avg_latency_us": 85.73180408807485, 00:09:53.007 "min_latency_us": 21.687336244541484, 00:09:53.007 "max_latency_us": 1480.9991266375546 00:09:53.007 } 00:09:53.007 ], 00:09:53.007 "core_count": 1 00:09:53.007 } 00:09:53.007 17:31:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:53.007 17:31:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 79939 00:09:53.007 17:31:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@950 -- # '[' -z 79939 ']' 00:09:53.007 17:31:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # kill -0 79939 00:09:53.007 17:31:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@955 -- # uname 00:09:53.007 17:31:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:09:53.007 17:31:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 79939 00:09:53.007 17:31:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:09:53.007 17:31:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:09:53.007 killing process with pid 79939 00:09:53.007 17:31:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 79939' 00:09:53.007 17:31:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@969 -- # kill 79939 00:09:53.007 [2024-11-27 17:31:24.172662] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:09:53.007 17:31:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@974 -- # wait 79939 00:09:53.267 [2024-11-27 17:31:24.222190] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:09:53.528 17:31:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.dVrmP5oyRq 00:09:53.528 17:31:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:09:53.528 17:31:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:09:53.528 17:31:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.00 00:09:53.528 17:31:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid1 00:09:53.528 17:31:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:09:53.528 17:31:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@199 -- # return 0 00:09:53.528 ************************************ 00:09:53.528 END TEST raid_read_error_test 00:09:53.528 ************************************ 00:09:53.528 17:31:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@847 -- # [[ 0.00 = \0\.\0\0 ]] 00:09:53.528 00:09:53.528 real 0m3.462s 00:09:53.528 user 0m4.241s 00:09:53.528 sys 0m0.604s 00:09:53.528 17:31:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:09:53.528 17:31:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:53.528 17:31:24 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test raid1 3 write 00:09:53.529 17:31:24 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:09:53.529 17:31:24 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:09:53.529 17:31:24 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:09:53.529 ************************************ 00:09:53.529 START TEST raid_write_error_test 00:09:53.529 ************************************ 00:09:53.529 17:31:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1125 -- # raid_io_error_test raid1 3 write 00:09:53.529 17:31:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid1 00:09:53.529 17:31:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=3 00:09:53.529 17:31:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:09:53.529 17:31:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:09:53.529 17:31:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:53.529 17:31:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:09:53.529 17:31:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:53.529 17:31:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:53.529 17:31:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:09:53.529 17:31:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:53.529 17:31:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:53.529 17:31:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:09:53.529 17:31:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:09:53.529 17:31:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:09:53.529 17:31:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:09:53.529 17:31:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:09:53.529 17:31:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:09:53.529 17:31:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:09:53.529 17:31:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:09:53.529 17:31:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:09:53.529 17:31:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:09:53.529 17:31:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid1 '!=' raid1 ']' 00:09:53.529 17:31:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@804 -- # strip_size=0 00:09:53.529 17:31:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:09:53.529 17:31:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.Tn61xa4aFB 00:09:53.529 17:31:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=80068 00:09:53.529 17:31:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:09:53.529 17:31:24 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 80068 00:09:53.529 17:31:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@831 -- # '[' -z 80068 ']' 00:09:53.529 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:53.529 17:31:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:53.529 17:31:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:09:53.529 17:31:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:53.529 17:31:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:09:53.529 17:31:24 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:53.790 [2024-11-27 17:31:24.781832] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:09:53.790 [2024-11-27 17:31:24.781978] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid80068 ] 00:09:53.790 [2024-11-27 17:31:24.928700] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:54.050 [2024-11-27 17:31:24.997722] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:09:54.050 [2024-11-27 17:31:25.073638] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:54.050 [2024-11-27 17:31:25.073677] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:54.621 17:31:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:09:54.621 17:31:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@864 -- # return 0 00:09:54.621 17:31:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:54.621 17:31:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:09:54.621 17:31:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:54.621 17:31:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:54.621 BaseBdev1_malloc 00:09:54.621 17:31:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:54.621 17:31:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:09:54.621 17:31:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:54.621 17:31:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:54.621 true 00:09:54.621 17:31:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:54.621 17:31:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:09:54.621 17:31:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:54.621 17:31:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:54.621 [2024-11-27 17:31:25.640031] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:09:54.621 [2024-11-27 17:31:25.640103] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:54.621 [2024-11-27 17:31:25.640128] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:09:54.621 [2024-11-27 17:31:25.640149] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:54.621 [2024-11-27 17:31:25.642561] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:54.621 [2024-11-27 17:31:25.642595] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:09:54.621 BaseBdev1 00:09:54.621 17:31:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:54.621 17:31:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:54.621 17:31:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:09:54.621 17:31:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:54.621 17:31:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:54.621 BaseBdev2_malloc 00:09:54.621 17:31:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:54.621 17:31:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:09:54.621 17:31:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:54.621 17:31:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:54.621 true 00:09:54.621 17:31:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:54.621 17:31:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:09:54.621 17:31:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:54.621 17:31:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:54.621 [2024-11-27 17:31:25.694474] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:09:54.621 [2024-11-27 17:31:25.694533] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:54.621 [2024-11-27 17:31:25.694554] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:09:54.621 [2024-11-27 17:31:25.694562] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:54.621 [2024-11-27 17:31:25.697017] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:54.621 [2024-11-27 17:31:25.697053] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:09:54.621 BaseBdev2 00:09:54.621 17:31:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:54.621 17:31:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:09:54.621 17:31:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:09:54.621 17:31:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:54.621 17:31:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:54.621 BaseBdev3_malloc 00:09:54.621 17:31:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:54.621 17:31:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:09:54.621 17:31:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:54.621 17:31:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:54.621 true 00:09:54.621 17:31:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:54.621 17:31:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:09:54.621 17:31:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:54.621 17:31:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:54.621 [2024-11-27 17:31:25.740884] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:09:54.621 [2024-11-27 17:31:25.740931] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:09:54.621 [2024-11-27 17:31:25.740967] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008780 00:09:54.621 [2024-11-27 17:31:25.740976] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:09:54.621 [2024-11-27 17:31:25.743372] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:09:54.621 [2024-11-27 17:31:25.743406] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:09:54.621 BaseBdev3 00:09:54.622 17:31:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:54.622 17:31:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n raid_bdev1 -s 00:09:54.622 17:31:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:54.622 17:31:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:54.622 [2024-11-27 17:31:25.752964] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:54.622 [2024-11-27 17:31:25.755133] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:54.622 [2024-11-27 17:31:25.755274] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:54.622 [2024-11-27 17:31:25.755471] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:09:54.622 [2024-11-27 17:31:25.755487] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:09:54.622 [2024-11-27 17:31:25.755734] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:09:54.622 [2024-11-27 17:31:25.755893] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:09:54.622 [2024-11-27 17:31:25.755909] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001c80 00:09:54.622 [2024-11-27 17:31:25.756034] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:54.622 17:31:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:54.622 17:31:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:09:54.622 17:31:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:54.622 17:31:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:54.622 17:31:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:54.622 17:31:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:54.622 17:31:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:09:54.622 17:31:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:54.622 17:31:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:54.622 17:31:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:54.622 17:31:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:54.622 17:31:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:54.622 17:31:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:54.622 17:31:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:54.622 17:31:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:54.622 17:31:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:54.622 17:31:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:54.622 "name": "raid_bdev1", 00:09:54.622 "uuid": "1a2780d4-6017-4e26-b4d7-36a9c84c6485", 00:09:54.622 "strip_size_kb": 0, 00:09:54.622 "state": "online", 00:09:54.622 "raid_level": "raid1", 00:09:54.622 "superblock": true, 00:09:54.622 "num_base_bdevs": 3, 00:09:54.622 "num_base_bdevs_discovered": 3, 00:09:54.622 "num_base_bdevs_operational": 3, 00:09:54.622 "base_bdevs_list": [ 00:09:54.622 { 00:09:54.622 "name": "BaseBdev1", 00:09:54.622 "uuid": "d0ee32ab-bc95-5cf3-8e0c-4204003e9137", 00:09:54.622 "is_configured": true, 00:09:54.622 "data_offset": 2048, 00:09:54.622 "data_size": 63488 00:09:54.622 }, 00:09:54.622 { 00:09:54.622 "name": "BaseBdev2", 00:09:54.622 "uuid": "17e52b5b-18b7-54bd-8136-2a8f6dc0ba84", 00:09:54.622 "is_configured": true, 00:09:54.622 "data_offset": 2048, 00:09:54.622 "data_size": 63488 00:09:54.622 }, 00:09:54.622 { 00:09:54.622 "name": "BaseBdev3", 00:09:54.622 "uuid": "800af45c-74d3-5390-867b-7f32088f3c97", 00:09:54.622 "is_configured": true, 00:09:54.622 "data_offset": 2048, 00:09:54.622 "data_size": 63488 00:09:54.622 } 00:09:54.622 ] 00:09:54.622 }' 00:09:54.622 17:31:25 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:54.622 17:31:25 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:55.192 17:31:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:09:55.192 17:31:26 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:09:55.192 [2024-11-27 17:31:26.248648] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:09:56.132 17:31:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:09:56.132 17:31:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:56.132 17:31:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:56.132 [2024-11-27 17:31:27.165341] bdev_raid.c:2272:_raid_bdev_fail_base_bdev: *NOTICE*: Failing base bdev in slot 0 ('BaseBdev1') of raid bdev 'raid_bdev1' 00:09:56.132 [2024-11-27 17:31:27.165543] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:09:56.132 [2024-11-27 17:31:27.165820] bdev_raid.c:1970:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 0 raid_ch: 0x60d000002600 00:09:56.132 17:31:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:56.132 17:31:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:09:56.132 17:31:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid1 = \r\a\i\d\1 ]] 00:09:56.132 17:31:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ write = \w\r\i\t\e ]] 00:09:56.132 17:31:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@833 -- # expected_num_base_bdevs=2 00:09:56.133 17:31:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:09:56.133 17:31:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:09:56.133 17:31:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:09:56.133 17:31:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:09:56.133 17:31:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:09:56.133 17:31:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:09:56.133 17:31:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:56.133 17:31:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:56.133 17:31:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:56.133 17:31:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:56.133 17:31:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:56.133 17:31:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:09:56.133 17:31:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:56.133 17:31:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:56.133 17:31:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:56.133 17:31:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:56.133 "name": "raid_bdev1", 00:09:56.133 "uuid": "1a2780d4-6017-4e26-b4d7-36a9c84c6485", 00:09:56.133 "strip_size_kb": 0, 00:09:56.133 "state": "online", 00:09:56.133 "raid_level": "raid1", 00:09:56.133 "superblock": true, 00:09:56.133 "num_base_bdevs": 3, 00:09:56.133 "num_base_bdevs_discovered": 2, 00:09:56.133 "num_base_bdevs_operational": 2, 00:09:56.133 "base_bdevs_list": [ 00:09:56.133 { 00:09:56.133 "name": null, 00:09:56.133 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:56.133 "is_configured": false, 00:09:56.133 "data_offset": 0, 00:09:56.133 "data_size": 63488 00:09:56.133 }, 00:09:56.133 { 00:09:56.133 "name": "BaseBdev2", 00:09:56.133 "uuid": "17e52b5b-18b7-54bd-8136-2a8f6dc0ba84", 00:09:56.133 "is_configured": true, 00:09:56.133 "data_offset": 2048, 00:09:56.133 "data_size": 63488 00:09:56.133 }, 00:09:56.133 { 00:09:56.133 "name": "BaseBdev3", 00:09:56.133 "uuid": "800af45c-74d3-5390-867b-7f32088f3c97", 00:09:56.133 "is_configured": true, 00:09:56.133 "data_offset": 2048, 00:09:56.133 "data_size": 63488 00:09:56.133 } 00:09:56.133 ] 00:09:56.133 }' 00:09:56.133 17:31:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:56.133 17:31:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:56.703 17:31:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:09:56.703 17:31:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:56.703 17:31:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:56.703 [2024-11-27 17:31:27.629676] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:09:56.703 [2024-11-27 17:31:27.629794] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:09:56.703 [2024-11-27 17:31:27.632356] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:09:56.703 [2024-11-27 17:31:27.632467] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:09:56.703 [2024-11-27 17:31:27.632591] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:09:56.703 [2024-11-27 17:31:27.632636] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name raid_bdev1, state offline 00:09:56.703 17:31:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:56.703 { 00:09:56.703 "results": [ 00:09:56.703 { 00:09:56.703 "job": "raid_bdev1", 00:09:56.703 "core_mask": "0x1", 00:09:56.703 "workload": "randrw", 00:09:56.703 "percentage": 50, 00:09:56.703 "status": "finished", 00:09:56.703 "queue_depth": 1, 00:09:56.703 "io_size": 131072, 00:09:56.703 "runtime": 1.381627, 00:09:56.703 "iops": 12934.750117072119, 00:09:56.703 "mibps": 1616.8437646340149, 00:09:56.703 "io_failed": 0, 00:09:56.703 "io_timeout": 0, 00:09:56.703 "avg_latency_us": 74.80686755811115, 00:09:56.703 "min_latency_us": 21.910917030567685, 00:09:56.703 "max_latency_us": 1559.6995633187773 00:09:56.703 } 00:09:56.703 ], 00:09:56.703 "core_count": 1 00:09:56.703 } 00:09:56.703 17:31:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 80068 00:09:56.703 17:31:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@950 -- # '[' -z 80068 ']' 00:09:56.703 17:31:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # kill -0 80068 00:09:56.703 17:31:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@955 -- # uname 00:09:56.703 17:31:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:09:56.703 17:31:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 80068 00:09:56.703 killing process with pid 80068 00:09:56.703 17:31:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:09:56.703 17:31:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:09:56.703 17:31:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 80068' 00:09:56.703 17:31:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@969 -- # kill 80068 00:09:56.703 [2024-11-27 17:31:27.669690] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:09:56.703 17:31:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@974 -- # wait 80068 00:09:56.703 [2024-11-27 17:31:27.718808] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:09:56.964 17:31:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.Tn61xa4aFB 00:09:56.964 17:31:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:09:56.964 17:31:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:09:56.964 17:31:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.00 00:09:56.964 17:31:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid1 00:09:56.964 ************************************ 00:09:56.964 END TEST raid_write_error_test 00:09:56.964 ************************************ 00:09:56.964 17:31:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:09:56.964 17:31:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@199 -- # return 0 00:09:56.964 17:31:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@847 -- # [[ 0.00 = \0\.\0\0 ]] 00:09:56.964 00:09:56.964 real 0m3.418s 00:09:56.964 user 0m4.149s 00:09:56.964 sys 0m0.636s 00:09:56.964 17:31:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:09:56.964 17:31:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:09:56.964 17:31:28 bdev_raid -- bdev/bdev_raid.sh@966 -- # for n in {2..4} 00:09:56.964 17:31:28 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:09:57.225 17:31:28 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test raid0 4 false 00:09:57.225 17:31:28 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:09:57.225 17:31:28 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:09:57.225 17:31:28 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:09:57.225 ************************************ 00:09:57.225 START TEST raid_state_function_test 00:09:57.225 ************************************ 00:09:57.225 17:31:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1125 -- # raid_state_function_test raid0 4 false 00:09:57.225 17:31:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=raid0 00:09:57.225 17:31:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=4 00:09:57.225 17:31:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:09:57.225 17:31:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:09:57.225 17:31:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:09:57.225 17:31:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:57.225 17:31:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:09:57.225 17:31:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:57.225 17:31:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:57.225 17:31:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:09:57.225 17:31:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:57.225 17:31:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:57.225 17:31:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:09:57.225 17:31:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:57.225 17:31:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:57.225 17:31:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev4 00:09:57.225 17:31:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:09:57.225 17:31:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:09:57.225 17:31:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:09:57.225 17:31:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:09:57.225 17:31:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:09:57.225 17:31:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:09:57.225 17:31:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:09:57.225 17:31:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:09:57.225 17:31:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' raid0 '!=' raid1 ']' 00:09:57.225 17:31:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:09:57.225 17:31:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:09:57.225 17:31:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:09:57.225 17:31:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:09:57.225 17:31:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=80201 00:09:57.225 17:31:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:09:57.225 17:31:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 80201' 00:09:57.225 Process raid pid: 80201 00:09:57.225 17:31:28 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 80201 00:09:57.225 17:31:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@831 -- # '[' -z 80201 ']' 00:09:57.225 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:57.225 17:31:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:57.225 17:31:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:09:57.225 17:31:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:57.225 17:31:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:09:57.225 17:31:28 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:57.225 [2024-11-27 17:31:28.267898] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:09:57.225 [2024-11-27 17:31:28.268033] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:09:57.485 [2024-11-27 17:31:28.414858] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:57.485 [2024-11-27 17:31:28.483841] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:09:57.485 [2024-11-27 17:31:28.559504] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:57.485 [2024-11-27 17:31:28.559552] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:09:58.055 17:31:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:09:58.055 17:31:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@864 -- # return 0 00:09:58.055 17:31:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:09:58.055 17:31:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:58.055 17:31:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:58.055 [2024-11-27 17:31:29.098685] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:58.055 [2024-11-27 17:31:29.098858] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:58.055 [2024-11-27 17:31:29.098888] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:58.055 [2024-11-27 17:31:29.098900] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:58.055 [2024-11-27 17:31:29.098906] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:58.055 [2024-11-27 17:31:29.098920] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:58.055 [2024-11-27 17:31:29.098926] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:09:58.055 [2024-11-27 17:31:29.098935] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:09:58.055 17:31:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:58.055 17:31:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:58.055 17:31:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:58.055 17:31:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:58.055 17:31:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:58.055 17:31:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:58.055 17:31:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:58.055 17:31:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:58.055 17:31:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:58.055 17:31:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:58.055 17:31:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:58.055 17:31:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:58.055 17:31:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:58.055 17:31:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:58.055 17:31:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:58.055 17:31:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:58.055 17:31:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:58.055 "name": "Existed_Raid", 00:09:58.055 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:58.055 "strip_size_kb": 64, 00:09:58.055 "state": "configuring", 00:09:58.055 "raid_level": "raid0", 00:09:58.055 "superblock": false, 00:09:58.055 "num_base_bdevs": 4, 00:09:58.055 "num_base_bdevs_discovered": 0, 00:09:58.055 "num_base_bdevs_operational": 4, 00:09:58.055 "base_bdevs_list": [ 00:09:58.055 { 00:09:58.055 "name": "BaseBdev1", 00:09:58.055 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:58.055 "is_configured": false, 00:09:58.055 "data_offset": 0, 00:09:58.055 "data_size": 0 00:09:58.055 }, 00:09:58.055 { 00:09:58.055 "name": "BaseBdev2", 00:09:58.055 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:58.055 "is_configured": false, 00:09:58.055 "data_offset": 0, 00:09:58.055 "data_size": 0 00:09:58.055 }, 00:09:58.055 { 00:09:58.055 "name": "BaseBdev3", 00:09:58.055 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:58.055 "is_configured": false, 00:09:58.055 "data_offset": 0, 00:09:58.055 "data_size": 0 00:09:58.055 }, 00:09:58.055 { 00:09:58.055 "name": "BaseBdev4", 00:09:58.055 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:58.055 "is_configured": false, 00:09:58.055 "data_offset": 0, 00:09:58.055 "data_size": 0 00:09:58.055 } 00:09:58.055 ] 00:09:58.055 }' 00:09:58.055 17:31:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:58.055 17:31:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:58.624 17:31:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:58.624 17:31:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:58.624 17:31:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:58.624 [2024-11-27 17:31:29.533839] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:58.624 [2024-11-27 17:31:29.533951] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:09:58.624 17:31:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:58.624 17:31:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:09:58.624 17:31:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:58.624 17:31:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:58.624 [2024-11-27 17:31:29.545840] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:09:58.624 [2024-11-27 17:31:29.545924] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:09:58.624 [2024-11-27 17:31:29.545966] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:58.624 [2024-11-27 17:31:29.545989] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:58.624 [2024-11-27 17:31:29.546006] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:58.624 [2024-11-27 17:31:29.546028] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:58.624 [2024-11-27 17:31:29.546045] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:09:58.624 [2024-11-27 17:31:29.546066] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:09:58.624 17:31:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:58.624 17:31:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:09:58.624 17:31:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:58.624 17:31:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:58.625 [2024-11-27 17:31:29.572742] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:58.625 BaseBdev1 00:09:58.625 17:31:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:58.625 17:31:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:09:58.625 17:31:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:09:58.625 17:31:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:09:58.625 17:31:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:09:58.625 17:31:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:09:58.625 17:31:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:09:58.625 17:31:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:09:58.625 17:31:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:58.625 17:31:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:58.625 17:31:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:58.625 17:31:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:09:58.625 17:31:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:58.625 17:31:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:58.625 [ 00:09:58.625 { 00:09:58.625 "name": "BaseBdev1", 00:09:58.625 "aliases": [ 00:09:58.625 "3f0a617d-1f56-44ce-9b7c-623a409b0b89" 00:09:58.625 ], 00:09:58.625 "product_name": "Malloc disk", 00:09:58.625 "block_size": 512, 00:09:58.625 "num_blocks": 65536, 00:09:58.625 "uuid": "3f0a617d-1f56-44ce-9b7c-623a409b0b89", 00:09:58.625 "assigned_rate_limits": { 00:09:58.625 "rw_ios_per_sec": 0, 00:09:58.625 "rw_mbytes_per_sec": 0, 00:09:58.625 "r_mbytes_per_sec": 0, 00:09:58.625 "w_mbytes_per_sec": 0 00:09:58.625 }, 00:09:58.625 "claimed": true, 00:09:58.625 "claim_type": "exclusive_write", 00:09:58.625 "zoned": false, 00:09:58.625 "supported_io_types": { 00:09:58.625 "read": true, 00:09:58.625 "write": true, 00:09:58.625 "unmap": true, 00:09:58.625 "flush": true, 00:09:58.625 "reset": true, 00:09:58.625 "nvme_admin": false, 00:09:58.625 "nvme_io": false, 00:09:58.625 "nvme_io_md": false, 00:09:58.625 "write_zeroes": true, 00:09:58.625 "zcopy": true, 00:09:58.625 "get_zone_info": false, 00:09:58.625 "zone_management": false, 00:09:58.625 "zone_append": false, 00:09:58.625 "compare": false, 00:09:58.625 "compare_and_write": false, 00:09:58.625 "abort": true, 00:09:58.625 "seek_hole": false, 00:09:58.625 "seek_data": false, 00:09:58.625 "copy": true, 00:09:58.625 "nvme_iov_md": false 00:09:58.625 }, 00:09:58.625 "memory_domains": [ 00:09:58.625 { 00:09:58.625 "dma_device_id": "system", 00:09:58.625 "dma_device_type": 1 00:09:58.625 }, 00:09:58.625 { 00:09:58.625 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:58.625 "dma_device_type": 2 00:09:58.625 } 00:09:58.625 ], 00:09:58.625 "driver_specific": {} 00:09:58.625 } 00:09:58.625 ] 00:09:58.625 17:31:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:58.625 17:31:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:09:58.625 17:31:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:58.625 17:31:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:58.625 17:31:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:58.625 17:31:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:58.625 17:31:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:58.625 17:31:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:58.625 17:31:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:58.625 17:31:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:58.625 17:31:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:58.625 17:31:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:58.625 17:31:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:58.625 17:31:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:58.625 17:31:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:58.625 17:31:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:58.625 17:31:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:58.625 17:31:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:58.625 "name": "Existed_Raid", 00:09:58.625 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:58.625 "strip_size_kb": 64, 00:09:58.625 "state": "configuring", 00:09:58.625 "raid_level": "raid0", 00:09:58.625 "superblock": false, 00:09:58.625 "num_base_bdevs": 4, 00:09:58.625 "num_base_bdevs_discovered": 1, 00:09:58.625 "num_base_bdevs_operational": 4, 00:09:58.625 "base_bdevs_list": [ 00:09:58.625 { 00:09:58.625 "name": "BaseBdev1", 00:09:58.625 "uuid": "3f0a617d-1f56-44ce-9b7c-623a409b0b89", 00:09:58.625 "is_configured": true, 00:09:58.625 "data_offset": 0, 00:09:58.625 "data_size": 65536 00:09:58.625 }, 00:09:58.625 { 00:09:58.625 "name": "BaseBdev2", 00:09:58.625 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:58.625 "is_configured": false, 00:09:58.625 "data_offset": 0, 00:09:58.625 "data_size": 0 00:09:58.625 }, 00:09:58.625 { 00:09:58.625 "name": "BaseBdev3", 00:09:58.625 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:58.625 "is_configured": false, 00:09:58.625 "data_offset": 0, 00:09:58.625 "data_size": 0 00:09:58.625 }, 00:09:58.625 { 00:09:58.625 "name": "BaseBdev4", 00:09:58.625 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:58.625 "is_configured": false, 00:09:58.625 "data_offset": 0, 00:09:58.625 "data_size": 0 00:09:58.625 } 00:09:58.625 ] 00:09:58.625 }' 00:09:58.625 17:31:29 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:58.625 17:31:29 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:58.885 17:31:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:09:58.885 17:31:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:58.885 17:31:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:58.885 [2024-11-27 17:31:30.012025] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:09:58.885 [2024-11-27 17:31:30.012124] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:09:58.885 17:31:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:58.885 17:31:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:09:58.885 17:31:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:58.885 17:31:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:58.885 [2024-11-27 17:31:30.024074] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:09:58.885 [2024-11-27 17:31:30.026329] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:09:58.885 [2024-11-27 17:31:30.026403] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:09:58.885 [2024-11-27 17:31:30.026431] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:09:58.885 [2024-11-27 17:31:30.026453] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:09:58.885 [2024-11-27 17:31:30.026471] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:09:58.885 [2024-11-27 17:31:30.026491] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:09:58.885 17:31:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:58.885 17:31:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:09:58.885 17:31:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:58.885 17:31:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:58.885 17:31:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:58.885 17:31:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:58.885 17:31:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:58.885 17:31:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:58.885 17:31:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:58.885 17:31:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:58.885 17:31:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:58.885 17:31:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:58.885 17:31:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:58.885 17:31:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:58.885 17:31:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:58.885 17:31:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:58.885 17:31:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:58.885 17:31:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:59.145 17:31:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:59.145 "name": "Existed_Raid", 00:09:59.145 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:59.145 "strip_size_kb": 64, 00:09:59.145 "state": "configuring", 00:09:59.145 "raid_level": "raid0", 00:09:59.145 "superblock": false, 00:09:59.145 "num_base_bdevs": 4, 00:09:59.145 "num_base_bdevs_discovered": 1, 00:09:59.145 "num_base_bdevs_operational": 4, 00:09:59.145 "base_bdevs_list": [ 00:09:59.145 { 00:09:59.145 "name": "BaseBdev1", 00:09:59.145 "uuid": "3f0a617d-1f56-44ce-9b7c-623a409b0b89", 00:09:59.145 "is_configured": true, 00:09:59.145 "data_offset": 0, 00:09:59.145 "data_size": 65536 00:09:59.145 }, 00:09:59.145 { 00:09:59.145 "name": "BaseBdev2", 00:09:59.145 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:59.145 "is_configured": false, 00:09:59.145 "data_offset": 0, 00:09:59.145 "data_size": 0 00:09:59.145 }, 00:09:59.145 { 00:09:59.145 "name": "BaseBdev3", 00:09:59.145 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:59.145 "is_configured": false, 00:09:59.145 "data_offset": 0, 00:09:59.145 "data_size": 0 00:09:59.145 }, 00:09:59.145 { 00:09:59.145 "name": "BaseBdev4", 00:09:59.145 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:59.145 "is_configured": false, 00:09:59.145 "data_offset": 0, 00:09:59.145 "data_size": 0 00:09:59.145 } 00:09:59.145 ] 00:09:59.145 }' 00:09:59.145 17:31:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:59.145 17:31:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:59.406 17:31:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:09:59.406 17:31:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:59.406 17:31:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:59.406 [2024-11-27 17:31:30.529321] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:09:59.406 BaseBdev2 00:09:59.406 17:31:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:59.406 17:31:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:09:59.406 17:31:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:09:59.406 17:31:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:09:59.406 17:31:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:09:59.406 17:31:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:09:59.406 17:31:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:09:59.406 17:31:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:09:59.406 17:31:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:59.406 17:31:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:59.406 17:31:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:59.406 17:31:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:09:59.406 17:31:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:59.406 17:31:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:59.406 [ 00:09:59.406 { 00:09:59.406 "name": "BaseBdev2", 00:09:59.406 "aliases": [ 00:09:59.406 "663380fe-16a7-4393-a318-fd3c93c4b073" 00:09:59.406 ], 00:09:59.406 "product_name": "Malloc disk", 00:09:59.406 "block_size": 512, 00:09:59.406 "num_blocks": 65536, 00:09:59.406 "uuid": "663380fe-16a7-4393-a318-fd3c93c4b073", 00:09:59.406 "assigned_rate_limits": { 00:09:59.406 "rw_ios_per_sec": 0, 00:09:59.406 "rw_mbytes_per_sec": 0, 00:09:59.406 "r_mbytes_per_sec": 0, 00:09:59.406 "w_mbytes_per_sec": 0 00:09:59.406 }, 00:09:59.406 "claimed": true, 00:09:59.406 "claim_type": "exclusive_write", 00:09:59.406 "zoned": false, 00:09:59.406 "supported_io_types": { 00:09:59.406 "read": true, 00:09:59.406 "write": true, 00:09:59.406 "unmap": true, 00:09:59.406 "flush": true, 00:09:59.406 "reset": true, 00:09:59.406 "nvme_admin": false, 00:09:59.406 "nvme_io": false, 00:09:59.406 "nvme_io_md": false, 00:09:59.406 "write_zeroes": true, 00:09:59.406 "zcopy": true, 00:09:59.406 "get_zone_info": false, 00:09:59.406 "zone_management": false, 00:09:59.406 "zone_append": false, 00:09:59.406 "compare": false, 00:09:59.406 "compare_and_write": false, 00:09:59.406 "abort": true, 00:09:59.406 "seek_hole": false, 00:09:59.406 "seek_data": false, 00:09:59.406 "copy": true, 00:09:59.406 "nvme_iov_md": false 00:09:59.406 }, 00:09:59.406 "memory_domains": [ 00:09:59.406 { 00:09:59.406 "dma_device_id": "system", 00:09:59.406 "dma_device_type": 1 00:09:59.406 }, 00:09:59.406 { 00:09:59.406 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:59.406 "dma_device_type": 2 00:09:59.406 } 00:09:59.406 ], 00:09:59.406 "driver_specific": {} 00:09:59.406 } 00:09:59.406 ] 00:09:59.406 17:31:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:59.406 17:31:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:09:59.406 17:31:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:09:59.406 17:31:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:59.406 17:31:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:59.406 17:31:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:59.406 17:31:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:59.406 17:31:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:59.406 17:31:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:59.406 17:31:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:59.406 17:31:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:59.406 17:31:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:59.406 17:31:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:59.406 17:31:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:59.406 17:31:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:59.406 17:31:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:59.406 17:31:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:59.406 17:31:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:59.406 17:31:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:59.666 17:31:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:59.666 "name": "Existed_Raid", 00:09:59.666 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:59.666 "strip_size_kb": 64, 00:09:59.666 "state": "configuring", 00:09:59.666 "raid_level": "raid0", 00:09:59.666 "superblock": false, 00:09:59.666 "num_base_bdevs": 4, 00:09:59.666 "num_base_bdevs_discovered": 2, 00:09:59.666 "num_base_bdevs_operational": 4, 00:09:59.666 "base_bdevs_list": [ 00:09:59.666 { 00:09:59.666 "name": "BaseBdev1", 00:09:59.666 "uuid": "3f0a617d-1f56-44ce-9b7c-623a409b0b89", 00:09:59.666 "is_configured": true, 00:09:59.666 "data_offset": 0, 00:09:59.666 "data_size": 65536 00:09:59.666 }, 00:09:59.666 { 00:09:59.666 "name": "BaseBdev2", 00:09:59.666 "uuid": "663380fe-16a7-4393-a318-fd3c93c4b073", 00:09:59.666 "is_configured": true, 00:09:59.666 "data_offset": 0, 00:09:59.666 "data_size": 65536 00:09:59.667 }, 00:09:59.667 { 00:09:59.667 "name": "BaseBdev3", 00:09:59.667 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:59.667 "is_configured": false, 00:09:59.667 "data_offset": 0, 00:09:59.667 "data_size": 0 00:09:59.667 }, 00:09:59.667 { 00:09:59.667 "name": "BaseBdev4", 00:09:59.667 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:59.667 "is_configured": false, 00:09:59.667 "data_offset": 0, 00:09:59.667 "data_size": 0 00:09:59.667 } 00:09:59.667 ] 00:09:59.667 }' 00:09:59.667 17:31:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:59.667 17:31:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:59.927 17:31:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:09:59.927 17:31:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:59.927 17:31:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:59.927 [2024-11-27 17:31:30.973409] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:09:59.927 BaseBdev3 00:09:59.927 17:31:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:59.927 17:31:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:09:59.927 17:31:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:09:59.927 17:31:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:09:59.927 17:31:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:09:59.927 17:31:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:09:59.927 17:31:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:09:59.927 17:31:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:09:59.927 17:31:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:59.927 17:31:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:59.927 17:31:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:59.927 17:31:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:09:59.927 17:31:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:59.927 17:31:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:59.927 [ 00:09:59.927 { 00:09:59.927 "name": "BaseBdev3", 00:09:59.927 "aliases": [ 00:09:59.927 "e494c7c5-a159-4f18-afc8-bd0d524d2d23" 00:09:59.927 ], 00:09:59.927 "product_name": "Malloc disk", 00:09:59.927 "block_size": 512, 00:09:59.927 "num_blocks": 65536, 00:09:59.927 "uuid": "e494c7c5-a159-4f18-afc8-bd0d524d2d23", 00:09:59.927 "assigned_rate_limits": { 00:09:59.927 "rw_ios_per_sec": 0, 00:09:59.927 "rw_mbytes_per_sec": 0, 00:09:59.927 "r_mbytes_per_sec": 0, 00:09:59.927 "w_mbytes_per_sec": 0 00:09:59.927 }, 00:09:59.927 "claimed": true, 00:09:59.927 "claim_type": "exclusive_write", 00:09:59.927 "zoned": false, 00:09:59.927 "supported_io_types": { 00:09:59.927 "read": true, 00:09:59.927 "write": true, 00:09:59.927 "unmap": true, 00:09:59.927 "flush": true, 00:09:59.927 "reset": true, 00:09:59.927 "nvme_admin": false, 00:09:59.927 "nvme_io": false, 00:09:59.927 "nvme_io_md": false, 00:09:59.927 "write_zeroes": true, 00:09:59.927 "zcopy": true, 00:09:59.927 "get_zone_info": false, 00:09:59.927 "zone_management": false, 00:09:59.927 "zone_append": false, 00:09:59.927 "compare": false, 00:09:59.927 "compare_and_write": false, 00:09:59.927 "abort": true, 00:09:59.927 "seek_hole": false, 00:09:59.927 "seek_data": false, 00:09:59.927 "copy": true, 00:09:59.927 "nvme_iov_md": false 00:09:59.927 }, 00:09:59.927 "memory_domains": [ 00:09:59.927 { 00:09:59.927 "dma_device_id": "system", 00:09:59.928 "dma_device_type": 1 00:09:59.928 }, 00:09:59.928 { 00:09:59.928 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:09:59.928 "dma_device_type": 2 00:09:59.928 } 00:09:59.928 ], 00:09:59.928 "driver_specific": {} 00:09:59.928 } 00:09:59.928 ] 00:09:59.928 17:31:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:59.928 17:31:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:09:59.928 17:31:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:09:59.928 17:31:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:09:59.928 17:31:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:09:59.928 17:31:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:09:59.928 17:31:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:09:59.928 17:31:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:09:59.928 17:31:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:09:59.928 17:31:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:09:59.928 17:31:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:09:59.928 17:31:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:09:59.928 17:31:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:09:59.928 17:31:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:09:59.928 17:31:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:09:59.928 17:31:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:59.928 17:31:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:09:59.928 17:31:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:09:59.928 17:31:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:59.928 17:31:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:09:59.928 "name": "Existed_Raid", 00:09:59.928 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:59.928 "strip_size_kb": 64, 00:09:59.928 "state": "configuring", 00:09:59.928 "raid_level": "raid0", 00:09:59.928 "superblock": false, 00:09:59.928 "num_base_bdevs": 4, 00:09:59.928 "num_base_bdevs_discovered": 3, 00:09:59.928 "num_base_bdevs_operational": 4, 00:09:59.928 "base_bdevs_list": [ 00:09:59.928 { 00:09:59.928 "name": "BaseBdev1", 00:09:59.928 "uuid": "3f0a617d-1f56-44ce-9b7c-623a409b0b89", 00:09:59.928 "is_configured": true, 00:09:59.928 "data_offset": 0, 00:09:59.928 "data_size": 65536 00:09:59.928 }, 00:09:59.928 { 00:09:59.928 "name": "BaseBdev2", 00:09:59.928 "uuid": "663380fe-16a7-4393-a318-fd3c93c4b073", 00:09:59.928 "is_configured": true, 00:09:59.928 "data_offset": 0, 00:09:59.928 "data_size": 65536 00:09:59.928 }, 00:09:59.928 { 00:09:59.928 "name": "BaseBdev3", 00:09:59.928 "uuid": "e494c7c5-a159-4f18-afc8-bd0d524d2d23", 00:09:59.928 "is_configured": true, 00:09:59.928 "data_offset": 0, 00:09:59.928 "data_size": 65536 00:09:59.928 }, 00:09:59.928 { 00:09:59.928 "name": "BaseBdev4", 00:09:59.928 "uuid": "00000000-0000-0000-0000-000000000000", 00:09:59.928 "is_configured": false, 00:09:59.928 "data_offset": 0, 00:09:59.928 "data_size": 0 00:09:59.928 } 00:09:59.928 ] 00:09:59.928 }' 00:09:59.928 17:31:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:09:59.928 17:31:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:00.498 17:31:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:10:00.498 17:31:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:00.498 17:31:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:00.498 [2024-11-27 17:31:31.449426] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:10:00.498 [2024-11-27 17:31:31.449473] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:10:00.498 [2024-11-27 17:31:31.449484] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 262144, blocklen 512 00:10:00.498 [2024-11-27 17:31:31.449801] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:10:00.498 [2024-11-27 17:31:31.449970] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:10:00.498 [2024-11-27 17:31:31.449992] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:10:00.498 [2024-11-27 17:31:31.450230] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:00.498 BaseBdev4 00:10:00.498 17:31:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:00.498 17:31:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev4 00:10:00.498 17:31:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev4 00:10:00.498 17:31:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:10:00.498 17:31:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:10:00.498 17:31:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:10:00.498 17:31:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:10:00.498 17:31:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:10:00.498 17:31:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:00.498 17:31:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:00.498 17:31:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:00.498 17:31:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:10:00.498 17:31:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:00.498 17:31:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:00.498 [ 00:10:00.498 { 00:10:00.498 "name": "BaseBdev4", 00:10:00.498 "aliases": [ 00:10:00.498 "7437379c-b4c2-4c5f-950d-5011e27cc1d8" 00:10:00.498 ], 00:10:00.498 "product_name": "Malloc disk", 00:10:00.498 "block_size": 512, 00:10:00.498 "num_blocks": 65536, 00:10:00.498 "uuid": "7437379c-b4c2-4c5f-950d-5011e27cc1d8", 00:10:00.498 "assigned_rate_limits": { 00:10:00.498 "rw_ios_per_sec": 0, 00:10:00.498 "rw_mbytes_per_sec": 0, 00:10:00.498 "r_mbytes_per_sec": 0, 00:10:00.498 "w_mbytes_per_sec": 0 00:10:00.498 }, 00:10:00.498 "claimed": true, 00:10:00.498 "claim_type": "exclusive_write", 00:10:00.498 "zoned": false, 00:10:00.498 "supported_io_types": { 00:10:00.498 "read": true, 00:10:00.498 "write": true, 00:10:00.498 "unmap": true, 00:10:00.498 "flush": true, 00:10:00.498 "reset": true, 00:10:00.498 "nvme_admin": false, 00:10:00.498 "nvme_io": false, 00:10:00.498 "nvme_io_md": false, 00:10:00.498 "write_zeroes": true, 00:10:00.498 "zcopy": true, 00:10:00.498 "get_zone_info": false, 00:10:00.498 "zone_management": false, 00:10:00.498 "zone_append": false, 00:10:00.498 "compare": false, 00:10:00.498 "compare_and_write": false, 00:10:00.498 "abort": true, 00:10:00.498 "seek_hole": false, 00:10:00.498 "seek_data": false, 00:10:00.498 "copy": true, 00:10:00.498 "nvme_iov_md": false 00:10:00.498 }, 00:10:00.498 "memory_domains": [ 00:10:00.498 { 00:10:00.498 "dma_device_id": "system", 00:10:00.498 "dma_device_type": 1 00:10:00.498 }, 00:10:00.498 { 00:10:00.498 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:00.498 "dma_device_type": 2 00:10:00.498 } 00:10:00.498 ], 00:10:00.498 "driver_specific": {} 00:10:00.498 } 00:10:00.498 ] 00:10:00.498 17:31:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:00.498 17:31:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:10:00.498 17:31:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:10:00.498 17:31:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:00.498 17:31:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid0 64 4 00:10:00.498 17:31:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:00.498 17:31:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:00.498 17:31:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:00.498 17:31:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:00.498 17:31:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:00.498 17:31:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:00.498 17:31:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:00.498 17:31:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:00.498 17:31:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:00.498 17:31:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:00.498 17:31:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:00.498 17:31:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:00.498 17:31:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:00.498 17:31:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:00.498 17:31:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:00.498 "name": "Existed_Raid", 00:10:00.498 "uuid": "845f430a-422e-4c58-8f83-587e1af1ab2f", 00:10:00.498 "strip_size_kb": 64, 00:10:00.498 "state": "online", 00:10:00.498 "raid_level": "raid0", 00:10:00.498 "superblock": false, 00:10:00.498 "num_base_bdevs": 4, 00:10:00.498 "num_base_bdevs_discovered": 4, 00:10:00.498 "num_base_bdevs_operational": 4, 00:10:00.498 "base_bdevs_list": [ 00:10:00.498 { 00:10:00.498 "name": "BaseBdev1", 00:10:00.498 "uuid": "3f0a617d-1f56-44ce-9b7c-623a409b0b89", 00:10:00.499 "is_configured": true, 00:10:00.499 "data_offset": 0, 00:10:00.499 "data_size": 65536 00:10:00.499 }, 00:10:00.499 { 00:10:00.499 "name": "BaseBdev2", 00:10:00.499 "uuid": "663380fe-16a7-4393-a318-fd3c93c4b073", 00:10:00.499 "is_configured": true, 00:10:00.499 "data_offset": 0, 00:10:00.499 "data_size": 65536 00:10:00.499 }, 00:10:00.499 { 00:10:00.499 "name": "BaseBdev3", 00:10:00.499 "uuid": "e494c7c5-a159-4f18-afc8-bd0d524d2d23", 00:10:00.499 "is_configured": true, 00:10:00.499 "data_offset": 0, 00:10:00.499 "data_size": 65536 00:10:00.499 }, 00:10:00.499 { 00:10:00.499 "name": "BaseBdev4", 00:10:00.499 "uuid": "7437379c-b4c2-4c5f-950d-5011e27cc1d8", 00:10:00.499 "is_configured": true, 00:10:00.499 "data_offset": 0, 00:10:00.499 "data_size": 65536 00:10:00.499 } 00:10:00.499 ] 00:10:00.499 }' 00:10:00.499 17:31:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:00.499 17:31:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:00.759 17:31:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:10:00.759 17:31:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:10:00.759 17:31:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:10:00.759 17:31:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:10:00.759 17:31:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:10:00.759 17:31:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:10:00.759 17:31:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:10:00.759 17:31:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:10:00.759 17:31:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:00.759 17:31:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:00.759 [2024-11-27 17:31:31.916962] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:00.759 17:31:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:01.019 17:31:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:01.019 "name": "Existed_Raid", 00:10:01.019 "aliases": [ 00:10:01.019 "845f430a-422e-4c58-8f83-587e1af1ab2f" 00:10:01.019 ], 00:10:01.019 "product_name": "Raid Volume", 00:10:01.019 "block_size": 512, 00:10:01.019 "num_blocks": 262144, 00:10:01.019 "uuid": "845f430a-422e-4c58-8f83-587e1af1ab2f", 00:10:01.019 "assigned_rate_limits": { 00:10:01.019 "rw_ios_per_sec": 0, 00:10:01.019 "rw_mbytes_per_sec": 0, 00:10:01.019 "r_mbytes_per_sec": 0, 00:10:01.019 "w_mbytes_per_sec": 0 00:10:01.019 }, 00:10:01.019 "claimed": false, 00:10:01.019 "zoned": false, 00:10:01.019 "supported_io_types": { 00:10:01.019 "read": true, 00:10:01.019 "write": true, 00:10:01.019 "unmap": true, 00:10:01.019 "flush": true, 00:10:01.019 "reset": true, 00:10:01.019 "nvme_admin": false, 00:10:01.019 "nvme_io": false, 00:10:01.019 "nvme_io_md": false, 00:10:01.019 "write_zeroes": true, 00:10:01.019 "zcopy": false, 00:10:01.019 "get_zone_info": false, 00:10:01.019 "zone_management": false, 00:10:01.019 "zone_append": false, 00:10:01.019 "compare": false, 00:10:01.019 "compare_and_write": false, 00:10:01.019 "abort": false, 00:10:01.019 "seek_hole": false, 00:10:01.019 "seek_data": false, 00:10:01.019 "copy": false, 00:10:01.019 "nvme_iov_md": false 00:10:01.019 }, 00:10:01.019 "memory_domains": [ 00:10:01.019 { 00:10:01.019 "dma_device_id": "system", 00:10:01.019 "dma_device_type": 1 00:10:01.019 }, 00:10:01.019 { 00:10:01.019 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:01.019 "dma_device_type": 2 00:10:01.019 }, 00:10:01.019 { 00:10:01.019 "dma_device_id": "system", 00:10:01.019 "dma_device_type": 1 00:10:01.019 }, 00:10:01.019 { 00:10:01.019 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:01.019 "dma_device_type": 2 00:10:01.019 }, 00:10:01.019 { 00:10:01.019 "dma_device_id": "system", 00:10:01.019 "dma_device_type": 1 00:10:01.019 }, 00:10:01.019 { 00:10:01.019 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:01.019 "dma_device_type": 2 00:10:01.019 }, 00:10:01.019 { 00:10:01.019 "dma_device_id": "system", 00:10:01.019 "dma_device_type": 1 00:10:01.019 }, 00:10:01.019 { 00:10:01.019 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:01.019 "dma_device_type": 2 00:10:01.019 } 00:10:01.019 ], 00:10:01.019 "driver_specific": { 00:10:01.019 "raid": { 00:10:01.019 "uuid": "845f430a-422e-4c58-8f83-587e1af1ab2f", 00:10:01.020 "strip_size_kb": 64, 00:10:01.020 "state": "online", 00:10:01.020 "raid_level": "raid0", 00:10:01.020 "superblock": false, 00:10:01.020 "num_base_bdevs": 4, 00:10:01.020 "num_base_bdevs_discovered": 4, 00:10:01.020 "num_base_bdevs_operational": 4, 00:10:01.020 "base_bdevs_list": [ 00:10:01.020 { 00:10:01.020 "name": "BaseBdev1", 00:10:01.020 "uuid": "3f0a617d-1f56-44ce-9b7c-623a409b0b89", 00:10:01.020 "is_configured": true, 00:10:01.020 "data_offset": 0, 00:10:01.020 "data_size": 65536 00:10:01.020 }, 00:10:01.020 { 00:10:01.020 "name": "BaseBdev2", 00:10:01.020 "uuid": "663380fe-16a7-4393-a318-fd3c93c4b073", 00:10:01.020 "is_configured": true, 00:10:01.020 "data_offset": 0, 00:10:01.020 "data_size": 65536 00:10:01.020 }, 00:10:01.020 { 00:10:01.020 "name": "BaseBdev3", 00:10:01.020 "uuid": "e494c7c5-a159-4f18-afc8-bd0d524d2d23", 00:10:01.020 "is_configured": true, 00:10:01.020 "data_offset": 0, 00:10:01.020 "data_size": 65536 00:10:01.020 }, 00:10:01.020 { 00:10:01.020 "name": "BaseBdev4", 00:10:01.020 "uuid": "7437379c-b4c2-4c5f-950d-5011e27cc1d8", 00:10:01.020 "is_configured": true, 00:10:01.020 "data_offset": 0, 00:10:01.020 "data_size": 65536 00:10:01.020 } 00:10:01.020 ] 00:10:01.020 } 00:10:01.020 } 00:10:01.020 }' 00:10:01.020 17:31:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:01.020 17:31:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:10:01.020 BaseBdev2 00:10:01.020 BaseBdev3 00:10:01.020 BaseBdev4' 00:10:01.020 17:31:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:01.020 17:31:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:01.020 17:31:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:01.020 17:31:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:10:01.020 17:31:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:01.020 17:31:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:01.020 17:31:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:01.020 17:31:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:01.020 17:31:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:01.020 17:31:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:01.020 17:31:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:01.020 17:31:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:01.020 17:31:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:10:01.020 17:31:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:01.020 17:31:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:01.020 17:31:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:01.020 17:31:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:01.020 17:31:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:01.020 17:31:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:01.020 17:31:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:10:01.020 17:31:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:01.020 17:31:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:01.020 17:31:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:01.020 17:31:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:01.020 17:31:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:01.020 17:31:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:01.020 17:31:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:01.020 17:31:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:10:01.020 17:31:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:01.020 17:31:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:01.020 17:31:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:01.020 17:31:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:01.280 17:31:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:01.280 17:31:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:01.280 17:31:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:10:01.280 17:31:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:01.280 17:31:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:01.280 [2024-11-27 17:31:32.248161] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:10:01.280 [2024-11-27 17:31:32.248191] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:01.280 [2024-11-27 17:31:32.248252] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:01.280 17:31:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:01.280 17:31:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:10:01.280 17:31:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy raid0 00:10:01.280 17:31:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:10:01.280 17:31:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # return 1 00:10:01.280 17:31:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:10:01.280 17:31:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline raid0 64 3 00:10:01.280 17:31:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:01.280 17:31:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:10:01.280 17:31:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:01.280 17:31:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:01.280 17:31:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:01.280 17:31:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:01.280 17:31:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:01.280 17:31:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:01.280 17:31:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:01.280 17:31:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:01.280 17:31:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:01.280 17:31:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:01.280 17:31:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:01.280 17:31:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:01.280 17:31:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:01.280 "name": "Existed_Raid", 00:10:01.280 "uuid": "845f430a-422e-4c58-8f83-587e1af1ab2f", 00:10:01.280 "strip_size_kb": 64, 00:10:01.280 "state": "offline", 00:10:01.280 "raid_level": "raid0", 00:10:01.280 "superblock": false, 00:10:01.280 "num_base_bdevs": 4, 00:10:01.280 "num_base_bdevs_discovered": 3, 00:10:01.280 "num_base_bdevs_operational": 3, 00:10:01.280 "base_bdevs_list": [ 00:10:01.280 { 00:10:01.280 "name": null, 00:10:01.280 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:01.280 "is_configured": false, 00:10:01.280 "data_offset": 0, 00:10:01.280 "data_size": 65536 00:10:01.280 }, 00:10:01.280 { 00:10:01.280 "name": "BaseBdev2", 00:10:01.280 "uuid": "663380fe-16a7-4393-a318-fd3c93c4b073", 00:10:01.280 "is_configured": true, 00:10:01.280 "data_offset": 0, 00:10:01.280 "data_size": 65536 00:10:01.280 }, 00:10:01.280 { 00:10:01.280 "name": "BaseBdev3", 00:10:01.280 "uuid": "e494c7c5-a159-4f18-afc8-bd0d524d2d23", 00:10:01.280 "is_configured": true, 00:10:01.280 "data_offset": 0, 00:10:01.280 "data_size": 65536 00:10:01.280 }, 00:10:01.280 { 00:10:01.280 "name": "BaseBdev4", 00:10:01.280 "uuid": "7437379c-b4c2-4c5f-950d-5011e27cc1d8", 00:10:01.280 "is_configured": true, 00:10:01.280 "data_offset": 0, 00:10:01.280 "data_size": 65536 00:10:01.280 } 00:10:01.280 ] 00:10:01.280 }' 00:10:01.280 17:31:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:01.280 17:31:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:01.541 17:31:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:10:01.541 17:31:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:01.541 17:31:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:01.541 17:31:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:10:01.541 17:31:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:01.541 17:31:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:01.541 17:31:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:01.801 17:31:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:10:01.801 17:31:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:10:01.801 17:31:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:10:01.801 17:31:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:01.801 17:31:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:01.801 [2024-11-27 17:31:32.744021] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:10:01.801 17:31:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:01.801 17:31:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:10:01.801 17:31:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:01.801 17:31:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:01.801 17:31:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:01.801 17:31:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:01.801 17:31:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:10:01.801 17:31:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:01.801 17:31:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:10:01.801 17:31:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:10:01.801 17:31:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:10:01.801 17:31:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:01.801 17:31:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:01.801 [2024-11-27 17:31:32.812462] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:10:01.801 17:31:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:01.801 17:31:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:10:01.801 17:31:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:01.801 17:31:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:01.801 17:31:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:10:01.801 17:31:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:01.801 17:31:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:01.801 17:31:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:01.801 17:31:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:10:01.802 17:31:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:10:01.802 17:31:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev4 00:10:01.802 17:31:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:01.802 17:31:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:01.802 [2024-11-27 17:31:32.888785] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev4 00:10:01.802 [2024-11-27 17:31:32.888835] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:10:01.802 17:31:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:01.802 17:31:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:10:01.802 17:31:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:01.802 17:31:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:01.802 17:31:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:10:01.802 17:31:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:01.802 17:31:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:01.802 17:31:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:01.802 17:31:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:10:01.802 17:31:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:10:01.802 17:31:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 4 -gt 2 ']' 00:10:01.802 17:31:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:10:01.802 17:31:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:01.802 17:31:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:10:01.802 17:31:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:01.802 17:31:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:01.802 BaseBdev2 00:10:01.802 17:31:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:01.802 17:31:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:10:01.802 17:31:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:10:01.802 17:31:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:10:01.802 17:31:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:10:01.802 17:31:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:10:01.802 17:31:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:10:01.802 17:31:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:10:01.802 17:31:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:01.802 17:31:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:02.063 17:31:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:02.063 17:31:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:10:02.063 17:31:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:02.063 17:31:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:02.063 [ 00:10:02.063 { 00:10:02.063 "name": "BaseBdev2", 00:10:02.063 "aliases": [ 00:10:02.063 "7e99e0df-38d1-40ba-88a4-aa9fb968df07" 00:10:02.063 ], 00:10:02.063 "product_name": "Malloc disk", 00:10:02.063 "block_size": 512, 00:10:02.063 "num_blocks": 65536, 00:10:02.063 "uuid": "7e99e0df-38d1-40ba-88a4-aa9fb968df07", 00:10:02.063 "assigned_rate_limits": { 00:10:02.063 "rw_ios_per_sec": 0, 00:10:02.063 "rw_mbytes_per_sec": 0, 00:10:02.063 "r_mbytes_per_sec": 0, 00:10:02.063 "w_mbytes_per_sec": 0 00:10:02.063 }, 00:10:02.063 "claimed": false, 00:10:02.063 "zoned": false, 00:10:02.063 "supported_io_types": { 00:10:02.063 "read": true, 00:10:02.063 "write": true, 00:10:02.063 "unmap": true, 00:10:02.063 "flush": true, 00:10:02.063 "reset": true, 00:10:02.063 "nvme_admin": false, 00:10:02.063 "nvme_io": false, 00:10:02.063 "nvme_io_md": false, 00:10:02.063 "write_zeroes": true, 00:10:02.063 "zcopy": true, 00:10:02.063 "get_zone_info": false, 00:10:02.063 "zone_management": false, 00:10:02.063 "zone_append": false, 00:10:02.063 "compare": false, 00:10:02.063 "compare_and_write": false, 00:10:02.063 "abort": true, 00:10:02.063 "seek_hole": false, 00:10:02.063 "seek_data": false, 00:10:02.063 "copy": true, 00:10:02.063 "nvme_iov_md": false 00:10:02.063 }, 00:10:02.063 "memory_domains": [ 00:10:02.063 { 00:10:02.063 "dma_device_id": "system", 00:10:02.063 "dma_device_type": 1 00:10:02.063 }, 00:10:02.063 { 00:10:02.063 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:02.063 "dma_device_type": 2 00:10:02.063 } 00:10:02.063 ], 00:10:02.063 "driver_specific": {} 00:10:02.063 } 00:10:02.063 ] 00:10:02.063 17:31:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:02.063 17:31:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:10:02.063 17:31:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:10:02.063 17:31:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:02.063 17:31:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:10:02.063 17:31:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:02.063 17:31:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:02.063 BaseBdev3 00:10:02.063 17:31:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:02.063 17:31:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:10:02.063 17:31:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:10:02.063 17:31:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:10:02.063 17:31:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:10:02.063 17:31:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:10:02.063 17:31:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:10:02.063 17:31:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:10:02.063 17:31:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:02.063 17:31:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:02.063 17:31:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:02.063 17:31:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:10:02.063 17:31:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:02.063 17:31:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:02.063 [ 00:10:02.063 { 00:10:02.063 "name": "BaseBdev3", 00:10:02.063 "aliases": [ 00:10:02.063 "f3442a73-a294-498e-8526-e7f69d3b5043" 00:10:02.063 ], 00:10:02.063 "product_name": "Malloc disk", 00:10:02.063 "block_size": 512, 00:10:02.063 "num_blocks": 65536, 00:10:02.063 "uuid": "f3442a73-a294-498e-8526-e7f69d3b5043", 00:10:02.063 "assigned_rate_limits": { 00:10:02.063 "rw_ios_per_sec": 0, 00:10:02.063 "rw_mbytes_per_sec": 0, 00:10:02.063 "r_mbytes_per_sec": 0, 00:10:02.063 "w_mbytes_per_sec": 0 00:10:02.063 }, 00:10:02.063 "claimed": false, 00:10:02.063 "zoned": false, 00:10:02.063 "supported_io_types": { 00:10:02.063 "read": true, 00:10:02.063 "write": true, 00:10:02.063 "unmap": true, 00:10:02.063 "flush": true, 00:10:02.063 "reset": true, 00:10:02.063 "nvme_admin": false, 00:10:02.063 "nvme_io": false, 00:10:02.063 "nvme_io_md": false, 00:10:02.063 "write_zeroes": true, 00:10:02.063 "zcopy": true, 00:10:02.063 "get_zone_info": false, 00:10:02.063 "zone_management": false, 00:10:02.063 "zone_append": false, 00:10:02.063 "compare": false, 00:10:02.063 "compare_and_write": false, 00:10:02.063 "abort": true, 00:10:02.063 "seek_hole": false, 00:10:02.063 "seek_data": false, 00:10:02.063 "copy": true, 00:10:02.063 "nvme_iov_md": false 00:10:02.063 }, 00:10:02.063 "memory_domains": [ 00:10:02.063 { 00:10:02.063 "dma_device_id": "system", 00:10:02.063 "dma_device_type": 1 00:10:02.063 }, 00:10:02.063 { 00:10:02.063 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:02.063 "dma_device_type": 2 00:10:02.063 } 00:10:02.063 ], 00:10:02.063 "driver_specific": {} 00:10:02.063 } 00:10:02.063 ] 00:10:02.063 17:31:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:02.063 17:31:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:10:02.063 17:31:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:10:02.063 17:31:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:02.063 17:31:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:10:02.063 17:31:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:02.063 17:31:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:02.063 BaseBdev4 00:10:02.064 17:31:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:02.064 17:31:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev4 00:10:02.064 17:31:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev4 00:10:02.064 17:31:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:10:02.064 17:31:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:10:02.064 17:31:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:10:02.064 17:31:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:10:02.064 17:31:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:10:02.064 17:31:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:02.064 17:31:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:02.064 17:31:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:02.064 17:31:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:10:02.064 17:31:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:02.064 17:31:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:02.064 [ 00:10:02.064 { 00:10:02.064 "name": "BaseBdev4", 00:10:02.064 "aliases": [ 00:10:02.064 "2801f3ec-cc2a-4df3-8ad8-d13266d87152" 00:10:02.064 ], 00:10:02.064 "product_name": "Malloc disk", 00:10:02.064 "block_size": 512, 00:10:02.064 "num_blocks": 65536, 00:10:02.064 "uuid": "2801f3ec-cc2a-4df3-8ad8-d13266d87152", 00:10:02.064 "assigned_rate_limits": { 00:10:02.064 "rw_ios_per_sec": 0, 00:10:02.064 "rw_mbytes_per_sec": 0, 00:10:02.064 "r_mbytes_per_sec": 0, 00:10:02.064 "w_mbytes_per_sec": 0 00:10:02.064 }, 00:10:02.064 "claimed": false, 00:10:02.064 "zoned": false, 00:10:02.064 "supported_io_types": { 00:10:02.064 "read": true, 00:10:02.064 "write": true, 00:10:02.064 "unmap": true, 00:10:02.064 "flush": true, 00:10:02.064 "reset": true, 00:10:02.064 "nvme_admin": false, 00:10:02.064 "nvme_io": false, 00:10:02.064 "nvme_io_md": false, 00:10:02.064 "write_zeroes": true, 00:10:02.064 "zcopy": true, 00:10:02.064 "get_zone_info": false, 00:10:02.064 "zone_management": false, 00:10:02.064 "zone_append": false, 00:10:02.064 "compare": false, 00:10:02.064 "compare_and_write": false, 00:10:02.064 "abort": true, 00:10:02.064 "seek_hole": false, 00:10:02.064 "seek_data": false, 00:10:02.064 "copy": true, 00:10:02.064 "nvme_iov_md": false 00:10:02.064 }, 00:10:02.064 "memory_domains": [ 00:10:02.064 { 00:10:02.064 "dma_device_id": "system", 00:10:02.064 "dma_device_type": 1 00:10:02.064 }, 00:10:02.064 { 00:10:02.064 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:02.064 "dma_device_type": 2 00:10:02.064 } 00:10:02.064 ], 00:10:02.064 "driver_specific": {} 00:10:02.064 } 00:10:02.064 ] 00:10:02.064 17:31:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:02.064 17:31:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:10:02.064 17:31:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:10:02.064 17:31:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:02.064 17:31:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:10:02.064 17:31:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:02.064 17:31:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:02.064 [2024-11-27 17:31:33.145290] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:10:02.064 [2024-11-27 17:31:33.145390] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:10:02.064 [2024-11-27 17:31:33.145449] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:02.064 [2024-11-27 17:31:33.147577] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:02.064 [2024-11-27 17:31:33.147666] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:10:02.064 17:31:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:02.064 17:31:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:10:02.064 17:31:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:02.064 17:31:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:02.064 17:31:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:02.064 17:31:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:02.064 17:31:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:02.064 17:31:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:02.064 17:31:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:02.064 17:31:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:02.064 17:31:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:02.064 17:31:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:02.064 17:31:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:02.064 17:31:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:02.064 17:31:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:02.064 17:31:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:02.064 17:31:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:02.064 "name": "Existed_Raid", 00:10:02.064 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:02.064 "strip_size_kb": 64, 00:10:02.064 "state": "configuring", 00:10:02.064 "raid_level": "raid0", 00:10:02.064 "superblock": false, 00:10:02.064 "num_base_bdevs": 4, 00:10:02.064 "num_base_bdevs_discovered": 3, 00:10:02.064 "num_base_bdevs_operational": 4, 00:10:02.064 "base_bdevs_list": [ 00:10:02.064 { 00:10:02.064 "name": "BaseBdev1", 00:10:02.064 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:02.064 "is_configured": false, 00:10:02.064 "data_offset": 0, 00:10:02.064 "data_size": 0 00:10:02.064 }, 00:10:02.064 { 00:10:02.064 "name": "BaseBdev2", 00:10:02.064 "uuid": "7e99e0df-38d1-40ba-88a4-aa9fb968df07", 00:10:02.064 "is_configured": true, 00:10:02.064 "data_offset": 0, 00:10:02.064 "data_size": 65536 00:10:02.064 }, 00:10:02.064 { 00:10:02.064 "name": "BaseBdev3", 00:10:02.064 "uuid": "f3442a73-a294-498e-8526-e7f69d3b5043", 00:10:02.064 "is_configured": true, 00:10:02.064 "data_offset": 0, 00:10:02.064 "data_size": 65536 00:10:02.064 }, 00:10:02.064 { 00:10:02.064 "name": "BaseBdev4", 00:10:02.064 "uuid": "2801f3ec-cc2a-4df3-8ad8-d13266d87152", 00:10:02.064 "is_configured": true, 00:10:02.064 "data_offset": 0, 00:10:02.064 "data_size": 65536 00:10:02.064 } 00:10:02.064 ] 00:10:02.064 }' 00:10:02.064 17:31:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:02.064 17:31:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:02.634 17:31:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:10:02.634 17:31:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:02.634 17:31:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:02.634 [2024-11-27 17:31:33.616458] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:10:02.634 17:31:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:02.634 17:31:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:10:02.634 17:31:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:02.634 17:31:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:02.634 17:31:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:02.634 17:31:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:02.634 17:31:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:02.634 17:31:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:02.634 17:31:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:02.634 17:31:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:02.634 17:31:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:02.634 17:31:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:02.634 17:31:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:02.634 17:31:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:02.634 17:31:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:02.634 17:31:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:02.634 17:31:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:02.634 "name": "Existed_Raid", 00:10:02.634 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:02.634 "strip_size_kb": 64, 00:10:02.634 "state": "configuring", 00:10:02.634 "raid_level": "raid0", 00:10:02.634 "superblock": false, 00:10:02.634 "num_base_bdevs": 4, 00:10:02.634 "num_base_bdevs_discovered": 2, 00:10:02.634 "num_base_bdevs_operational": 4, 00:10:02.634 "base_bdevs_list": [ 00:10:02.634 { 00:10:02.634 "name": "BaseBdev1", 00:10:02.634 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:02.634 "is_configured": false, 00:10:02.634 "data_offset": 0, 00:10:02.634 "data_size": 0 00:10:02.634 }, 00:10:02.634 { 00:10:02.634 "name": null, 00:10:02.634 "uuid": "7e99e0df-38d1-40ba-88a4-aa9fb968df07", 00:10:02.634 "is_configured": false, 00:10:02.634 "data_offset": 0, 00:10:02.634 "data_size": 65536 00:10:02.635 }, 00:10:02.635 { 00:10:02.635 "name": "BaseBdev3", 00:10:02.635 "uuid": "f3442a73-a294-498e-8526-e7f69d3b5043", 00:10:02.635 "is_configured": true, 00:10:02.635 "data_offset": 0, 00:10:02.635 "data_size": 65536 00:10:02.635 }, 00:10:02.635 { 00:10:02.635 "name": "BaseBdev4", 00:10:02.635 "uuid": "2801f3ec-cc2a-4df3-8ad8-d13266d87152", 00:10:02.635 "is_configured": true, 00:10:02.635 "data_offset": 0, 00:10:02.635 "data_size": 65536 00:10:02.635 } 00:10:02.635 ] 00:10:02.635 }' 00:10:02.635 17:31:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:02.635 17:31:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:02.895 17:31:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:02.895 17:31:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:02.895 17:31:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:02.895 17:31:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:10:02.895 17:31:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:03.155 17:31:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:10:03.155 17:31:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:10:03.155 17:31:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:03.155 17:31:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:03.155 [2024-11-27 17:31:34.132421] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:03.155 BaseBdev1 00:10:03.155 17:31:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:03.155 17:31:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:10:03.155 17:31:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:10:03.155 17:31:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:10:03.155 17:31:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:10:03.155 17:31:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:10:03.155 17:31:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:10:03.155 17:31:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:10:03.155 17:31:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:03.155 17:31:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:03.155 17:31:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:03.155 17:31:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:10:03.155 17:31:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:03.155 17:31:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:03.155 [ 00:10:03.155 { 00:10:03.155 "name": "BaseBdev1", 00:10:03.155 "aliases": [ 00:10:03.155 "7ef8bd30-ab64-4d06-b669-d6d6a3e84206" 00:10:03.155 ], 00:10:03.155 "product_name": "Malloc disk", 00:10:03.155 "block_size": 512, 00:10:03.155 "num_blocks": 65536, 00:10:03.155 "uuid": "7ef8bd30-ab64-4d06-b669-d6d6a3e84206", 00:10:03.155 "assigned_rate_limits": { 00:10:03.155 "rw_ios_per_sec": 0, 00:10:03.155 "rw_mbytes_per_sec": 0, 00:10:03.155 "r_mbytes_per_sec": 0, 00:10:03.155 "w_mbytes_per_sec": 0 00:10:03.155 }, 00:10:03.155 "claimed": true, 00:10:03.155 "claim_type": "exclusive_write", 00:10:03.155 "zoned": false, 00:10:03.155 "supported_io_types": { 00:10:03.155 "read": true, 00:10:03.155 "write": true, 00:10:03.155 "unmap": true, 00:10:03.155 "flush": true, 00:10:03.155 "reset": true, 00:10:03.155 "nvme_admin": false, 00:10:03.155 "nvme_io": false, 00:10:03.155 "nvme_io_md": false, 00:10:03.155 "write_zeroes": true, 00:10:03.155 "zcopy": true, 00:10:03.155 "get_zone_info": false, 00:10:03.155 "zone_management": false, 00:10:03.155 "zone_append": false, 00:10:03.155 "compare": false, 00:10:03.155 "compare_and_write": false, 00:10:03.155 "abort": true, 00:10:03.155 "seek_hole": false, 00:10:03.155 "seek_data": false, 00:10:03.155 "copy": true, 00:10:03.155 "nvme_iov_md": false 00:10:03.155 }, 00:10:03.155 "memory_domains": [ 00:10:03.155 { 00:10:03.155 "dma_device_id": "system", 00:10:03.155 "dma_device_type": 1 00:10:03.155 }, 00:10:03.155 { 00:10:03.155 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:03.155 "dma_device_type": 2 00:10:03.155 } 00:10:03.155 ], 00:10:03.155 "driver_specific": {} 00:10:03.155 } 00:10:03.155 ] 00:10:03.155 17:31:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:03.155 17:31:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:10:03.155 17:31:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:10:03.155 17:31:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:03.155 17:31:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:03.155 17:31:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:03.155 17:31:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:03.155 17:31:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:03.155 17:31:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:03.155 17:31:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:03.155 17:31:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:03.155 17:31:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:03.155 17:31:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:03.155 17:31:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:03.155 17:31:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:03.155 17:31:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:03.155 17:31:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:03.155 17:31:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:03.155 "name": "Existed_Raid", 00:10:03.155 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:03.155 "strip_size_kb": 64, 00:10:03.155 "state": "configuring", 00:10:03.155 "raid_level": "raid0", 00:10:03.155 "superblock": false, 00:10:03.155 "num_base_bdevs": 4, 00:10:03.155 "num_base_bdevs_discovered": 3, 00:10:03.155 "num_base_bdevs_operational": 4, 00:10:03.155 "base_bdevs_list": [ 00:10:03.155 { 00:10:03.155 "name": "BaseBdev1", 00:10:03.155 "uuid": "7ef8bd30-ab64-4d06-b669-d6d6a3e84206", 00:10:03.155 "is_configured": true, 00:10:03.155 "data_offset": 0, 00:10:03.155 "data_size": 65536 00:10:03.155 }, 00:10:03.155 { 00:10:03.155 "name": null, 00:10:03.155 "uuid": "7e99e0df-38d1-40ba-88a4-aa9fb968df07", 00:10:03.155 "is_configured": false, 00:10:03.155 "data_offset": 0, 00:10:03.155 "data_size": 65536 00:10:03.155 }, 00:10:03.155 { 00:10:03.155 "name": "BaseBdev3", 00:10:03.155 "uuid": "f3442a73-a294-498e-8526-e7f69d3b5043", 00:10:03.155 "is_configured": true, 00:10:03.155 "data_offset": 0, 00:10:03.155 "data_size": 65536 00:10:03.155 }, 00:10:03.155 { 00:10:03.155 "name": "BaseBdev4", 00:10:03.155 "uuid": "2801f3ec-cc2a-4df3-8ad8-d13266d87152", 00:10:03.155 "is_configured": true, 00:10:03.155 "data_offset": 0, 00:10:03.155 "data_size": 65536 00:10:03.155 } 00:10:03.155 ] 00:10:03.155 }' 00:10:03.155 17:31:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:03.155 17:31:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:03.415 17:31:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:10:03.415 17:31:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:03.415 17:31:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:03.415 17:31:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:03.675 17:31:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:03.675 17:31:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:10:03.675 17:31:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:10:03.675 17:31:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:03.675 17:31:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:03.675 [2024-11-27 17:31:34.651574] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:10:03.675 17:31:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:03.676 17:31:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:10:03.676 17:31:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:03.676 17:31:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:03.676 17:31:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:03.676 17:31:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:03.676 17:31:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:03.676 17:31:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:03.676 17:31:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:03.676 17:31:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:03.676 17:31:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:03.676 17:31:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:03.676 17:31:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:03.676 17:31:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:03.676 17:31:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:03.676 17:31:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:03.676 17:31:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:03.676 "name": "Existed_Raid", 00:10:03.676 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:03.676 "strip_size_kb": 64, 00:10:03.676 "state": "configuring", 00:10:03.676 "raid_level": "raid0", 00:10:03.676 "superblock": false, 00:10:03.676 "num_base_bdevs": 4, 00:10:03.676 "num_base_bdevs_discovered": 2, 00:10:03.676 "num_base_bdevs_operational": 4, 00:10:03.676 "base_bdevs_list": [ 00:10:03.676 { 00:10:03.676 "name": "BaseBdev1", 00:10:03.676 "uuid": "7ef8bd30-ab64-4d06-b669-d6d6a3e84206", 00:10:03.676 "is_configured": true, 00:10:03.676 "data_offset": 0, 00:10:03.676 "data_size": 65536 00:10:03.676 }, 00:10:03.676 { 00:10:03.676 "name": null, 00:10:03.676 "uuid": "7e99e0df-38d1-40ba-88a4-aa9fb968df07", 00:10:03.676 "is_configured": false, 00:10:03.676 "data_offset": 0, 00:10:03.676 "data_size": 65536 00:10:03.676 }, 00:10:03.676 { 00:10:03.676 "name": null, 00:10:03.676 "uuid": "f3442a73-a294-498e-8526-e7f69d3b5043", 00:10:03.676 "is_configured": false, 00:10:03.676 "data_offset": 0, 00:10:03.676 "data_size": 65536 00:10:03.676 }, 00:10:03.676 { 00:10:03.676 "name": "BaseBdev4", 00:10:03.676 "uuid": "2801f3ec-cc2a-4df3-8ad8-d13266d87152", 00:10:03.676 "is_configured": true, 00:10:03.676 "data_offset": 0, 00:10:03.676 "data_size": 65536 00:10:03.676 } 00:10:03.676 ] 00:10:03.676 }' 00:10:03.676 17:31:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:03.676 17:31:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:03.941 17:31:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:03.941 17:31:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:10:03.941 17:31:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:03.941 17:31:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:03.941 17:31:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:03.941 17:31:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:10:03.941 17:31:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:10:03.941 17:31:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:03.941 17:31:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:04.208 [2024-11-27 17:31:35.130892] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:04.208 17:31:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:04.208 17:31:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:10:04.208 17:31:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:04.208 17:31:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:04.208 17:31:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:04.208 17:31:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:04.208 17:31:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:04.208 17:31:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:04.208 17:31:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:04.208 17:31:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:04.208 17:31:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:04.208 17:31:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:04.208 17:31:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:04.208 17:31:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:04.208 17:31:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:04.208 17:31:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:04.208 17:31:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:04.208 "name": "Existed_Raid", 00:10:04.208 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:04.208 "strip_size_kb": 64, 00:10:04.208 "state": "configuring", 00:10:04.208 "raid_level": "raid0", 00:10:04.208 "superblock": false, 00:10:04.208 "num_base_bdevs": 4, 00:10:04.208 "num_base_bdevs_discovered": 3, 00:10:04.208 "num_base_bdevs_operational": 4, 00:10:04.208 "base_bdevs_list": [ 00:10:04.208 { 00:10:04.208 "name": "BaseBdev1", 00:10:04.208 "uuid": "7ef8bd30-ab64-4d06-b669-d6d6a3e84206", 00:10:04.208 "is_configured": true, 00:10:04.208 "data_offset": 0, 00:10:04.208 "data_size": 65536 00:10:04.208 }, 00:10:04.208 { 00:10:04.208 "name": null, 00:10:04.208 "uuid": "7e99e0df-38d1-40ba-88a4-aa9fb968df07", 00:10:04.208 "is_configured": false, 00:10:04.208 "data_offset": 0, 00:10:04.208 "data_size": 65536 00:10:04.208 }, 00:10:04.208 { 00:10:04.208 "name": "BaseBdev3", 00:10:04.208 "uuid": "f3442a73-a294-498e-8526-e7f69d3b5043", 00:10:04.208 "is_configured": true, 00:10:04.208 "data_offset": 0, 00:10:04.208 "data_size": 65536 00:10:04.208 }, 00:10:04.208 { 00:10:04.208 "name": "BaseBdev4", 00:10:04.208 "uuid": "2801f3ec-cc2a-4df3-8ad8-d13266d87152", 00:10:04.208 "is_configured": true, 00:10:04.208 "data_offset": 0, 00:10:04.208 "data_size": 65536 00:10:04.208 } 00:10:04.208 ] 00:10:04.208 }' 00:10:04.208 17:31:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:04.208 17:31:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:04.468 17:31:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:10:04.468 17:31:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:04.468 17:31:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:04.468 17:31:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:04.468 17:31:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:04.468 17:31:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:10:04.468 17:31:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:10:04.468 17:31:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:04.468 17:31:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:04.468 [2024-11-27 17:31:35.618038] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:10:04.468 17:31:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:04.468 17:31:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:10:04.468 17:31:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:04.468 17:31:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:04.468 17:31:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:04.468 17:31:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:04.468 17:31:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:04.468 17:31:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:04.468 17:31:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:04.468 17:31:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:04.469 17:31:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:04.469 17:31:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:04.469 17:31:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:04.469 17:31:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:04.469 17:31:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:04.729 17:31:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:04.729 17:31:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:04.729 "name": "Existed_Raid", 00:10:04.729 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:04.729 "strip_size_kb": 64, 00:10:04.729 "state": "configuring", 00:10:04.729 "raid_level": "raid0", 00:10:04.729 "superblock": false, 00:10:04.729 "num_base_bdevs": 4, 00:10:04.729 "num_base_bdevs_discovered": 2, 00:10:04.729 "num_base_bdevs_operational": 4, 00:10:04.729 "base_bdevs_list": [ 00:10:04.729 { 00:10:04.729 "name": null, 00:10:04.729 "uuid": "7ef8bd30-ab64-4d06-b669-d6d6a3e84206", 00:10:04.729 "is_configured": false, 00:10:04.729 "data_offset": 0, 00:10:04.729 "data_size": 65536 00:10:04.729 }, 00:10:04.729 { 00:10:04.729 "name": null, 00:10:04.729 "uuid": "7e99e0df-38d1-40ba-88a4-aa9fb968df07", 00:10:04.729 "is_configured": false, 00:10:04.729 "data_offset": 0, 00:10:04.729 "data_size": 65536 00:10:04.729 }, 00:10:04.729 { 00:10:04.729 "name": "BaseBdev3", 00:10:04.729 "uuid": "f3442a73-a294-498e-8526-e7f69d3b5043", 00:10:04.729 "is_configured": true, 00:10:04.729 "data_offset": 0, 00:10:04.729 "data_size": 65536 00:10:04.729 }, 00:10:04.729 { 00:10:04.729 "name": "BaseBdev4", 00:10:04.729 "uuid": "2801f3ec-cc2a-4df3-8ad8-d13266d87152", 00:10:04.729 "is_configured": true, 00:10:04.729 "data_offset": 0, 00:10:04.729 "data_size": 65536 00:10:04.729 } 00:10:04.729 ] 00:10:04.729 }' 00:10:04.729 17:31:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:04.729 17:31:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:04.989 17:31:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:10:04.989 17:31:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:04.989 17:31:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:04.989 17:31:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:04.989 17:31:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:04.989 17:31:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:10:04.989 17:31:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:10:04.989 17:31:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:04.989 17:31:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:04.989 [2024-11-27 17:31:36.116985] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:04.989 17:31:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:04.989 17:31:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:10:04.989 17:31:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:04.989 17:31:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:04.989 17:31:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:04.989 17:31:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:04.989 17:31:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:04.989 17:31:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:04.989 17:31:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:04.989 17:31:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:04.989 17:31:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:04.989 17:31:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:04.989 17:31:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:04.989 17:31:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:04.989 17:31:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:04.989 17:31:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:04.989 17:31:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:04.989 "name": "Existed_Raid", 00:10:04.989 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:04.989 "strip_size_kb": 64, 00:10:04.989 "state": "configuring", 00:10:04.989 "raid_level": "raid0", 00:10:04.989 "superblock": false, 00:10:04.989 "num_base_bdevs": 4, 00:10:04.989 "num_base_bdevs_discovered": 3, 00:10:04.989 "num_base_bdevs_operational": 4, 00:10:04.989 "base_bdevs_list": [ 00:10:04.989 { 00:10:04.990 "name": null, 00:10:04.990 "uuid": "7ef8bd30-ab64-4d06-b669-d6d6a3e84206", 00:10:04.990 "is_configured": false, 00:10:04.990 "data_offset": 0, 00:10:04.990 "data_size": 65536 00:10:04.990 }, 00:10:04.990 { 00:10:04.990 "name": "BaseBdev2", 00:10:04.990 "uuid": "7e99e0df-38d1-40ba-88a4-aa9fb968df07", 00:10:04.990 "is_configured": true, 00:10:04.990 "data_offset": 0, 00:10:04.990 "data_size": 65536 00:10:04.990 }, 00:10:04.990 { 00:10:04.990 "name": "BaseBdev3", 00:10:04.990 "uuid": "f3442a73-a294-498e-8526-e7f69d3b5043", 00:10:04.990 "is_configured": true, 00:10:04.990 "data_offset": 0, 00:10:04.990 "data_size": 65536 00:10:04.990 }, 00:10:04.990 { 00:10:04.990 "name": "BaseBdev4", 00:10:04.990 "uuid": "2801f3ec-cc2a-4df3-8ad8-d13266d87152", 00:10:04.990 "is_configured": true, 00:10:04.990 "data_offset": 0, 00:10:04.990 "data_size": 65536 00:10:04.990 } 00:10:04.990 ] 00:10:04.990 }' 00:10:04.990 17:31:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:04.990 17:31:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:05.559 17:31:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:10:05.559 17:31:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:05.559 17:31:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:05.559 17:31:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:05.559 17:31:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:05.559 17:31:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:10:05.559 17:31:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:10:05.559 17:31:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:05.559 17:31:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:05.559 17:31:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:05.559 17:31:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:05.559 17:31:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 7ef8bd30-ab64-4d06-b669-d6d6a3e84206 00:10:05.559 17:31:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:05.559 17:31:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:05.559 [2024-11-27 17:31:36.664774] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:10:05.559 [2024-11-27 17:31:36.664876] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:10:05.559 [2024-11-27 17:31:36.664901] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 262144, blocklen 512 00:10:05.559 [2024-11-27 17:31:36.665244] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002a10 00:10:05.559 [2024-11-27 17:31:36.665410] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:10:05.559 [2024-11-27 17:31:36.665450] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001c80 00:10:05.559 [2024-11-27 17:31:36.665675] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:05.559 NewBaseBdev 00:10:05.559 17:31:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:05.559 17:31:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:10:05.559 17:31:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=NewBaseBdev 00:10:05.559 17:31:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:10:05.559 17:31:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:10:05.559 17:31:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:10:05.559 17:31:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:10:05.559 17:31:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:10:05.559 17:31:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:05.559 17:31:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:05.559 17:31:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:05.559 17:31:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:10:05.559 17:31:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:05.559 17:31:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:05.559 [ 00:10:05.559 { 00:10:05.559 "name": "NewBaseBdev", 00:10:05.559 "aliases": [ 00:10:05.559 "7ef8bd30-ab64-4d06-b669-d6d6a3e84206" 00:10:05.559 ], 00:10:05.559 "product_name": "Malloc disk", 00:10:05.559 "block_size": 512, 00:10:05.559 "num_blocks": 65536, 00:10:05.559 "uuid": "7ef8bd30-ab64-4d06-b669-d6d6a3e84206", 00:10:05.559 "assigned_rate_limits": { 00:10:05.559 "rw_ios_per_sec": 0, 00:10:05.559 "rw_mbytes_per_sec": 0, 00:10:05.559 "r_mbytes_per_sec": 0, 00:10:05.559 "w_mbytes_per_sec": 0 00:10:05.559 }, 00:10:05.559 "claimed": true, 00:10:05.559 "claim_type": "exclusive_write", 00:10:05.559 "zoned": false, 00:10:05.559 "supported_io_types": { 00:10:05.559 "read": true, 00:10:05.559 "write": true, 00:10:05.559 "unmap": true, 00:10:05.559 "flush": true, 00:10:05.559 "reset": true, 00:10:05.559 "nvme_admin": false, 00:10:05.559 "nvme_io": false, 00:10:05.559 "nvme_io_md": false, 00:10:05.559 "write_zeroes": true, 00:10:05.559 "zcopy": true, 00:10:05.559 "get_zone_info": false, 00:10:05.559 "zone_management": false, 00:10:05.559 "zone_append": false, 00:10:05.559 "compare": false, 00:10:05.559 "compare_and_write": false, 00:10:05.559 "abort": true, 00:10:05.559 "seek_hole": false, 00:10:05.559 "seek_data": false, 00:10:05.559 "copy": true, 00:10:05.559 "nvme_iov_md": false 00:10:05.559 }, 00:10:05.559 "memory_domains": [ 00:10:05.559 { 00:10:05.559 "dma_device_id": "system", 00:10:05.559 "dma_device_type": 1 00:10:05.559 }, 00:10:05.559 { 00:10:05.559 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:05.559 "dma_device_type": 2 00:10:05.559 } 00:10:05.559 ], 00:10:05.559 "driver_specific": {} 00:10:05.559 } 00:10:05.559 ] 00:10:05.559 17:31:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:05.559 17:31:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:10:05.559 17:31:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid0 64 4 00:10:05.559 17:31:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:05.559 17:31:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:05.559 17:31:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:05.559 17:31:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:05.559 17:31:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:05.559 17:31:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:05.559 17:31:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:05.559 17:31:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:05.559 17:31:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:05.559 17:31:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:05.559 17:31:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:05.559 17:31:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:05.559 17:31:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:05.559 17:31:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:05.819 17:31:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:05.819 "name": "Existed_Raid", 00:10:05.819 "uuid": "35fe90c5-a454-4183-9382-a5ada0c19a96", 00:10:05.819 "strip_size_kb": 64, 00:10:05.819 "state": "online", 00:10:05.819 "raid_level": "raid0", 00:10:05.819 "superblock": false, 00:10:05.819 "num_base_bdevs": 4, 00:10:05.819 "num_base_bdevs_discovered": 4, 00:10:05.819 "num_base_bdevs_operational": 4, 00:10:05.819 "base_bdevs_list": [ 00:10:05.819 { 00:10:05.819 "name": "NewBaseBdev", 00:10:05.819 "uuid": "7ef8bd30-ab64-4d06-b669-d6d6a3e84206", 00:10:05.819 "is_configured": true, 00:10:05.819 "data_offset": 0, 00:10:05.819 "data_size": 65536 00:10:05.819 }, 00:10:05.819 { 00:10:05.819 "name": "BaseBdev2", 00:10:05.819 "uuid": "7e99e0df-38d1-40ba-88a4-aa9fb968df07", 00:10:05.819 "is_configured": true, 00:10:05.819 "data_offset": 0, 00:10:05.819 "data_size": 65536 00:10:05.819 }, 00:10:05.819 { 00:10:05.819 "name": "BaseBdev3", 00:10:05.819 "uuid": "f3442a73-a294-498e-8526-e7f69d3b5043", 00:10:05.819 "is_configured": true, 00:10:05.819 "data_offset": 0, 00:10:05.819 "data_size": 65536 00:10:05.819 }, 00:10:05.819 { 00:10:05.819 "name": "BaseBdev4", 00:10:05.819 "uuid": "2801f3ec-cc2a-4df3-8ad8-d13266d87152", 00:10:05.819 "is_configured": true, 00:10:05.819 "data_offset": 0, 00:10:05.819 "data_size": 65536 00:10:05.819 } 00:10:05.819 ] 00:10:05.819 }' 00:10:05.819 17:31:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:05.819 17:31:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:06.079 17:31:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:10:06.079 17:31:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:10:06.079 17:31:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:10:06.079 17:31:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:10:06.079 17:31:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:10:06.079 17:31:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:10:06.079 17:31:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:10:06.079 17:31:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:06.079 17:31:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:06.079 17:31:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:10:06.079 [2024-11-27 17:31:37.080411] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:06.079 17:31:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:06.079 17:31:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:06.079 "name": "Existed_Raid", 00:10:06.079 "aliases": [ 00:10:06.079 "35fe90c5-a454-4183-9382-a5ada0c19a96" 00:10:06.079 ], 00:10:06.079 "product_name": "Raid Volume", 00:10:06.079 "block_size": 512, 00:10:06.079 "num_blocks": 262144, 00:10:06.079 "uuid": "35fe90c5-a454-4183-9382-a5ada0c19a96", 00:10:06.079 "assigned_rate_limits": { 00:10:06.079 "rw_ios_per_sec": 0, 00:10:06.079 "rw_mbytes_per_sec": 0, 00:10:06.079 "r_mbytes_per_sec": 0, 00:10:06.079 "w_mbytes_per_sec": 0 00:10:06.079 }, 00:10:06.079 "claimed": false, 00:10:06.079 "zoned": false, 00:10:06.079 "supported_io_types": { 00:10:06.079 "read": true, 00:10:06.079 "write": true, 00:10:06.079 "unmap": true, 00:10:06.079 "flush": true, 00:10:06.079 "reset": true, 00:10:06.079 "nvme_admin": false, 00:10:06.079 "nvme_io": false, 00:10:06.079 "nvme_io_md": false, 00:10:06.079 "write_zeroes": true, 00:10:06.079 "zcopy": false, 00:10:06.079 "get_zone_info": false, 00:10:06.079 "zone_management": false, 00:10:06.079 "zone_append": false, 00:10:06.079 "compare": false, 00:10:06.079 "compare_and_write": false, 00:10:06.079 "abort": false, 00:10:06.079 "seek_hole": false, 00:10:06.080 "seek_data": false, 00:10:06.080 "copy": false, 00:10:06.080 "nvme_iov_md": false 00:10:06.080 }, 00:10:06.080 "memory_domains": [ 00:10:06.080 { 00:10:06.080 "dma_device_id": "system", 00:10:06.080 "dma_device_type": 1 00:10:06.080 }, 00:10:06.080 { 00:10:06.080 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:06.080 "dma_device_type": 2 00:10:06.080 }, 00:10:06.080 { 00:10:06.080 "dma_device_id": "system", 00:10:06.080 "dma_device_type": 1 00:10:06.080 }, 00:10:06.080 { 00:10:06.080 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:06.080 "dma_device_type": 2 00:10:06.080 }, 00:10:06.080 { 00:10:06.080 "dma_device_id": "system", 00:10:06.080 "dma_device_type": 1 00:10:06.080 }, 00:10:06.080 { 00:10:06.080 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:06.080 "dma_device_type": 2 00:10:06.080 }, 00:10:06.080 { 00:10:06.080 "dma_device_id": "system", 00:10:06.080 "dma_device_type": 1 00:10:06.080 }, 00:10:06.080 { 00:10:06.080 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:06.080 "dma_device_type": 2 00:10:06.080 } 00:10:06.080 ], 00:10:06.080 "driver_specific": { 00:10:06.080 "raid": { 00:10:06.080 "uuid": "35fe90c5-a454-4183-9382-a5ada0c19a96", 00:10:06.080 "strip_size_kb": 64, 00:10:06.080 "state": "online", 00:10:06.080 "raid_level": "raid0", 00:10:06.080 "superblock": false, 00:10:06.080 "num_base_bdevs": 4, 00:10:06.080 "num_base_bdevs_discovered": 4, 00:10:06.080 "num_base_bdevs_operational": 4, 00:10:06.080 "base_bdevs_list": [ 00:10:06.080 { 00:10:06.080 "name": "NewBaseBdev", 00:10:06.080 "uuid": "7ef8bd30-ab64-4d06-b669-d6d6a3e84206", 00:10:06.080 "is_configured": true, 00:10:06.080 "data_offset": 0, 00:10:06.080 "data_size": 65536 00:10:06.080 }, 00:10:06.080 { 00:10:06.080 "name": "BaseBdev2", 00:10:06.080 "uuid": "7e99e0df-38d1-40ba-88a4-aa9fb968df07", 00:10:06.080 "is_configured": true, 00:10:06.080 "data_offset": 0, 00:10:06.080 "data_size": 65536 00:10:06.080 }, 00:10:06.080 { 00:10:06.080 "name": "BaseBdev3", 00:10:06.080 "uuid": "f3442a73-a294-498e-8526-e7f69d3b5043", 00:10:06.080 "is_configured": true, 00:10:06.080 "data_offset": 0, 00:10:06.080 "data_size": 65536 00:10:06.080 }, 00:10:06.080 { 00:10:06.080 "name": "BaseBdev4", 00:10:06.080 "uuid": "2801f3ec-cc2a-4df3-8ad8-d13266d87152", 00:10:06.080 "is_configured": true, 00:10:06.080 "data_offset": 0, 00:10:06.080 "data_size": 65536 00:10:06.080 } 00:10:06.080 ] 00:10:06.080 } 00:10:06.080 } 00:10:06.080 }' 00:10:06.080 17:31:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:06.080 17:31:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:10:06.080 BaseBdev2 00:10:06.080 BaseBdev3 00:10:06.080 BaseBdev4' 00:10:06.080 17:31:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:06.080 17:31:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:06.080 17:31:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:06.080 17:31:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:06.080 17:31:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:10:06.080 17:31:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:06.080 17:31:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:06.080 17:31:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:06.080 17:31:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:06.080 17:31:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:06.080 17:31:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:06.080 17:31:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:10:06.080 17:31:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:06.080 17:31:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:06.080 17:31:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:06.340 17:31:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:06.340 17:31:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:06.340 17:31:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:06.340 17:31:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:06.340 17:31:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:10:06.341 17:31:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:06.341 17:31:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:06.341 17:31:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:06.341 17:31:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:06.341 17:31:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:06.341 17:31:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:06.341 17:31:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:06.341 17:31:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:10:06.341 17:31:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:06.341 17:31:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:06.341 17:31:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:06.341 17:31:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:06.341 17:31:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:06.341 17:31:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:06.341 17:31:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:10:06.341 17:31:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:06.341 17:31:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:06.341 [2024-11-27 17:31:37.411502] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:10:06.341 [2024-11-27 17:31:37.411530] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:06.341 [2024-11-27 17:31:37.411614] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:06.341 [2024-11-27 17:31:37.411683] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:06.341 [2024-11-27 17:31:37.411692] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name Existed_Raid, state offline 00:10:06.341 17:31:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:06.341 17:31:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 80201 00:10:06.341 17:31:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@950 -- # '[' -z 80201 ']' 00:10:06.341 17:31:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # kill -0 80201 00:10:06.341 17:31:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@955 -- # uname 00:10:06.341 17:31:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:10:06.341 17:31:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 80201 00:10:06.341 17:31:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:10:06.341 17:31:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:10:06.341 17:31:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 80201' 00:10:06.341 killing process with pid 80201 00:10:06.341 17:31:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@969 -- # kill 80201 00:10:06.341 [2024-11-27 17:31:37.455626] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:10:06.341 17:31:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@974 -- # wait 80201 00:10:06.600 [2024-11-27 17:31:37.532616] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:10:06.860 17:31:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:10:06.860 00:10:06.860 real 0m9.729s 00:10:06.860 user 0m16.273s 00:10:06.860 sys 0m2.100s 00:10:06.860 17:31:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:10:06.860 17:31:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:06.860 ************************************ 00:10:06.860 END TEST raid_state_function_test 00:10:06.860 ************************************ 00:10:06.860 17:31:37 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test raid0 4 true 00:10:06.860 17:31:37 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:10:06.860 17:31:37 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:10:06.860 17:31:37 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:10:06.860 ************************************ 00:10:06.860 START TEST raid_state_function_test_sb 00:10:06.860 ************************************ 00:10:06.860 17:31:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1125 -- # raid_state_function_test raid0 4 true 00:10:06.860 17:31:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=raid0 00:10:06.860 17:31:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=4 00:10:06.860 17:31:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:10:06.860 17:31:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:10:06.860 17:31:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:10:06.860 17:31:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:06.860 17:31:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:10:06.860 17:31:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:06.860 17:31:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:06.860 17:31:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:10:06.860 17:31:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:06.860 17:31:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:06.860 17:31:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:10:06.860 17:31:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:06.860 17:31:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:06.860 17:31:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev4 00:10:06.860 17:31:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:06.860 17:31:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:06.860 17:31:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:10:06.860 17:31:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:10:06.860 17:31:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:10:06.860 17:31:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:10:06.860 17:31:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:10:06.860 17:31:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:10:06.860 17:31:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' raid0 '!=' raid1 ']' 00:10:06.860 17:31:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:10:06.860 17:31:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:10:06.860 17:31:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:10:06.860 17:31:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:10:06.860 Process raid pid: 80850 00:10:06.860 17:31:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=80850 00:10:06.860 17:31:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:10:06.860 17:31:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 80850' 00:10:06.860 17:31:37 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 80850 00:10:06.860 17:31:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@831 -- # '[' -z 80850 ']' 00:10:06.860 17:31:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:06.860 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:06.861 17:31:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@836 -- # local max_retries=100 00:10:06.861 17:31:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:06.861 17:31:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # xtrace_disable 00:10:06.861 17:31:37 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:07.120 [2024-11-27 17:31:38.074488] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:10:07.120 [2024-11-27 17:31:38.074710] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:10:07.120 [2024-11-27 17:31:38.218424] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:07.120 [2024-11-27 17:31:38.288829] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:10:07.379 [2024-11-27 17:31:38.364629] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:07.379 [2024-11-27 17:31:38.364673] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:07.949 17:31:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:10:07.949 17:31:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@864 -- # return 0 00:10:07.949 17:31:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:10:07.949 17:31:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:07.949 17:31:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:07.949 [2024-11-27 17:31:38.907660] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:10:07.949 [2024-11-27 17:31:38.907727] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:10:07.949 [2024-11-27 17:31:38.907747] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:10:07.949 [2024-11-27 17:31:38.907759] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:10:07.949 [2024-11-27 17:31:38.907766] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:10:07.949 [2024-11-27 17:31:38.907778] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:10:07.949 [2024-11-27 17:31:38.907783] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:10:07.949 [2024-11-27 17:31:38.907792] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:10:07.949 17:31:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:07.949 17:31:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:10:07.949 17:31:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:07.949 17:31:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:07.949 17:31:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:07.949 17:31:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:07.950 17:31:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:07.950 17:31:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:07.950 17:31:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:07.950 17:31:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:07.950 17:31:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:07.950 17:31:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:07.950 17:31:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:07.950 17:31:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:07.950 17:31:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:07.950 17:31:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:07.950 17:31:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:07.950 "name": "Existed_Raid", 00:10:07.950 "uuid": "c348eb7f-640a-4b86-9432-83ca4c23723c", 00:10:07.950 "strip_size_kb": 64, 00:10:07.950 "state": "configuring", 00:10:07.950 "raid_level": "raid0", 00:10:07.950 "superblock": true, 00:10:07.950 "num_base_bdevs": 4, 00:10:07.950 "num_base_bdevs_discovered": 0, 00:10:07.950 "num_base_bdevs_operational": 4, 00:10:07.950 "base_bdevs_list": [ 00:10:07.950 { 00:10:07.950 "name": "BaseBdev1", 00:10:07.950 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:07.950 "is_configured": false, 00:10:07.950 "data_offset": 0, 00:10:07.950 "data_size": 0 00:10:07.950 }, 00:10:07.950 { 00:10:07.950 "name": "BaseBdev2", 00:10:07.950 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:07.950 "is_configured": false, 00:10:07.950 "data_offset": 0, 00:10:07.950 "data_size": 0 00:10:07.950 }, 00:10:07.950 { 00:10:07.950 "name": "BaseBdev3", 00:10:07.950 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:07.950 "is_configured": false, 00:10:07.950 "data_offset": 0, 00:10:07.950 "data_size": 0 00:10:07.950 }, 00:10:07.950 { 00:10:07.950 "name": "BaseBdev4", 00:10:07.950 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:07.950 "is_configured": false, 00:10:07.950 "data_offset": 0, 00:10:07.950 "data_size": 0 00:10:07.950 } 00:10:07.950 ] 00:10:07.950 }' 00:10:07.950 17:31:38 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:07.950 17:31:38 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:08.215 17:31:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:10:08.216 17:31:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:08.216 17:31:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:08.216 [2024-11-27 17:31:39.370744] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:10:08.216 [2024-11-27 17:31:39.370848] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:10:08.216 17:31:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:08.216 17:31:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:10:08.216 17:31:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:08.216 17:31:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:08.216 [2024-11-27 17:31:39.382748] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:10:08.216 [2024-11-27 17:31:39.382829] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:10:08.216 [2024-11-27 17:31:39.382874] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:10:08.216 [2024-11-27 17:31:39.382898] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:10:08.216 [2024-11-27 17:31:39.382918] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:10:08.216 [2024-11-27 17:31:39.382948] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:10:08.216 [2024-11-27 17:31:39.382968] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:10:08.216 [2024-11-27 17:31:39.382990] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:10:08.216 17:31:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:08.216 17:31:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:10:08.216 17:31:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:08.216 17:31:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:08.479 [2024-11-27 17:31:39.410054] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:08.479 BaseBdev1 00:10:08.479 17:31:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:08.479 17:31:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:10:08.479 17:31:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:10:08.479 17:31:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:10:08.479 17:31:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:10:08.479 17:31:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:10:08.479 17:31:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:10:08.479 17:31:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:10:08.479 17:31:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:08.479 17:31:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:08.479 17:31:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:08.479 17:31:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:10:08.479 17:31:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:08.479 17:31:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:08.479 [ 00:10:08.479 { 00:10:08.479 "name": "BaseBdev1", 00:10:08.479 "aliases": [ 00:10:08.479 "e8d1c6db-0421-4402-82f2-5d530d912be1" 00:10:08.479 ], 00:10:08.479 "product_name": "Malloc disk", 00:10:08.479 "block_size": 512, 00:10:08.479 "num_blocks": 65536, 00:10:08.479 "uuid": "e8d1c6db-0421-4402-82f2-5d530d912be1", 00:10:08.479 "assigned_rate_limits": { 00:10:08.479 "rw_ios_per_sec": 0, 00:10:08.479 "rw_mbytes_per_sec": 0, 00:10:08.479 "r_mbytes_per_sec": 0, 00:10:08.479 "w_mbytes_per_sec": 0 00:10:08.479 }, 00:10:08.479 "claimed": true, 00:10:08.479 "claim_type": "exclusive_write", 00:10:08.479 "zoned": false, 00:10:08.479 "supported_io_types": { 00:10:08.479 "read": true, 00:10:08.479 "write": true, 00:10:08.479 "unmap": true, 00:10:08.479 "flush": true, 00:10:08.479 "reset": true, 00:10:08.479 "nvme_admin": false, 00:10:08.479 "nvme_io": false, 00:10:08.479 "nvme_io_md": false, 00:10:08.479 "write_zeroes": true, 00:10:08.479 "zcopy": true, 00:10:08.479 "get_zone_info": false, 00:10:08.479 "zone_management": false, 00:10:08.479 "zone_append": false, 00:10:08.479 "compare": false, 00:10:08.479 "compare_and_write": false, 00:10:08.479 "abort": true, 00:10:08.479 "seek_hole": false, 00:10:08.479 "seek_data": false, 00:10:08.479 "copy": true, 00:10:08.479 "nvme_iov_md": false 00:10:08.479 }, 00:10:08.479 "memory_domains": [ 00:10:08.479 { 00:10:08.479 "dma_device_id": "system", 00:10:08.479 "dma_device_type": 1 00:10:08.479 }, 00:10:08.480 { 00:10:08.480 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:08.480 "dma_device_type": 2 00:10:08.480 } 00:10:08.480 ], 00:10:08.480 "driver_specific": {} 00:10:08.480 } 00:10:08.480 ] 00:10:08.480 17:31:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:08.480 17:31:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:10:08.480 17:31:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:10:08.480 17:31:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:08.480 17:31:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:08.480 17:31:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:08.480 17:31:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:08.480 17:31:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:08.480 17:31:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:08.480 17:31:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:08.480 17:31:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:08.480 17:31:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:08.480 17:31:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:08.480 17:31:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:08.480 17:31:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:08.480 17:31:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:08.480 17:31:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:08.480 17:31:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:08.480 "name": "Existed_Raid", 00:10:08.480 "uuid": "f44df467-33c3-46a3-8c2c-e5c421e9203e", 00:10:08.480 "strip_size_kb": 64, 00:10:08.480 "state": "configuring", 00:10:08.480 "raid_level": "raid0", 00:10:08.480 "superblock": true, 00:10:08.480 "num_base_bdevs": 4, 00:10:08.480 "num_base_bdevs_discovered": 1, 00:10:08.480 "num_base_bdevs_operational": 4, 00:10:08.480 "base_bdevs_list": [ 00:10:08.480 { 00:10:08.480 "name": "BaseBdev1", 00:10:08.480 "uuid": "e8d1c6db-0421-4402-82f2-5d530d912be1", 00:10:08.480 "is_configured": true, 00:10:08.480 "data_offset": 2048, 00:10:08.480 "data_size": 63488 00:10:08.480 }, 00:10:08.480 { 00:10:08.480 "name": "BaseBdev2", 00:10:08.480 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:08.480 "is_configured": false, 00:10:08.480 "data_offset": 0, 00:10:08.480 "data_size": 0 00:10:08.480 }, 00:10:08.480 { 00:10:08.480 "name": "BaseBdev3", 00:10:08.480 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:08.480 "is_configured": false, 00:10:08.480 "data_offset": 0, 00:10:08.480 "data_size": 0 00:10:08.480 }, 00:10:08.480 { 00:10:08.480 "name": "BaseBdev4", 00:10:08.480 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:08.480 "is_configured": false, 00:10:08.480 "data_offset": 0, 00:10:08.480 "data_size": 0 00:10:08.480 } 00:10:08.480 ] 00:10:08.480 }' 00:10:08.480 17:31:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:08.480 17:31:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:08.740 17:31:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:10:08.740 17:31:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:08.740 17:31:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:08.740 [2024-11-27 17:31:39.889242] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:10:08.740 [2024-11-27 17:31:39.889340] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:10:08.741 17:31:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:08.741 17:31:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:10:08.741 17:31:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:08.741 17:31:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:08.741 [2024-11-27 17:31:39.901284] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:08.741 [2024-11-27 17:31:39.903428] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:10:08.741 [2024-11-27 17:31:39.903510] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:10:08.741 [2024-11-27 17:31:39.903525] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:10:08.741 [2024-11-27 17:31:39.903534] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:10:08.741 [2024-11-27 17:31:39.903540] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:10:08.741 [2024-11-27 17:31:39.903549] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:10:08.741 17:31:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:08.741 17:31:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:10:08.741 17:31:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:08.741 17:31:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:10:08.741 17:31:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:08.741 17:31:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:08.741 17:31:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:08.741 17:31:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:08.741 17:31:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:08.741 17:31:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:08.741 17:31:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:08.741 17:31:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:08.741 17:31:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:08.741 17:31:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:08.741 17:31:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:08.741 17:31:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:08.741 17:31:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:08.741 17:31:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:09.000 17:31:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:09.000 "name": "Existed_Raid", 00:10:09.000 "uuid": "28abdeee-fd43-4400-ad0d-a50dbb2ce588", 00:10:09.000 "strip_size_kb": 64, 00:10:09.000 "state": "configuring", 00:10:09.000 "raid_level": "raid0", 00:10:09.000 "superblock": true, 00:10:09.000 "num_base_bdevs": 4, 00:10:09.000 "num_base_bdevs_discovered": 1, 00:10:09.000 "num_base_bdevs_operational": 4, 00:10:09.000 "base_bdevs_list": [ 00:10:09.000 { 00:10:09.000 "name": "BaseBdev1", 00:10:09.000 "uuid": "e8d1c6db-0421-4402-82f2-5d530d912be1", 00:10:09.000 "is_configured": true, 00:10:09.000 "data_offset": 2048, 00:10:09.000 "data_size": 63488 00:10:09.000 }, 00:10:09.000 { 00:10:09.000 "name": "BaseBdev2", 00:10:09.000 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:09.000 "is_configured": false, 00:10:09.000 "data_offset": 0, 00:10:09.001 "data_size": 0 00:10:09.001 }, 00:10:09.001 { 00:10:09.001 "name": "BaseBdev3", 00:10:09.001 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:09.001 "is_configured": false, 00:10:09.001 "data_offset": 0, 00:10:09.001 "data_size": 0 00:10:09.001 }, 00:10:09.001 { 00:10:09.001 "name": "BaseBdev4", 00:10:09.001 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:09.001 "is_configured": false, 00:10:09.001 "data_offset": 0, 00:10:09.001 "data_size": 0 00:10:09.001 } 00:10:09.001 ] 00:10:09.001 }' 00:10:09.001 17:31:39 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:09.001 17:31:39 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:09.260 17:31:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:10:09.260 17:31:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:09.260 17:31:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:09.260 [2024-11-27 17:31:40.371859] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:09.260 BaseBdev2 00:10:09.260 17:31:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:09.260 17:31:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:10:09.260 17:31:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:10:09.260 17:31:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:10:09.260 17:31:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:10:09.260 17:31:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:10:09.260 17:31:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:10:09.260 17:31:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:10:09.260 17:31:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:09.260 17:31:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:09.260 17:31:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:09.260 17:31:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:10:09.260 17:31:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:09.260 17:31:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:09.260 [ 00:10:09.260 { 00:10:09.260 "name": "BaseBdev2", 00:10:09.260 "aliases": [ 00:10:09.260 "4bada4e1-68cf-426f-8186-6d64e2747b15" 00:10:09.260 ], 00:10:09.261 "product_name": "Malloc disk", 00:10:09.261 "block_size": 512, 00:10:09.261 "num_blocks": 65536, 00:10:09.261 "uuid": "4bada4e1-68cf-426f-8186-6d64e2747b15", 00:10:09.261 "assigned_rate_limits": { 00:10:09.261 "rw_ios_per_sec": 0, 00:10:09.261 "rw_mbytes_per_sec": 0, 00:10:09.261 "r_mbytes_per_sec": 0, 00:10:09.261 "w_mbytes_per_sec": 0 00:10:09.261 }, 00:10:09.261 "claimed": true, 00:10:09.261 "claim_type": "exclusive_write", 00:10:09.261 "zoned": false, 00:10:09.261 "supported_io_types": { 00:10:09.261 "read": true, 00:10:09.261 "write": true, 00:10:09.261 "unmap": true, 00:10:09.261 "flush": true, 00:10:09.261 "reset": true, 00:10:09.261 "nvme_admin": false, 00:10:09.261 "nvme_io": false, 00:10:09.261 "nvme_io_md": false, 00:10:09.261 "write_zeroes": true, 00:10:09.261 "zcopy": true, 00:10:09.261 "get_zone_info": false, 00:10:09.261 "zone_management": false, 00:10:09.261 "zone_append": false, 00:10:09.261 "compare": false, 00:10:09.261 "compare_and_write": false, 00:10:09.261 "abort": true, 00:10:09.261 "seek_hole": false, 00:10:09.261 "seek_data": false, 00:10:09.261 "copy": true, 00:10:09.261 "nvme_iov_md": false 00:10:09.261 }, 00:10:09.261 "memory_domains": [ 00:10:09.261 { 00:10:09.261 "dma_device_id": "system", 00:10:09.261 "dma_device_type": 1 00:10:09.261 }, 00:10:09.261 { 00:10:09.261 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:09.261 "dma_device_type": 2 00:10:09.261 } 00:10:09.261 ], 00:10:09.261 "driver_specific": {} 00:10:09.261 } 00:10:09.261 ] 00:10:09.261 17:31:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:09.261 17:31:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:10:09.261 17:31:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:10:09.261 17:31:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:09.261 17:31:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:10:09.261 17:31:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:09.261 17:31:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:09.261 17:31:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:09.261 17:31:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:09.261 17:31:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:09.261 17:31:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:09.261 17:31:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:09.261 17:31:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:09.261 17:31:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:09.261 17:31:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:09.261 17:31:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:09.261 17:31:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:09.261 17:31:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:09.261 17:31:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:09.521 17:31:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:09.521 "name": "Existed_Raid", 00:10:09.521 "uuid": "28abdeee-fd43-4400-ad0d-a50dbb2ce588", 00:10:09.521 "strip_size_kb": 64, 00:10:09.521 "state": "configuring", 00:10:09.521 "raid_level": "raid0", 00:10:09.521 "superblock": true, 00:10:09.521 "num_base_bdevs": 4, 00:10:09.521 "num_base_bdevs_discovered": 2, 00:10:09.521 "num_base_bdevs_operational": 4, 00:10:09.521 "base_bdevs_list": [ 00:10:09.521 { 00:10:09.521 "name": "BaseBdev1", 00:10:09.521 "uuid": "e8d1c6db-0421-4402-82f2-5d530d912be1", 00:10:09.521 "is_configured": true, 00:10:09.521 "data_offset": 2048, 00:10:09.521 "data_size": 63488 00:10:09.521 }, 00:10:09.521 { 00:10:09.521 "name": "BaseBdev2", 00:10:09.521 "uuid": "4bada4e1-68cf-426f-8186-6d64e2747b15", 00:10:09.521 "is_configured": true, 00:10:09.521 "data_offset": 2048, 00:10:09.521 "data_size": 63488 00:10:09.521 }, 00:10:09.521 { 00:10:09.521 "name": "BaseBdev3", 00:10:09.521 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:09.521 "is_configured": false, 00:10:09.521 "data_offset": 0, 00:10:09.521 "data_size": 0 00:10:09.521 }, 00:10:09.521 { 00:10:09.521 "name": "BaseBdev4", 00:10:09.521 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:09.521 "is_configured": false, 00:10:09.521 "data_offset": 0, 00:10:09.521 "data_size": 0 00:10:09.521 } 00:10:09.521 ] 00:10:09.521 }' 00:10:09.521 17:31:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:09.521 17:31:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:09.781 17:31:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:10:09.781 17:31:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:09.781 17:31:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:09.781 [2024-11-27 17:31:40.883647] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:09.781 BaseBdev3 00:10:09.781 17:31:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:09.781 17:31:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:10:09.781 17:31:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:10:09.781 17:31:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:10:09.781 17:31:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:10:09.781 17:31:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:10:09.781 17:31:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:10:09.781 17:31:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:10:09.781 17:31:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:09.781 17:31:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:09.781 17:31:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:09.781 17:31:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:10:09.781 17:31:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:09.781 17:31:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:09.781 [ 00:10:09.781 { 00:10:09.781 "name": "BaseBdev3", 00:10:09.781 "aliases": [ 00:10:09.781 "a9515ebc-4169-4d6a-a46a-0ba1873f7a96" 00:10:09.781 ], 00:10:09.781 "product_name": "Malloc disk", 00:10:09.781 "block_size": 512, 00:10:09.781 "num_blocks": 65536, 00:10:09.781 "uuid": "a9515ebc-4169-4d6a-a46a-0ba1873f7a96", 00:10:09.781 "assigned_rate_limits": { 00:10:09.781 "rw_ios_per_sec": 0, 00:10:09.781 "rw_mbytes_per_sec": 0, 00:10:09.781 "r_mbytes_per_sec": 0, 00:10:09.781 "w_mbytes_per_sec": 0 00:10:09.781 }, 00:10:09.781 "claimed": true, 00:10:09.781 "claim_type": "exclusive_write", 00:10:09.781 "zoned": false, 00:10:09.781 "supported_io_types": { 00:10:09.781 "read": true, 00:10:09.781 "write": true, 00:10:09.781 "unmap": true, 00:10:09.781 "flush": true, 00:10:09.781 "reset": true, 00:10:09.781 "nvme_admin": false, 00:10:09.781 "nvme_io": false, 00:10:09.781 "nvme_io_md": false, 00:10:09.781 "write_zeroes": true, 00:10:09.781 "zcopy": true, 00:10:09.781 "get_zone_info": false, 00:10:09.781 "zone_management": false, 00:10:09.781 "zone_append": false, 00:10:09.781 "compare": false, 00:10:09.781 "compare_and_write": false, 00:10:09.781 "abort": true, 00:10:09.781 "seek_hole": false, 00:10:09.781 "seek_data": false, 00:10:09.781 "copy": true, 00:10:09.781 "nvme_iov_md": false 00:10:09.781 }, 00:10:09.781 "memory_domains": [ 00:10:09.781 { 00:10:09.781 "dma_device_id": "system", 00:10:09.781 "dma_device_type": 1 00:10:09.781 }, 00:10:09.781 { 00:10:09.781 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:09.781 "dma_device_type": 2 00:10:09.781 } 00:10:09.781 ], 00:10:09.781 "driver_specific": {} 00:10:09.781 } 00:10:09.781 ] 00:10:09.781 17:31:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:09.781 17:31:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:10:09.781 17:31:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:10:09.781 17:31:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:09.781 17:31:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:10:09.781 17:31:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:09.781 17:31:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:09.781 17:31:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:09.781 17:31:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:09.781 17:31:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:09.781 17:31:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:09.781 17:31:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:09.781 17:31:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:09.781 17:31:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:09.781 17:31:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:09.781 17:31:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:09.781 17:31:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:09.781 17:31:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:09.781 17:31:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:10.041 17:31:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:10.041 "name": "Existed_Raid", 00:10:10.041 "uuid": "28abdeee-fd43-4400-ad0d-a50dbb2ce588", 00:10:10.041 "strip_size_kb": 64, 00:10:10.041 "state": "configuring", 00:10:10.041 "raid_level": "raid0", 00:10:10.041 "superblock": true, 00:10:10.041 "num_base_bdevs": 4, 00:10:10.041 "num_base_bdevs_discovered": 3, 00:10:10.041 "num_base_bdevs_operational": 4, 00:10:10.041 "base_bdevs_list": [ 00:10:10.041 { 00:10:10.041 "name": "BaseBdev1", 00:10:10.041 "uuid": "e8d1c6db-0421-4402-82f2-5d530d912be1", 00:10:10.041 "is_configured": true, 00:10:10.041 "data_offset": 2048, 00:10:10.041 "data_size": 63488 00:10:10.041 }, 00:10:10.041 { 00:10:10.041 "name": "BaseBdev2", 00:10:10.041 "uuid": "4bada4e1-68cf-426f-8186-6d64e2747b15", 00:10:10.041 "is_configured": true, 00:10:10.041 "data_offset": 2048, 00:10:10.041 "data_size": 63488 00:10:10.041 }, 00:10:10.041 { 00:10:10.041 "name": "BaseBdev3", 00:10:10.041 "uuid": "a9515ebc-4169-4d6a-a46a-0ba1873f7a96", 00:10:10.041 "is_configured": true, 00:10:10.041 "data_offset": 2048, 00:10:10.041 "data_size": 63488 00:10:10.041 }, 00:10:10.041 { 00:10:10.041 "name": "BaseBdev4", 00:10:10.041 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:10.041 "is_configured": false, 00:10:10.041 "data_offset": 0, 00:10:10.041 "data_size": 0 00:10:10.041 } 00:10:10.041 ] 00:10:10.041 }' 00:10:10.041 17:31:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:10.041 17:31:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:10.301 17:31:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:10:10.301 17:31:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:10.301 17:31:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:10.301 [2024-11-27 17:31:41.343604] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:10:10.301 [2024-11-27 17:31:41.343925] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:10:10.301 [2024-11-27 17:31:41.343980] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:10:10.301 BaseBdev4 00:10:10.301 [2024-11-27 17:31:41.344327] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:10:10.301 [2024-11-27 17:31:41.344520] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:10:10.301 [2024-11-27 17:31:41.344575] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:10:10.301 [2024-11-27 17:31:41.344749] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:10.301 17:31:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:10.301 17:31:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev4 00:10:10.301 17:31:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev4 00:10:10.301 17:31:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:10:10.301 17:31:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:10:10.301 17:31:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:10:10.301 17:31:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:10:10.301 17:31:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:10:10.301 17:31:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:10.301 17:31:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:10.301 17:31:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:10.301 17:31:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:10:10.301 17:31:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:10.301 17:31:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:10.301 [ 00:10:10.301 { 00:10:10.301 "name": "BaseBdev4", 00:10:10.301 "aliases": [ 00:10:10.301 "4dc7b9df-13e9-4f07-a065-39fc02bd21e7" 00:10:10.301 ], 00:10:10.301 "product_name": "Malloc disk", 00:10:10.301 "block_size": 512, 00:10:10.301 "num_blocks": 65536, 00:10:10.301 "uuid": "4dc7b9df-13e9-4f07-a065-39fc02bd21e7", 00:10:10.301 "assigned_rate_limits": { 00:10:10.301 "rw_ios_per_sec": 0, 00:10:10.301 "rw_mbytes_per_sec": 0, 00:10:10.301 "r_mbytes_per_sec": 0, 00:10:10.301 "w_mbytes_per_sec": 0 00:10:10.301 }, 00:10:10.301 "claimed": true, 00:10:10.302 "claim_type": "exclusive_write", 00:10:10.302 "zoned": false, 00:10:10.302 "supported_io_types": { 00:10:10.302 "read": true, 00:10:10.302 "write": true, 00:10:10.302 "unmap": true, 00:10:10.302 "flush": true, 00:10:10.302 "reset": true, 00:10:10.302 "nvme_admin": false, 00:10:10.302 "nvme_io": false, 00:10:10.302 "nvme_io_md": false, 00:10:10.302 "write_zeroes": true, 00:10:10.302 "zcopy": true, 00:10:10.302 "get_zone_info": false, 00:10:10.302 "zone_management": false, 00:10:10.302 "zone_append": false, 00:10:10.302 "compare": false, 00:10:10.302 "compare_and_write": false, 00:10:10.302 "abort": true, 00:10:10.302 "seek_hole": false, 00:10:10.302 "seek_data": false, 00:10:10.302 "copy": true, 00:10:10.302 "nvme_iov_md": false 00:10:10.302 }, 00:10:10.302 "memory_domains": [ 00:10:10.302 { 00:10:10.302 "dma_device_id": "system", 00:10:10.302 "dma_device_type": 1 00:10:10.302 }, 00:10:10.302 { 00:10:10.302 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:10.302 "dma_device_type": 2 00:10:10.302 } 00:10:10.302 ], 00:10:10.302 "driver_specific": {} 00:10:10.302 } 00:10:10.302 ] 00:10:10.302 17:31:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:10.302 17:31:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:10:10.302 17:31:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:10:10.302 17:31:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:10.302 17:31:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid0 64 4 00:10:10.302 17:31:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:10.302 17:31:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:10.302 17:31:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:10.302 17:31:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:10.302 17:31:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:10.302 17:31:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:10.302 17:31:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:10.302 17:31:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:10.302 17:31:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:10.302 17:31:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:10.302 17:31:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:10.302 17:31:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:10.302 17:31:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:10.302 17:31:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:10.302 17:31:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:10.302 "name": "Existed_Raid", 00:10:10.302 "uuid": "28abdeee-fd43-4400-ad0d-a50dbb2ce588", 00:10:10.302 "strip_size_kb": 64, 00:10:10.302 "state": "online", 00:10:10.302 "raid_level": "raid0", 00:10:10.302 "superblock": true, 00:10:10.302 "num_base_bdevs": 4, 00:10:10.302 "num_base_bdevs_discovered": 4, 00:10:10.302 "num_base_bdevs_operational": 4, 00:10:10.302 "base_bdevs_list": [ 00:10:10.302 { 00:10:10.302 "name": "BaseBdev1", 00:10:10.302 "uuid": "e8d1c6db-0421-4402-82f2-5d530d912be1", 00:10:10.302 "is_configured": true, 00:10:10.302 "data_offset": 2048, 00:10:10.302 "data_size": 63488 00:10:10.302 }, 00:10:10.302 { 00:10:10.302 "name": "BaseBdev2", 00:10:10.302 "uuid": "4bada4e1-68cf-426f-8186-6d64e2747b15", 00:10:10.302 "is_configured": true, 00:10:10.302 "data_offset": 2048, 00:10:10.302 "data_size": 63488 00:10:10.302 }, 00:10:10.302 { 00:10:10.302 "name": "BaseBdev3", 00:10:10.302 "uuid": "a9515ebc-4169-4d6a-a46a-0ba1873f7a96", 00:10:10.302 "is_configured": true, 00:10:10.302 "data_offset": 2048, 00:10:10.302 "data_size": 63488 00:10:10.302 }, 00:10:10.302 { 00:10:10.302 "name": "BaseBdev4", 00:10:10.302 "uuid": "4dc7b9df-13e9-4f07-a065-39fc02bd21e7", 00:10:10.302 "is_configured": true, 00:10:10.302 "data_offset": 2048, 00:10:10.302 "data_size": 63488 00:10:10.302 } 00:10:10.302 ] 00:10:10.302 }' 00:10:10.302 17:31:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:10.302 17:31:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:10.872 17:31:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:10:10.872 17:31:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:10:10.872 17:31:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:10:10.872 17:31:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:10:10.872 17:31:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:10:10.872 17:31:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:10:10.872 17:31:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:10:10.872 17:31:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:10:10.872 17:31:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:10.872 17:31:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:10.872 [2024-11-27 17:31:41.811172] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:10.872 17:31:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:10.872 17:31:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:10.872 "name": "Existed_Raid", 00:10:10.872 "aliases": [ 00:10:10.872 "28abdeee-fd43-4400-ad0d-a50dbb2ce588" 00:10:10.872 ], 00:10:10.872 "product_name": "Raid Volume", 00:10:10.872 "block_size": 512, 00:10:10.872 "num_blocks": 253952, 00:10:10.872 "uuid": "28abdeee-fd43-4400-ad0d-a50dbb2ce588", 00:10:10.872 "assigned_rate_limits": { 00:10:10.872 "rw_ios_per_sec": 0, 00:10:10.872 "rw_mbytes_per_sec": 0, 00:10:10.872 "r_mbytes_per_sec": 0, 00:10:10.872 "w_mbytes_per_sec": 0 00:10:10.872 }, 00:10:10.872 "claimed": false, 00:10:10.872 "zoned": false, 00:10:10.872 "supported_io_types": { 00:10:10.872 "read": true, 00:10:10.872 "write": true, 00:10:10.872 "unmap": true, 00:10:10.872 "flush": true, 00:10:10.872 "reset": true, 00:10:10.872 "nvme_admin": false, 00:10:10.872 "nvme_io": false, 00:10:10.872 "nvme_io_md": false, 00:10:10.872 "write_zeroes": true, 00:10:10.872 "zcopy": false, 00:10:10.872 "get_zone_info": false, 00:10:10.872 "zone_management": false, 00:10:10.872 "zone_append": false, 00:10:10.872 "compare": false, 00:10:10.872 "compare_and_write": false, 00:10:10.872 "abort": false, 00:10:10.872 "seek_hole": false, 00:10:10.872 "seek_data": false, 00:10:10.872 "copy": false, 00:10:10.872 "nvme_iov_md": false 00:10:10.872 }, 00:10:10.872 "memory_domains": [ 00:10:10.872 { 00:10:10.872 "dma_device_id": "system", 00:10:10.872 "dma_device_type": 1 00:10:10.872 }, 00:10:10.872 { 00:10:10.872 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:10.872 "dma_device_type": 2 00:10:10.872 }, 00:10:10.872 { 00:10:10.872 "dma_device_id": "system", 00:10:10.872 "dma_device_type": 1 00:10:10.872 }, 00:10:10.872 { 00:10:10.872 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:10.872 "dma_device_type": 2 00:10:10.872 }, 00:10:10.872 { 00:10:10.872 "dma_device_id": "system", 00:10:10.872 "dma_device_type": 1 00:10:10.872 }, 00:10:10.872 { 00:10:10.872 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:10.872 "dma_device_type": 2 00:10:10.872 }, 00:10:10.872 { 00:10:10.872 "dma_device_id": "system", 00:10:10.872 "dma_device_type": 1 00:10:10.872 }, 00:10:10.872 { 00:10:10.872 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:10.872 "dma_device_type": 2 00:10:10.872 } 00:10:10.872 ], 00:10:10.872 "driver_specific": { 00:10:10.872 "raid": { 00:10:10.872 "uuid": "28abdeee-fd43-4400-ad0d-a50dbb2ce588", 00:10:10.872 "strip_size_kb": 64, 00:10:10.872 "state": "online", 00:10:10.872 "raid_level": "raid0", 00:10:10.872 "superblock": true, 00:10:10.872 "num_base_bdevs": 4, 00:10:10.872 "num_base_bdevs_discovered": 4, 00:10:10.872 "num_base_bdevs_operational": 4, 00:10:10.872 "base_bdevs_list": [ 00:10:10.872 { 00:10:10.872 "name": "BaseBdev1", 00:10:10.872 "uuid": "e8d1c6db-0421-4402-82f2-5d530d912be1", 00:10:10.872 "is_configured": true, 00:10:10.872 "data_offset": 2048, 00:10:10.872 "data_size": 63488 00:10:10.872 }, 00:10:10.872 { 00:10:10.873 "name": "BaseBdev2", 00:10:10.873 "uuid": "4bada4e1-68cf-426f-8186-6d64e2747b15", 00:10:10.873 "is_configured": true, 00:10:10.873 "data_offset": 2048, 00:10:10.873 "data_size": 63488 00:10:10.873 }, 00:10:10.873 { 00:10:10.873 "name": "BaseBdev3", 00:10:10.873 "uuid": "a9515ebc-4169-4d6a-a46a-0ba1873f7a96", 00:10:10.873 "is_configured": true, 00:10:10.873 "data_offset": 2048, 00:10:10.873 "data_size": 63488 00:10:10.873 }, 00:10:10.873 { 00:10:10.873 "name": "BaseBdev4", 00:10:10.873 "uuid": "4dc7b9df-13e9-4f07-a065-39fc02bd21e7", 00:10:10.873 "is_configured": true, 00:10:10.873 "data_offset": 2048, 00:10:10.873 "data_size": 63488 00:10:10.873 } 00:10:10.873 ] 00:10:10.873 } 00:10:10.873 } 00:10:10.873 }' 00:10:10.873 17:31:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:10.873 17:31:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:10:10.873 BaseBdev2 00:10:10.873 BaseBdev3 00:10:10.873 BaseBdev4' 00:10:10.873 17:31:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:10.873 17:31:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:10.873 17:31:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:10.873 17:31:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:10.873 17:31:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:10:10.873 17:31:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:10.873 17:31:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:10.873 17:31:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:10.873 17:31:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:10.873 17:31:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:10.873 17:31:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:10.873 17:31:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:10:10.873 17:31:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:10.873 17:31:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:10.873 17:31:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:10.873 17:31:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:10.873 17:31:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:10.873 17:31:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:10.873 17:31:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:10.873 17:31:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:10:10.873 17:31:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:10.873 17:31:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:10.873 17:31:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:10.873 17:31:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:11.133 17:31:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:11.133 17:31:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:11.133 17:31:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:11.133 17:31:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:10:11.133 17:31:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:11.133 17:31:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:11.133 17:31:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:11.133 17:31:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:11.133 17:31:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:11.133 17:31:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:11.133 17:31:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:10:11.133 17:31:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:11.133 17:31:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:11.133 [2024-11-27 17:31:42.142321] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:10:11.133 [2024-11-27 17:31:42.142393] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:11.133 [2024-11-27 17:31:42.142481] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:11.133 17:31:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:11.134 17:31:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:10:11.134 17:31:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy raid0 00:10:11.134 17:31:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:10:11.134 17:31:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # return 1 00:10:11.134 17:31:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:10:11.134 17:31:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline raid0 64 3 00:10:11.134 17:31:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:11.134 17:31:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:10:11.134 17:31:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:11.134 17:31:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:11.134 17:31:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:11.134 17:31:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:11.134 17:31:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:11.134 17:31:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:11.134 17:31:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:11.134 17:31:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:11.134 17:31:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:11.134 17:31:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:11.134 17:31:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:11.134 17:31:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:11.134 17:31:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:11.134 "name": "Existed_Raid", 00:10:11.134 "uuid": "28abdeee-fd43-4400-ad0d-a50dbb2ce588", 00:10:11.134 "strip_size_kb": 64, 00:10:11.134 "state": "offline", 00:10:11.134 "raid_level": "raid0", 00:10:11.134 "superblock": true, 00:10:11.134 "num_base_bdevs": 4, 00:10:11.134 "num_base_bdevs_discovered": 3, 00:10:11.134 "num_base_bdevs_operational": 3, 00:10:11.134 "base_bdevs_list": [ 00:10:11.134 { 00:10:11.134 "name": null, 00:10:11.134 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:11.134 "is_configured": false, 00:10:11.134 "data_offset": 0, 00:10:11.134 "data_size": 63488 00:10:11.134 }, 00:10:11.134 { 00:10:11.134 "name": "BaseBdev2", 00:10:11.134 "uuid": "4bada4e1-68cf-426f-8186-6d64e2747b15", 00:10:11.134 "is_configured": true, 00:10:11.134 "data_offset": 2048, 00:10:11.134 "data_size": 63488 00:10:11.134 }, 00:10:11.134 { 00:10:11.134 "name": "BaseBdev3", 00:10:11.134 "uuid": "a9515ebc-4169-4d6a-a46a-0ba1873f7a96", 00:10:11.134 "is_configured": true, 00:10:11.134 "data_offset": 2048, 00:10:11.134 "data_size": 63488 00:10:11.134 }, 00:10:11.134 { 00:10:11.134 "name": "BaseBdev4", 00:10:11.134 "uuid": "4dc7b9df-13e9-4f07-a065-39fc02bd21e7", 00:10:11.134 "is_configured": true, 00:10:11.134 "data_offset": 2048, 00:10:11.134 "data_size": 63488 00:10:11.134 } 00:10:11.134 ] 00:10:11.134 }' 00:10:11.134 17:31:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:11.134 17:31:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:11.704 17:31:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:10:11.704 17:31:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:11.704 17:31:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:11.704 17:31:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:11.704 17:31:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:11.704 17:31:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:10:11.704 17:31:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:11.704 17:31:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:10:11.704 17:31:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:10:11.704 17:31:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:10:11.704 17:31:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:11.704 17:31:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:11.704 [2024-11-27 17:31:42.661856] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:10:11.704 17:31:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:11.704 17:31:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:10:11.704 17:31:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:11.704 17:31:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:11.704 17:31:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:10:11.704 17:31:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:11.704 17:31:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:11.704 17:31:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:11.704 17:31:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:10:11.704 17:31:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:10:11.704 17:31:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:10:11.704 17:31:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:11.704 17:31:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:11.704 [2024-11-27 17:31:42.738240] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:10:11.704 17:31:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:11.704 17:31:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:10:11.704 17:31:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:11.704 17:31:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:11.704 17:31:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:11.704 17:31:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:11.704 17:31:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:10:11.704 17:31:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:11.704 17:31:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:10:11.704 17:31:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:10:11.704 17:31:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev4 00:10:11.704 17:31:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:11.704 17:31:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:11.704 [2024-11-27 17:31:42.818516] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev4 00:10:11.704 [2024-11-27 17:31:42.818610] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:10:11.704 17:31:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:11.704 17:31:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:10:11.704 17:31:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:11.704 17:31:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:10:11.704 17:31:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:11.704 17:31:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:11.704 17:31:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:11.704 17:31:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:11.704 17:31:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:10:11.704 17:31:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:10:11.704 17:31:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 4 -gt 2 ']' 00:10:11.704 17:31:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:10:11.704 17:31:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:11.704 17:31:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:10:11.966 17:31:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:11.966 17:31:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:11.966 BaseBdev2 00:10:11.966 17:31:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:11.966 17:31:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:10:11.966 17:31:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:10:11.966 17:31:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:10:11.966 17:31:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:10:11.966 17:31:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:10:11.966 17:31:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:10:11.966 17:31:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:10:11.966 17:31:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:11.966 17:31:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:11.966 17:31:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:11.966 17:31:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:10:11.966 17:31:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:11.966 17:31:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:11.966 [ 00:10:11.966 { 00:10:11.966 "name": "BaseBdev2", 00:10:11.966 "aliases": [ 00:10:11.966 "9b157ea2-17ce-4a23-a7fa-f9310170f402" 00:10:11.966 ], 00:10:11.966 "product_name": "Malloc disk", 00:10:11.966 "block_size": 512, 00:10:11.966 "num_blocks": 65536, 00:10:11.966 "uuid": "9b157ea2-17ce-4a23-a7fa-f9310170f402", 00:10:11.966 "assigned_rate_limits": { 00:10:11.966 "rw_ios_per_sec": 0, 00:10:11.966 "rw_mbytes_per_sec": 0, 00:10:11.966 "r_mbytes_per_sec": 0, 00:10:11.966 "w_mbytes_per_sec": 0 00:10:11.966 }, 00:10:11.966 "claimed": false, 00:10:11.966 "zoned": false, 00:10:11.966 "supported_io_types": { 00:10:11.966 "read": true, 00:10:11.966 "write": true, 00:10:11.966 "unmap": true, 00:10:11.966 "flush": true, 00:10:11.966 "reset": true, 00:10:11.966 "nvme_admin": false, 00:10:11.966 "nvme_io": false, 00:10:11.966 "nvme_io_md": false, 00:10:11.966 "write_zeroes": true, 00:10:11.966 "zcopy": true, 00:10:11.966 "get_zone_info": false, 00:10:11.966 "zone_management": false, 00:10:11.966 "zone_append": false, 00:10:11.966 "compare": false, 00:10:11.966 "compare_and_write": false, 00:10:11.966 "abort": true, 00:10:11.966 "seek_hole": false, 00:10:11.966 "seek_data": false, 00:10:11.966 "copy": true, 00:10:11.966 "nvme_iov_md": false 00:10:11.966 }, 00:10:11.966 "memory_domains": [ 00:10:11.966 { 00:10:11.966 "dma_device_id": "system", 00:10:11.966 "dma_device_type": 1 00:10:11.966 }, 00:10:11.966 { 00:10:11.966 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:11.966 "dma_device_type": 2 00:10:11.966 } 00:10:11.966 ], 00:10:11.966 "driver_specific": {} 00:10:11.966 } 00:10:11.966 ] 00:10:11.966 17:31:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:11.966 17:31:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:10:11.966 17:31:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:10:11.966 17:31:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:11.966 17:31:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:10:11.966 17:31:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:11.966 17:31:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:11.966 BaseBdev3 00:10:11.966 17:31:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:11.966 17:31:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:10:11.966 17:31:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:10:11.966 17:31:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:10:11.966 17:31:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:10:11.966 17:31:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:10:11.966 17:31:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:10:11.966 17:31:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:10:11.966 17:31:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:11.966 17:31:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:11.966 17:31:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:11.966 17:31:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:10:11.966 17:31:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:11.966 17:31:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:11.966 [ 00:10:11.966 { 00:10:11.966 "name": "BaseBdev3", 00:10:11.966 "aliases": [ 00:10:11.966 "870d6740-7f69-41d7-9e66-def1fd7cfbf9" 00:10:11.966 ], 00:10:11.966 "product_name": "Malloc disk", 00:10:11.966 "block_size": 512, 00:10:11.966 "num_blocks": 65536, 00:10:11.966 "uuid": "870d6740-7f69-41d7-9e66-def1fd7cfbf9", 00:10:11.966 "assigned_rate_limits": { 00:10:11.966 "rw_ios_per_sec": 0, 00:10:11.966 "rw_mbytes_per_sec": 0, 00:10:11.966 "r_mbytes_per_sec": 0, 00:10:11.966 "w_mbytes_per_sec": 0 00:10:11.966 }, 00:10:11.966 "claimed": false, 00:10:11.966 "zoned": false, 00:10:11.966 "supported_io_types": { 00:10:11.966 "read": true, 00:10:11.966 "write": true, 00:10:11.966 "unmap": true, 00:10:11.966 "flush": true, 00:10:11.966 "reset": true, 00:10:11.966 "nvme_admin": false, 00:10:11.966 "nvme_io": false, 00:10:11.966 "nvme_io_md": false, 00:10:11.967 "write_zeroes": true, 00:10:11.967 "zcopy": true, 00:10:11.967 "get_zone_info": false, 00:10:11.967 "zone_management": false, 00:10:11.967 "zone_append": false, 00:10:11.967 "compare": false, 00:10:11.967 "compare_and_write": false, 00:10:11.967 "abort": true, 00:10:11.967 "seek_hole": false, 00:10:11.967 "seek_data": false, 00:10:11.967 "copy": true, 00:10:11.967 "nvme_iov_md": false 00:10:11.967 }, 00:10:11.967 "memory_domains": [ 00:10:11.967 { 00:10:11.967 "dma_device_id": "system", 00:10:11.967 "dma_device_type": 1 00:10:11.967 }, 00:10:11.967 { 00:10:11.967 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:11.967 "dma_device_type": 2 00:10:11.967 } 00:10:11.967 ], 00:10:11.967 "driver_specific": {} 00:10:11.967 } 00:10:11.967 ] 00:10:11.967 17:31:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:11.967 17:31:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:10:11.967 17:31:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:10:11.967 17:31:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:11.967 17:31:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:10:11.967 17:31:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:11.967 17:31:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:11.967 BaseBdev4 00:10:11.967 17:31:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:11.967 17:31:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev4 00:10:11.967 17:31:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev4 00:10:11.967 17:31:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:10:11.967 17:31:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:10:11.967 17:31:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:10:11.967 17:31:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:10:11.967 17:31:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:10:11.967 17:31:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:11.967 17:31:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:11.967 17:31:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:11.967 17:31:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:10:11.967 17:31:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:11.967 17:31:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:11.967 [ 00:10:11.967 { 00:10:11.967 "name": "BaseBdev4", 00:10:11.967 "aliases": [ 00:10:11.967 "9daa1097-6cca-4a82-a047-0394016ca5d4" 00:10:11.967 ], 00:10:11.967 "product_name": "Malloc disk", 00:10:11.967 "block_size": 512, 00:10:11.967 "num_blocks": 65536, 00:10:11.967 "uuid": "9daa1097-6cca-4a82-a047-0394016ca5d4", 00:10:11.967 "assigned_rate_limits": { 00:10:11.967 "rw_ios_per_sec": 0, 00:10:11.967 "rw_mbytes_per_sec": 0, 00:10:11.967 "r_mbytes_per_sec": 0, 00:10:11.967 "w_mbytes_per_sec": 0 00:10:11.967 }, 00:10:11.967 "claimed": false, 00:10:11.967 "zoned": false, 00:10:11.967 "supported_io_types": { 00:10:11.967 "read": true, 00:10:11.967 "write": true, 00:10:11.967 "unmap": true, 00:10:11.967 "flush": true, 00:10:11.967 "reset": true, 00:10:11.967 "nvme_admin": false, 00:10:11.967 "nvme_io": false, 00:10:11.967 "nvme_io_md": false, 00:10:11.967 "write_zeroes": true, 00:10:11.967 "zcopy": true, 00:10:11.967 "get_zone_info": false, 00:10:11.967 "zone_management": false, 00:10:11.967 "zone_append": false, 00:10:11.967 "compare": false, 00:10:11.967 "compare_and_write": false, 00:10:11.967 "abort": true, 00:10:11.967 "seek_hole": false, 00:10:11.967 "seek_data": false, 00:10:11.967 "copy": true, 00:10:11.967 "nvme_iov_md": false 00:10:11.967 }, 00:10:11.967 "memory_domains": [ 00:10:11.967 { 00:10:11.967 "dma_device_id": "system", 00:10:11.967 "dma_device_type": 1 00:10:11.967 }, 00:10:11.967 { 00:10:11.967 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:11.967 "dma_device_type": 2 00:10:11.967 } 00:10:11.967 ], 00:10:11.967 "driver_specific": {} 00:10:11.967 } 00:10:11.967 ] 00:10:11.967 17:31:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:11.967 17:31:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:10:11.967 17:31:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:10:11.967 17:31:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:11.967 17:31:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:10:11.967 17:31:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:11.967 17:31:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:11.967 [2024-11-27 17:31:43.075205] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:10:11.967 [2024-11-27 17:31:43.075317] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:10:11.967 [2024-11-27 17:31:43.075378] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:11.967 [2024-11-27 17:31:43.077467] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:11.967 [2024-11-27 17:31:43.077553] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:10:11.967 17:31:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:11.967 17:31:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:10:11.967 17:31:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:11.967 17:31:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:11.967 17:31:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:11.967 17:31:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:11.967 17:31:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:11.967 17:31:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:11.967 17:31:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:11.967 17:31:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:11.967 17:31:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:11.967 17:31:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:11.967 17:31:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:11.967 17:31:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:11.967 17:31:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:11.967 17:31:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:11.967 17:31:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:11.967 "name": "Existed_Raid", 00:10:11.967 "uuid": "9caeee85-2c90-4e9e-9c6f-21b72604993a", 00:10:11.967 "strip_size_kb": 64, 00:10:11.967 "state": "configuring", 00:10:11.967 "raid_level": "raid0", 00:10:11.967 "superblock": true, 00:10:11.967 "num_base_bdevs": 4, 00:10:11.967 "num_base_bdevs_discovered": 3, 00:10:11.967 "num_base_bdevs_operational": 4, 00:10:11.967 "base_bdevs_list": [ 00:10:11.967 { 00:10:11.967 "name": "BaseBdev1", 00:10:11.967 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:11.967 "is_configured": false, 00:10:11.967 "data_offset": 0, 00:10:11.967 "data_size": 0 00:10:11.967 }, 00:10:11.967 { 00:10:11.967 "name": "BaseBdev2", 00:10:11.967 "uuid": "9b157ea2-17ce-4a23-a7fa-f9310170f402", 00:10:11.967 "is_configured": true, 00:10:11.967 "data_offset": 2048, 00:10:11.967 "data_size": 63488 00:10:11.967 }, 00:10:11.967 { 00:10:11.967 "name": "BaseBdev3", 00:10:11.967 "uuid": "870d6740-7f69-41d7-9e66-def1fd7cfbf9", 00:10:11.967 "is_configured": true, 00:10:11.967 "data_offset": 2048, 00:10:11.967 "data_size": 63488 00:10:11.967 }, 00:10:11.967 { 00:10:11.967 "name": "BaseBdev4", 00:10:11.967 "uuid": "9daa1097-6cca-4a82-a047-0394016ca5d4", 00:10:11.967 "is_configured": true, 00:10:11.967 "data_offset": 2048, 00:10:11.967 "data_size": 63488 00:10:11.967 } 00:10:11.967 ] 00:10:11.967 }' 00:10:11.968 17:31:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:11.968 17:31:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:12.537 17:31:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:10:12.537 17:31:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:12.537 17:31:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:12.537 [2024-11-27 17:31:43.526417] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:10:12.537 17:31:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:12.537 17:31:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:10:12.537 17:31:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:12.537 17:31:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:12.537 17:31:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:12.537 17:31:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:12.537 17:31:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:12.537 17:31:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:12.537 17:31:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:12.537 17:31:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:12.537 17:31:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:12.537 17:31:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:12.537 17:31:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:12.537 17:31:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:12.537 17:31:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:12.537 17:31:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:12.537 17:31:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:12.537 "name": "Existed_Raid", 00:10:12.537 "uuid": "9caeee85-2c90-4e9e-9c6f-21b72604993a", 00:10:12.537 "strip_size_kb": 64, 00:10:12.537 "state": "configuring", 00:10:12.537 "raid_level": "raid0", 00:10:12.537 "superblock": true, 00:10:12.537 "num_base_bdevs": 4, 00:10:12.537 "num_base_bdevs_discovered": 2, 00:10:12.537 "num_base_bdevs_operational": 4, 00:10:12.538 "base_bdevs_list": [ 00:10:12.538 { 00:10:12.538 "name": "BaseBdev1", 00:10:12.538 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:12.538 "is_configured": false, 00:10:12.538 "data_offset": 0, 00:10:12.538 "data_size": 0 00:10:12.538 }, 00:10:12.538 { 00:10:12.538 "name": null, 00:10:12.538 "uuid": "9b157ea2-17ce-4a23-a7fa-f9310170f402", 00:10:12.538 "is_configured": false, 00:10:12.538 "data_offset": 0, 00:10:12.538 "data_size": 63488 00:10:12.538 }, 00:10:12.538 { 00:10:12.538 "name": "BaseBdev3", 00:10:12.538 "uuid": "870d6740-7f69-41d7-9e66-def1fd7cfbf9", 00:10:12.538 "is_configured": true, 00:10:12.538 "data_offset": 2048, 00:10:12.538 "data_size": 63488 00:10:12.538 }, 00:10:12.538 { 00:10:12.538 "name": "BaseBdev4", 00:10:12.538 "uuid": "9daa1097-6cca-4a82-a047-0394016ca5d4", 00:10:12.538 "is_configured": true, 00:10:12.538 "data_offset": 2048, 00:10:12.538 "data_size": 63488 00:10:12.538 } 00:10:12.538 ] 00:10:12.538 }' 00:10:12.538 17:31:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:12.538 17:31:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:12.797 17:31:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:10:12.797 17:31:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:12.797 17:31:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:12.797 17:31:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:12.797 17:31:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:12.797 17:31:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:10:12.797 17:31:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:10:12.797 17:31:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:12.797 17:31:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:13.057 [2024-11-27 17:31:43.990543] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:13.057 BaseBdev1 00:10:13.058 17:31:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:13.058 17:31:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:10:13.058 17:31:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:10:13.058 17:31:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:10:13.058 17:31:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:10:13.058 17:31:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:10:13.058 17:31:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:10:13.058 17:31:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:10:13.058 17:31:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:13.058 17:31:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:13.058 17:31:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:13.058 17:31:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:10:13.058 17:31:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:13.058 17:31:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:13.058 [ 00:10:13.058 { 00:10:13.058 "name": "BaseBdev1", 00:10:13.058 "aliases": [ 00:10:13.058 "fcdc3321-c813-42cb-bcda-07f061c745c7" 00:10:13.058 ], 00:10:13.058 "product_name": "Malloc disk", 00:10:13.058 "block_size": 512, 00:10:13.058 "num_blocks": 65536, 00:10:13.058 "uuid": "fcdc3321-c813-42cb-bcda-07f061c745c7", 00:10:13.058 "assigned_rate_limits": { 00:10:13.058 "rw_ios_per_sec": 0, 00:10:13.058 "rw_mbytes_per_sec": 0, 00:10:13.058 "r_mbytes_per_sec": 0, 00:10:13.058 "w_mbytes_per_sec": 0 00:10:13.058 }, 00:10:13.058 "claimed": true, 00:10:13.058 "claim_type": "exclusive_write", 00:10:13.058 "zoned": false, 00:10:13.058 "supported_io_types": { 00:10:13.058 "read": true, 00:10:13.058 "write": true, 00:10:13.058 "unmap": true, 00:10:13.058 "flush": true, 00:10:13.058 "reset": true, 00:10:13.058 "nvme_admin": false, 00:10:13.058 "nvme_io": false, 00:10:13.058 "nvme_io_md": false, 00:10:13.058 "write_zeroes": true, 00:10:13.058 "zcopy": true, 00:10:13.058 "get_zone_info": false, 00:10:13.058 "zone_management": false, 00:10:13.058 "zone_append": false, 00:10:13.058 "compare": false, 00:10:13.058 "compare_and_write": false, 00:10:13.058 "abort": true, 00:10:13.058 "seek_hole": false, 00:10:13.058 "seek_data": false, 00:10:13.058 "copy": true, 00:10:13.058 "nvme_iov_md": false 00:10:13.058 }, 00:10:13.058 "memory_domains": [ 00:10:13.058 { 00:10:13.058 "dma_device_id": "system", 00:10:13.058 "dma_device_type": 1 00:10:13.058 }, 00:10:13.058 { 00:10:13.058 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:13.058 "dma_device_type": 2 00:10:13.058 } 00:10:13.058 ], 00:10:13.058 "driver_specific": {} 00:10:13.058 } 00:10:13.058 ] 00:10:13.058 17:31:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:13.058 17:31:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:10:13.058 17:31:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:10:13.058 17:31:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:13.058 17:31:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:13.058 17:31:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:13.058 17:31:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:13.058 17:31:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:13.058 17:31:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:13.058 17:31:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:13.058 17:31:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:13.058 17:31:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:13.058 17:31:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:13.058 17:31:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:13.058 17:31:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:13.058 17:31:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:13.058 17:31:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:13.058 17:31:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:13.058 "name": "Existed_Raid", 00:10:13.058 "uuid": "9caeee85-2c90-4e9e-9c6f-21b72604993a", 00:10:13.058 "strip_size_kb": 64, 00:10:13.058 "state": "configuring", 00:10:13.058 "raid_level": "raid0", 00:10:13.058 "superblock": true, 00:10:13.058 "num_base_bdevs": 4, 00:10:13.058 "num_base_bdevs_discovered": 3, 00:10:13.058 "num_base_bdevs_operational": 4, 00:10:13.058 "base_bdevs_list": [ 00:10:13.058 { 00:10:13.058 "name": "BaseBdev1", 00:10:13.058 "uuid": "fcdc3321-c813-42cb-bcda-07f061c745c7", 00:10:13.058 "is_configured": true, 00:10:13.058 "data_offset": 2048, 00:10:13.058 "data_size": 63488 00:10:13.058 }, 00:10:13.058 { 00:10:13.058 "name": null, 00:10:13.058 "uuid": "9b157ea2-17ce-4a23-a7fa-f9310170f402", 00:10:13.058 "is_configured": false, 00:10:13.058 "data_offset": 0, 00:10:13.058 "data_size": 63488 00:10:13.058 }, 00:10:13.058 { 00:10:13.058 "name": "BaseBdev3", 00:10:13.058 "uuid": "870d6740-7f69-41d7-9e66-def1fd7cfbf9", 00:10:13.058 "is_configured": true, 00:10:13.058 "data_offset": 2048, 00:10:13.058 "data_size": 63488 00:10:13.058 }, 00:10:13.058 { 00:10:13.058 "name": "BaseBdev4", 00:10:13.058 "uuid": "9daa1097-6cca-4a82-a047-0394016ca5d4", 00:10:13.058 "is_configured": true, 00:10:13.058 "data_offset": 2048, 00:10:13.058 "data_size": 63488 00:10:13.058 } 00:10:13.058 ] 00:10:13.058 }' 00:10:13.058 17:31:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:13.058 17:31:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:13.318 17:31:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:13.318 17:31:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:10:13.318 17:31:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:13.318 17:31:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:13.318 17:31:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:13.578 17:31:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:10:13.578 17:31:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:10:13.578 17:31:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:13.578 17:31:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:13.578 [2024-11-27 17:31:44.533654] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:10:13.578 17:31:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:13.578 17:31:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:10:13.578 17:31:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:13.578 17:31:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:13.578 17:31:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:13.578 17:31:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:13.578 17:31:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:13.578 17:31:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:13.579 17:31:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:13.579 17:31:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:13.579 17:31:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:13.579 17:31:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:13.579 17:31:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:13.579 17:31:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:13.579 17:31:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:13.579 17:31:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:13.579 17:31:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:13.579 "name": "Existed_Raid", 00:10:13.579 "uuid": "9caeee85-2c90-4e9e-9c6f-21b72604993a", 00:10:13.579 "strip_size_kb": 64, 00:10:13.579 "state": "configuring", 00:10:13.579 "raid_level": "raid0", 00:10:13.579 "superblock": true, 00:10:13.579 "num_base_bdevs": 4, 00:10:13.579 "num_base_bdevs_discovered": 2, 00:10:13.579 "num_base_bdevs_operational": 4, 00:10:13.579 "base_bdevs_list": [ 00:10:13.579 { 00:10:13.579 "name": "BaseBdev1", 00:10:13.579 "uuid": "fcdc3321-c813-42cb-bcda-07f061c745c7", 00:10:13.579 "is_configured": true, 00:10:13.579 "data_offset": 2048, 00:10:13.579 "data_size": 63488 00:10:13.579 }, 00:10:13.579 { 00:10:13.579 "name": null, 00:10:13.579 "uuid": "9b157ea2-17ce-4a23-a7fa-f9310170f402", 00:10:13.579 "is_configured": false, 00:10:13.579 "data_offset": 0, 00:10:13.579 "data_size": 63488 00:10:13.579 }, 00:10:13.579 { 00:10:13.579 "name": null, 00:10:13.579 "uuid": "870d6740-7f69-41d7-9e66-def1fd7cfbf9", 00:10:13.579 "is_configured": false, 00:10:13.579 "data_offset": 0, 00:10:13.579 "data_size": 63488 00:10:13.579 }, 00:10:13.579 { 00:10:13.579 "name": "BaseBdev4", 00:10:13.579 "uuid": "9daa1097-6cca-4a82-a047-0394016ca5d4", 00:10:13.579 "is_configured": true, 00:10:13.579 "data_offset": 2048, 00:10:13.579 "data_size": 63488 00:10:13.579 } 00:10:13.579 ] 00:10:13.579 }' 00:10:13.579 17:31:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:13.579 17:31:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:13.839 17:31:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:13.839 17:31:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:13.839 17:31:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:13.839 17:31:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:10:13.839 17:31:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:13.839 17:31:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:10:13.839 17:31:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:10:13.839 17:31:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:13.839 17:31:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:13.839 [2024-11-27 17:31:45.004925] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:13.839 17:31:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:13.839 17:31:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:10:13.839 17:31:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:13.839 17:31:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:13.839 17:31:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:13.839 17:31:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:13.839 17:31:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:13.839 17:31:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:13.839 17:31:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:13.839 17:31:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:13.839 17:31:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:13.839 17:31:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:13.839 17:31:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:13.839 17:31:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:13.839 17:31:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:14.099 17:31:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:14.099 17:31:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:14.099 "name": "Existed_Raid", 00:10:14.099 "uuid": "9caeee85-2c90-4e9e-9c6f-21b72604993a", 00:10:14.099 "strip_size_kb": 64, 00:10:14.099 "state": "configuring", 00:10:14.099 "raid_level": "raid0", 00:10:14.099 "superblock": true, 00:10:14.099 "num_base_bdevs": 4, 00:10:14.099 "num_base_bdevs_discovered": 3, 00:10:14.099 "num_base_bdevs_operational": 4, 00:10:14.099 "base_bdevs_list": [ 00:10:14.099 { 00:10:14.099 "name": "BaseBdev1", 00:10:14.099 "uuid": "fcdc3321-c813-42cb-bcda-07f061c745c7", 00:10:14.099 "is_configured": true, 00:10:14.099 "data_offset": 2048, 00:10:14.099 "data_size": 63488 00:10:14.099 }, 00:10:14.099 { 00:10:14.099 "name": null, 00:10:14.099 "uuid": "9b157ea2-17ce-4a23-a7fa-f9310170f402", 00:10:14.099 "is_configured": false, 00:10:14.099 "data_offset": 0, 00:10:14.099 "data_size": 63488 00:10:14.099 }, 00:10:14.099 { 00:10:14.099 "name": "BaseBdev3", 00:10:14.099 "uuid": "870d6740-7f69-41d7-9e66-def1fd7cfbf9", 00:10:14.099 "is_configured": true, 00:10:14.099 "data_offset": 2048, 00:10:14.099 "data_size": 63488 00:10:14.099 }, 00:10:14.099 { 00:10:14.099 "name": "BaseBdev4", 00:10:14.099 "uuid": "9daa1097-6cca-4a82-a047-0394016ca5d4", 00:10:14.099 "is_configured": true, 00:10:14.099 "data_offset": 2048, 00:10:14.099 "data_size": 63488 00:10:14.099 } 00:10:14.099 ] 00:10:14.099 }' 00:10:14.099 17:31:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:14.099 17:31:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:14.359 17:31:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:14.359 17:31:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:14.359 17:31:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:14.359 17:31:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:10:14.359 17:31:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:14.359 17:31:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:10:14.359 17:31:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:10:14.359 17:31:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:14.359 17:31:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:14.359 [2024-11-27 17:31:45.492092] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:10:14.359 17:31:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:14.359 17:31:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:10:14.359 17:31:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:14.359 17:31:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:14.359 17:31:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:14.359 17:31:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:14.359 17:31:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:14.359 17:31:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:14.359 17:31:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:14.359 17:31:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:14.359 17:31:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:14.359 17:31:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:14.359 17:31:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:14.359 17:31:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:14.359 17:31:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:14.359 17:31:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:14.619 17:31:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:14.619 "name": "Existed_Raid", 00:10:14.619 "uuid": "9caeee85-2c90-4e9e-9c6f-21b72604993a", 00:10:14.619 "strip_size_kb": 64, 00:10:14.619 "state": "configuring", 00:10:14.619 "raid_level": "raid0", 00:10:14.619 "superblock": true, 00:10:14.619 "num_base_bdevs": 4, 00:10:14.619 "num_base_bdevs_discovered": 2, 00:10:14.619 "num_base_bdevs_operational": 4, 00:10:14.619 "base_bdevs_list": [ 00:10:14.619 { 00:10:14.619 "name": null, 00:10:14.619 "uuid": "fcdc3321-c813-42cb-bcda-07f061c745c7", 00:10:14.619 "is_configured": false, 00:10:14.619 "data_offset": 0, 00:10:14.619 "data_size": 63488 00:10:14.619 }, 00:10:14.619 { 00:10:14.619 "name": null, 00:10:14.619 "uuid": "9b157ea2-17ce-4a23-a7fa-f9310170f402", 00:10:14.619 "is_configured": false, 00:10:14.619 "data_offset": 0, 00:10:14.619 "data_size": 63488 00:10:14.619 }, 00:10:14.619 { 00:10:14.619 "name": "BaseBdev3", 00:10:14.619 "uuid": "870d6740-7f69-41d7-9e66-def1fd7cfbf9", 00:10:14.619 "is_configured": true, 00:10:14.619 "data_offset": 2048, 00:10:14.619 "data_size": 63488 00:10:14.619 }, 00:10:14.619 { 00:10:14.619 "name": "BaseBdev4", 00:10:14.619 "uuid": "9daa1097-6cca-4a82-a047-0394016ca5d4", 00:10:14.619 "is_configured": true, 00:10:14.619 "data_offset": 2048, 00:10:14.619 "data_size": 63488 00:10:14.619 } 00:10:14.619 ] 00:10:14.619 }' 00:10:14.619 17:31:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:14.619 17:31:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:14.879 17:31:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:14.879 17:31:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:10:14.879 17:31:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:14.879 17:31:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:14.879 17:31:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:14.879 17:31:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:10:14.879 17:31:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:10:14.879 17:31:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:14.879 17:31:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:14.879 [2024-11-27 17:31:45.986889] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:14.879 17:31:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:14.879 17:31:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid0 64 4 00:10:14.879 17:31:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:14.879 17:31:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:14.879 17:31:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:14.879 17:31:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:14.879 17:31:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:14.879 17:31:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:14.879 17:31:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:14.879 17:31:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:14.879 17:31:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:14.879 17:31:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:14.879 17:31:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:14.879 17:31:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:14.879 17:31:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:14.879 17:31:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:14.879 17:31:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:14.879 "name": "Existed_Raid", 00:10:14.879 "uuid": "9caeee85-2c90-4e9e-9c6f-21b72604993a", 00:10:14.879 "strip_size_kb": 64, 00:10:14.879 "state": "configuring", 00:10:14.879 "raid_level": "raid0", 00:10:14.879 "superblock": true, 00:10:14.879 "num_base_bdevs": 4, 00:10:14.879 "num_base_bdevs_discovered": 3, 00:10:14.879 "num_base_bdevs_operational": 4, 00:10:14.879 "base_bdevs_list": [ 00:10:14.879 { 00:10:14.879 "name": null, 00:10:14.879 "uuid": "fcdc3321-c813-42cb-bcda-07f061c745c7", 00:10:14.879 "is_configured": false, 00:10:14.879 "data_offset": 0, 00:10:14.879 "data_size": 63488 00:10:14.879 }, 00:10:14.879 { 00:10:14.879 "name": "BaseBdev2", 00:10:14.879 "uuid": "9b157ea2-17ce-4a23-a7fa-f9310170f402", 00:10:14.879 "is_configured": true, 00:10:14.879 "data_offset": 2048, 00:10:14.879 "data_size": 63488 00:10:14.879 }, 00:10:14.879 { 00:10:14.879 "name": "BaseBdev3", 00:10:14.879 "uuid": "870d6740-7f69-41d7-9e66-def1fd7cfbf9", 00:10:14.879 "is_configured": true, 00:10:14.879 "data_offset": 2048, 00:10:14.879 "data_size": 63488 00:10:14.879 }, 00:10:14.879 { 00:10:14.879 "name": "BaseBdev4", 00:10:14.879 "uuid": "9daa1097-6cca-4a82-a047-0394016ca5d4", 00:10:14.879 "is_configured": true, 00:10:14.879 "data_offset": 2048, 00:10:14.879 "data_size": 63488 00:10:14.879 } 00:10:14.879 ] 00:10:14.879 }' 00:10:14.879 17:31:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:14.879 17:31:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:15.450 17:31:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:10:15.450 17:31:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:15.450 17:31:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:15.450 17:31:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:15.450 17:31:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:15.450 17:31:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:10:15.450 17:31:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:15.450 17:31:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:10:15.450 17:31:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:15.450 17:31:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:15.450 17:31:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:15.450 17:31:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u fcdc3321-c813-42cb-bcda-07f061c745c7 00:10:15.450 17:31:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:15.450 17:31:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:15.450 NewBaseBdev 00:10:15.450 [2024-11-27 17:31:46.550758] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:10:15.450 [2024-11-27 17:31:46.550957] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:10:15.450 [2024-11-27 17:31:46.550971] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:10:15.450 [2024-11-27 17:31:46.551285] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002a10 00:10:15.450 [2024-11-27 17:31:46.551416] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:10:15.450 [2024-11-27 17:31:46.551427] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001c80 00:10:15.450 [2024-11-27 17:31:46.551537] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:15.450 17:31:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:15.450 17:31:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:10:15.450 17:31:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=NewBaseBdev 00:10:15.450 17:31:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:10:15.450 17:31:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:10:15.450 17:31:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:10:15.450 17:31:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:10:15.450 17:31:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:10:15.450 17:31:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:15.450 17:31:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:15.450 17:31:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:15.450 17:31:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:10:15.450 17:31:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:15.450 17:31:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:15.450 [ 00:10:15.450 { 00:10:15.450 "name": "NewBaseBdev", 00:10:15.450 "aliases": [ 00:10:15.450 "fcdc3321-c813-42cb-bcda-07f061c745c7" 00:10:15.450 ], 00:10:15.450 "product_name": "Malloc disk", 00:10:15.450 "block_size": 512, 00:10:15.450 "num_blocks": 65536, 00:10:15.450 "uuid": "fcdc3321-c813-42cb-bcda-07f061c745c7", 00:10:15.450 "assigned_rate_limits": { 00:10:15.450 "rw_ios_per_sec": 0, 00:10:15.450 "rw_mbytes_per_sec": 0, 00:10:15.450 "r_mbytes_per_sec": 0, 00:10:15.450 "w_mbytes_per_sec": 0 00:10:15.450 }, 00:10:15.450 "claimed": true, 00:10:15.450 "claim_type": "exclusive_write", 00:10:15.450 "zoned": false, 00:10:15.450 "supported_io_types": { 00:10:15.450 "read": true, 00:10:15.450 "write": true, 00:10:15.450 "unmap": true, 00:10:15.450 "flush": true, 00:10:15.450 "reset": true, 00:10:15.450 "nvme_admin": false, 00:10:15.450 "nvme_io": false, 00:10:15.450 "nvme_io_md": false, 00:10:15.450 "write_zeroes": true, 00:10:15.450 "zcopy": true, 00:10:15.450 "get_zone_info": false, 00:10:15.450 "zone_management": false, 00:10:15.450 "zone_append": false, 00:10:15.450 "compare": false, 00:10:15.450 "compare_and_write": false, 00:10:15.450 "abort": true, 00:10:15.450 "seek_hole": false, 00:10:15.450 "seek_data": false, 00:10:15.450 "copy": true, 00:10:15.450 "nvme_iov_md": false 00:10:15.450 }, 00:10:15.450 "memory_domains": [ 00:10:15.450 { 00:10:15.450 "dma_device_id": "system", 00:10:15.450 "dma_device_type": 1 00:10:15.450 }, 00:10:15.450 { 00:10:15.450 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:15.450 "dma_device_type": 2 00:10:15.450 } 00:10:15.450 ], 00:10:15.450 "driver_specific": {} 00:10:15.450 } 00:10:15.450 ] 00:10:15.450 17:31:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:15.450 17:31:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:10:15.450 17:31:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid0 64 4 00:10:15.450 17:31:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:15.450 17:31:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:15.450 17:31:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:15.450 17:31:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:15.450 17:31:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:15.450 17:31:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:15.450 17:31:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:15.450 17:31:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:15.450 17:31:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:15.450 17:31:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:15.450 17:31:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:15.450 17:31:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:15.450 17:31:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:15.450 17:31:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:15.716 17:31:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:15.716 "name": "Existed_Raid", 00:10:15.716 "uuid": "9caeee85-2c90-4e9e-9c6f-21b72604993a", 00:10:15.716 "strip_size_kb": 64, 00:10:15.716 "state": "online", 00:10:15.716 "raid_level": "raid0", 00:10:15.716 "superblock": true, 00:10:15.716 "num_base_bdevs": 4, 00:10:15.716 "num_base_bdevs_discovered": 4, 00:10:15.716 "num_base_bdevs_operational": 4, 00:10:15.716 "base_bdevs_list": [ 00:10:15.716 { 00:10:15.716 "name": "NewBaseBdev", 00:10:15.716 "uuid": "fcdc3321-c813-42cb-bcda-07f061c745c7", 00:10:15.716 "is_configured": true, 00:10:15.716 "data_offset": 2048, 00:10:15.716 "data_size": 63488 00:10:15.716 }, 00:10:15.716 { 00:10:15.716 "name": "BaseBdev2", 00:10:15.716 "uuid": "9b157ea2-17ce-4a23-a7fa-f9310170f402", 00:10:15.716 "is_configured": true, 00:10:15.716 "data_offset": 2048, 00:10:15.716 "data_size": 63488 00:10:15.716 }, 00:10:15.716 { 00:10:15.716 "name": "BaseBdev3", 00:10:15.716 "uuid": "870d6740-7f69-41d7-9e66-def1fd7cfbf9", 00:10:15.716 "is_configured": true, 00:10:15.716 "data_offset": 2048, 00:10:15.716 "data_size": 63488 00:10:15.716 }, 00:10:15.716 { 00:10:15.716 "name": "BaseBdev4", 00:10:15.716 "uuid": "9daa1097-6cca-4a82-a047-0394016ca5d4", 00:10:15.716 "is_configured": true, 00:10:15.716 "data_offset": 2048, 00:10:15.716 "data_size": 63488 00:10:15.716 } 00:10:15.716 ] 00:10:15.716 }' 00:10:15.716 17:31:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:15.716 17:31:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:16.002 17:31:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:10:16.002 17:31:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:10:16.002 17:31:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:10:16.002 17:31:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:10:16.002 17:31:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:10:16.002 17:31:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:10:16.002 17:31:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:10:16.002 17:31:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:10:16.002 17:31:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:16.002 17:31:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:16.002 [2024-11-27 17:31:47.046287] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:16.002 17:31:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:16.002 17:31:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:16.002 "name": "Existed_Raid", 00:10:16.002 "aliases": [ 00:10:16.002 "9caeee85-2c90-4e9e-9c6f-21b72604993a" 00:10:16.002 ], 00:10:16.002 "product_name": "Raid Volume", 00:10:16.002 "block_size": 512, 00:10:16.002 "num_blocks": 253952, 00:10:16.002 "uuid": "9caeee85-2c90-4e9e-9c6f-21b72604993a", 00:10:16.002 "assigned_rate_limits": { 00:10:16.002 "rw_ios_per_sec": 0, 00:10:16.002 "rw_mbytes_per_sec": 0, 00:10:16.002 "r_mbytes_per_sec": 0, 00:10:16.002 "w_mbytes_per_sec": 0 00:10:16.002 }, 00:10:16.002 "claimed": false, 00:10:16.002 "zoned": false, 00:10:16.002 "supported_io_types": { 00:10:16.002 "read": true, 00:10:16.002 "write": true, 00:10:16.002 "unmap": true, 00:10:16.002 "flush": true, 00:10:16.002 "reset": true, 00:10:16.002 "nvme_admin": false, 00:10:16.002 "nvme_io": false, 00:10:16.002 "nvme_io_md": false, 00:10:16.002 "write_zeroes": true, 00:10:16.002 "zcopy": false, 00:10:16.002 "get_zone_info": false, 00:10:16.002 "zone_management": false, 00:10:16.002 "zone_append": false, 00:10:16.002 "compare": false, 00:10:16.002 "compare_and_write": false, 00:10:16.002 "abort": false, 00:10:16.002 "seek_hole": false, 00:10:16.002 "seek_data": false, 00:10:16.002 "copy": false, 00:10:16.002 "nvme_iov_md": false 00:10:16.002 }, 00:10:16.002 "memory_domains": [ 00:10:16.002 { 00:10:16.002 "dma_device_id": "system", 00:10:16.002 "dma_device_type": 1 00:10:16.002 }, 00:10:16.002 { 00:10:16.002 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:16.002 "dma_device_type": 2 00:10:16.002 }, 00:10:16.002 { 00:10:16.002 "dma_device_id": "system", 00:10:16.002 "dma_device_type": 1 00:10:16.002 }, 00:10:16.002 { 00:10:16.002 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:16.002 "dma_device_type": 2 00:10:16.002 }, 00:10:16.002 { 00:10:16.002 "dma_device_id": "system", 00:10:16.002 "dma_device_type": 1 00:10:16.002 }, 00:10:16.002 { 00:10:16.002 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:16.002 "dma_device_type": 2 00:10:16.002 }, 00:10:16.002 { 00:10:16.002 "dma_device_id": "system", 00:10:16.002 "dma_device_type": 1 00:10:16.002 }, 00:10:16.002 { 00:10:16.002 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:16.002 "dma_device_type": 2 00:10:16.002 } 00:10:16.002 ], 00:10:16.002 "driver_specific": { 00:10:16.002 "raid": { 00:10:16.002 "uuid": "9caeee85-2c90-4e9e-9c6f-21b72604993a", 00:10:16.002 "strip_size_kb": 64, 00:10:16.002 "state": "online", 00:10:16.002 "raid_level": "raid0", 00:10:16.002 "superblock": true, 00:10:16.002 "num_base_bdevs": 4, 00:10:16.002 "num_base_bdevs_discovered": 4, 00:10:16.002 "num_base_bdevs_operational": 4, 00:10:16.002 "base_bdevs_list": [ 00:10:16.002 { 00:10:16.002 "name": "NewBaseBdev", 00:10:16.002 "uuid": "fcdc3321-c813-42cb-bcda-07f061c745c7", 00:10:16.002 "is_configured": true, 00:10:16.002 "data_offset": 2048, 00:10:16.002 "data_size": 63488 00:10:16.002 }, 00:10:16.002 { 00:10:16.002 "name": "BaseBdev2", 00:10:16.002 "uuid": "9b157ea2-17ce-4a23-a7fa-f9310170f402", 00:10:16.002 "is_configured": true, 00:10:16.002 "data_offset": 2048, 00:10:16.002 "data_size": 63488 00:10:16.002 }, 00:10:16.002 { 00:10:16.002 "name": "BaseBdev3", 00:10:16.002 "uuid": "870d6740-7f69-41d7-9e66-def1fd7cfbf9", 00:10:16.002 "is_configured": true, 00:10:16.002 "data_offset": 2048, 00:10:16.002 "data_size": 63488 00:10:16.002 }, 00:10:16.002 { 00:10:16.002 "name": "BaseBdev4", 00:10:16.002 "uuid": "9daa1097-6cca-4a82-a047-0394016ca5d4", 00:10:16.002 "is_configured": true, 00:10:16.002 "data_offset": 2048, 00:10:16.002 "data_size": 63488 00:10:16.002 } 00:10:16.002 ] 00:10:16.002 } 00:10:16.002 } 00:10:16.002 }' 00:10:16.002 17:31:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:16.002 17:31:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:10:16.002 BaseBdev2 00:10:16.002 BaseBdev3 00:10:16.002 BaseBdev4' 00:10:16.002 17:31:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:16.002 17:31:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:16.002 17:31:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:16.002 17:31:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:16.002 17:31:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:10:16.002 17:31:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:16.002 17:31:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:16.276 17:31:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:16.276 17:31:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:16.276 17:31:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:16.276 17:31:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:16.276 17:31:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:16.276 17:31:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:10:16.276 17:31:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:16.276 17:31:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:16.276 17:31:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:16.276 17:31:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:16.276 17:31:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:16.276 17:31:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:16.276 17:31:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:16.276 17:31:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:10:16.276 17:31:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:16.276 17:31:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:16.276 17:31:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:16.276 17:31:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:16.276 17:31:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:16.276 17:31:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:16.276 17:31:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:16.276 17:31:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:10:16.276 17:31:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:16.276 17:31:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:16.276 17:31:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:16.276 17:31:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:16.276 17:31:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:16.276 17:31:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:10:16.276 17:31:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:16.276 17:31:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:16.276 [2024-11-27 17:31:47.353405] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:10:16.276 [2024-11-27 17:31:47.353436] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:16.276 [2024-11-27 17:31:47.353515] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:16.276 [2024-11-27 17:31:47.353588] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:16.276 [2024-11-27 17:31:47.353606] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name Existed_Raid, state offline 00:10:16.276 17:31:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:16.276 17:31:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 80850 00:10:16.276 17:31:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@950 -- # '[' -z 80850 ']' 00:10:16.276 17:31:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # kill -0 80850 00:10:16.276 17:31:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@955 -- # uname 00:10:16.276 17:31:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:10:16.276 17:31:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 80850 00:10:16.276 killing process with pid 80850 00:10:16.276 17:31:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:10:16.276 17:31:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:10:16.276 17:31:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@968 -- # echo 'killing process with pid 80850' 00:10:16.276 17:31:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@969 -- # kill 80850 00:10:16.276 [2024-11-27 17:31:47.403167] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:10:16.276 17:31:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@974 -- # wait 80850 00:10:16.536 [2024-11-27 17:31:47.482129] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:10:16.796 17:31:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:10:16.796 00:10:16.796 real 0m9.881s 00:10:16.796 user 0m16.592s 00:10:16.796 sys 0m2.091s 00:10:16.796 17:31:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1126 -- # xtrace_disable 00:10:16.796 17:31:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:16.796 ************************************ 00:10:16.796 END TEST raid_state_function_test_sb 00:10:16.796 ************************************ 00:10:16.796 17:31:47 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test raid0 4 00:10:16.796 17:31:47 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:10:16.796 17:31:47 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:10:16.796 17:31:47 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:10:16.796 ************************************ 00:10:16.796 START TEST raid_superblock_test 00:10:16.796 ************************************ 00:10:16.796 17:31:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1125 -- # raid_superblock_test raid0 4 00:10:16.796 17:31:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=raid0 00:10:16.796 17:31:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=4 00:10:16.796 17:31:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:10:16.796 17:31:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:10:16.796 17:31:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:10:16.796 17:31:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:10:16.796 17:31:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:10:16.796 17:31:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:10:16.796 17:31:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:10:16.796 17:31:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:10:16.796 17:31:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:10:16.796 17:31:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:10:16.796 17:31:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:10:16.796 17:31:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' raid0 '!=' raid1 ']' 00:10:16.796 17:31:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@405 -- # strip_size=64 00:10:16.796 17:31:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@406 -- # strip_size_create_arg='-z 64' 00:10:16.796 17:31:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=81510 00:10:16.796 17:31:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:10:16.796 17:31:47 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 81510 00:10:16.796 17:31:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@831 -- # '[' -z 81510 ']' 00:10:16.796 17:31:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:16.796 17:31:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:10:16.796 17:31:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:16.796 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:16.796 17:31:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:10:16.796 17:31:47 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:17.056 [2024-11-27 17:31:48.022743] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:10:17.056 [2024-11-27 17:31:48.022980] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid81510 ] 00:10:17.056 [2024-11-27 17:31:48.166633] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:17.056 [2024-11-27 17:31:48.234924] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:10:17.316 [2024-11-27 17:31:48.310811] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:17.316 [2024-11-27 17:31:48.310970] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:17.886 17:31:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:10:17.886 17:31:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@864 -- # return 0 00:10:17.886 17:31:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:10:17.886 17:31:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:10:17.886 17:31:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:10:17.886 17:31:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:10:17.886 17:31:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:10:17.886 17:31:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:10:17.886 17:31:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:10:17.886 17:31:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:10:17.886 17:31:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:10:17.886 17:31:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:17.886 17:31:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:17.886 malloc1 00:10:17.886 17:31:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:17.886 17:31:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:10:17.886 17:31:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:17.886 17:31:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:17.886 [2024-11-27 17:31:48.869076] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:10:17.886 [2024-11-27 17:31:48.869182] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:17.886 [2024-11-27 17:31:48.869210] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:10:17.886 [2024-11-27 17:31:48.869233] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:17.886 [2024-11-27 17:31:48.871703] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:17.886 [2024-11-27 17:31:48.871794] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:10:17.886 pt1 00:10:17.886 17:31:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:17.886 17:31:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:10:17.886 17:31:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:10:17.886 17:31:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:10:17.886 17:31:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:10:17.886 17:31:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:10:17.886 17:31:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:10:17.886 17:31:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:10:17.886 17:31:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:10:17.886 17:31:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:10:17.886 17:31:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:17.886 17:31:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:17.886 malloc2 00:10:17.886 17:31:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:17.886 17:31:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:10:17.886 17:31:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:17.886 17:31:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:17.886 [2024-11-27 17:31:48.920632] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:10:17.886 [2024-11-27 17:31:48.920835] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:17.886 [2024-11-27 17:31:48.920882] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:10:17.886 [2024-11-27 17:31:48.920908] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:17.886 [2024-11-27 17:31:48.925263] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:17.886 [2024-11-27 17:31:48.925316] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:10:17.887 pt2 00:10:17.887 17:31:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:17.887 17:31:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:10:17.887 17:31:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:10:17.887 17:31:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc3 00:10:17.887 17:31:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt3 00:10:17.887 17:31:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000003 00:10:17.887 17:31:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:10:17.887 17:31:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:10:17.887 17:31:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:10:17.887 17:31:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc3 00:10:17.887 17:31:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:17.887 17:31:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:17.887 malloc3 00:10:17.887 17:31:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:17.887 17:31:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:10:17.887 17:31:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:17.887 17:31:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:17.887 [2024-11-27 17:31:48.956120] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:10:17.887 [2024-11-27 17:31:48.956206] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:17.887 [2024-11-27 17:31:48.956225] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:10:17.887 [2024-11-27 17:31:48.956237] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:17.887 [2024-11-27 17:31:48.958524] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:17.887 [2024-11-27 17:31:48.958603] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:10:17.887 pt3 00:10:17.887 17:31:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:17.887 17:31:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:10:17.887 17:31:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:10:17.887 17:31:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc4 00:10:17.887 17:31:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt4 00:10:17.887 17:31:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000004 00:10:17.887 17:31:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:10:17.887 17:31:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:10:17.887 17:31:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:10:17.887 17:31:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc4 00:10:17.887 17:31:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:17.887 17:31:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:17.887 malloc4 00:10:17.887 17:31:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:17.887 17:31:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:10:17.887 17:31:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:17.887 17:31:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:17.887 [2024-11-27 17:31:48.990534] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:10:17.887 [2024-11-27 17:31:48.990587] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:17.887 [2024-11-27 17:31:48.990620] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:10:17.887 [2024-11-27 17:31:48.990634] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:17.887 [2024-11-27 17:31:48.993105] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:17.887 [2024-11-27 17:31:48.993215] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:10:17.887 pt4 00:10:17.887 17:31:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:17.887 17:31:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:10:17.887 17:31:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:10:17.887 17:31:48 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''pt1 pt2 pt3 pt4'\''' -n raid_bdev1 -s 00:10:17.887 17:31:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:17.887 17:31:48 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:17.887 [2024-11-27 17:31:49.002568] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:10:17.887 [2024-11-27 17:31:49.004667] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:10:17.887 [2024-11-27 17:31:49.004785] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:10:17.887 [2024-11-27 17:31:49.004834] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:10:17.887 [2024-11-27 17:31:49.004987] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:10:17.887 [2024-11-27 17:31:49.005001] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:10:17.887 [2024-11-27 17:31:49.005286] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:10:17.887 [2024-11-27 17:31:49.005433] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:10:17.887 [2024-11-27 17:31:49.005443] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:10:17.887 [2024-11-27 17:31:49.005566] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:17.887 17:31:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:17.887 17:31:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 4 00:10:17.887 17:31:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:17.887 17:31:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:17.887 17:31:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:17.887 17:31:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:17.887 17:31:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:17.887 17:31:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:17.887 17:31:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:17.887 17:31:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:17.887 17:31:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:17.887 17:31:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:17.887 17:31:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:17.887 17:31:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:17.887 17:31:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:17.887 17:31:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:17.887 17:31:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:17.887 "name": "raid_bdev1", 00:10:17.887 "uuid": "1cd0032f-8f5a-4107-80ea-00ee92c58b13", 00:10:17.887 "strip_size_kb": 64, 00:10:17.887 "state": "online", 00:10:17.887 "raid_level": "raid0", 00:10:17.887 "superblock": true, 00:10:17.887 "num_base_bdevs": 4, 00:10:17.887 "num_base_bdevs_discovered": 4, 00:10:17.887 "num_base_bdevs_operational": 4, 00:10:17.887 "base_bdevs_list": [ 00:10:17.887 { 00:10:17.887 "name": "pt1", 00:10:17.887 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:17.887 "is_configured": true, 00:10:17.887 "data_offset": 2048, 00:10:17.887 "data_size": 63488 00:10:17.887 }, 00:10:17.887 { 00:10:17.887 "name": "pt2", 00:10:17.887 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:17.887 "is_configured": true, 00:10:17.887 "data_offset": 2048, 00:10:17.887 "data_size": 63488 00:10:17.887 }, 00:10:17.887 { 00:10:17.887 "name": "pt3", 00:10:17.887 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:17.887 "is_configured": true, 00:10:17.887 "data_offset": 2048, 00:10:17.887 "data_size": 63488 00:10:17.887 }, 00:10:17.887 { 00:10:17.887 "name": "pt4", 00:10:17.887 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:17.887 "is_configured": true, 00:10:17.887 "data_offset": 2048, 00:10:17.887 "data_size": 63488 00:10:17.887 } 00:10:17.887 ] 00:10:17.887 }' 00:10:17.887 17:31:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:17.887 17:31:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:18.456 17:31:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:10:18.456 17:31:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:10:18.456 17:31:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:10:18.456 17:31:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:10:18.456 17:31:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:10:18.456 17:31:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:10:18.456 17:31:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:10:18.456 17:31:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:10:18.456 17:31:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:18.456 17:31:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:18.456 [2024-11-27 17:31:49.478031] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:18.456 17:31:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:18.456 17:31:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:18.456 "name": "raid_bdev1", 00:10:18.456 "aliases": [ 00:10:18.456 "1cd0032f-8f5a-4107-80ea-00ee92c58b13" 00:10:18.456 ], 00:10:18.456 "product_name": "Raid Volume", 00:10:18.456 "block_size": 512, 00:10:18.456 "num_blocks": 253952, 00:10:18.456 "uuid": "1cd0032f-8f5a-4107-80ea-00ee92c58b13", 00:10:18.456 "assigned_rate_limits": { 00:10:18.456 "rw_ios_per_sec": 0, 00:10:18.456 "rw_mbytes_per_sec": 0, 00:10:18.456 "r_mbytes_per_sec": 0, 00:10:18.456 "w_mbytes_per_sec": 0 00:10:18.456 }, 00:10:18.456 "claimed": false, 00:10:18.456 "zoned": false, 00:10:18.456 "supported_io_types": { 00:10:18.456 "read": true, 00:10:18.456 "write": true, 00:10:18.456 "unmap": true, 00:10:18.456 "flush": true, 00:10:18.456 "reset": true, 00:10:18.456 "nvme_admin": false, 00:10:18.456 "nvme_io": false, 00:10:18.456 "nvme_io_md": false, 00:10:18.456 "write_zeroes": true, 00:10:18.456 "zcopy": false, 00:10:18.456 "get_zone_info": false, 00:10:18.456 "zone_management": false, 00:10:18.456 "zone_append": false, 00:10:18.456 "compare": false, 00:10:18.456 "compare_and_write": false, 00:10:18.456 "abort": false, 00:10:18.456 "seek_hole": false, 00:10:18.456 "seek_data": false, 00:10:18.456 "copy": false, 00:10:18.456 "nvme_iov_md": false 00:10:18.456 }, 00:10:18.456 "memory_domains": [ 00:10:18.456 { 00:10:18.456 "dma_device_id": "system", 00:10:18.456 "dma_device_type": 1 00:10:18.456 }, 00:10:18.456 { 00:10:18.456 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:18.456 "dma_device_type": 2 00:10:18.456 }, 00:10:18.456 { 00:10:18.456 "dma_device_id": "system", 00:10:18.456 "dma_device_type": 1 00:10:18.456 }, 00:10:18.456 { 00:10:18.456 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:18.456 "dma_device_type": 2 00:10:18.456 }, 00:10:18.456 { 00:10:18.456 "dma_device_id": "system", 00:10:18.456 "dma_device_type": 1 00:10:18.456 }, 00:10:18.456 { 00:10:18.456 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:18.456 "dma_device_type": 2 00:10:18.456 }, 00:10:18.456 { 00:10:18.456 "dma_device_id": "system", 00:10:18.456 "dma_device_type": 1 00:10:18.456 }, 00:10:18.456 { 00:10:18.456 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:18.456 "dma_device_type": 2 00:10:18.456 } 00:10:18.456 ], 00:10:18.456 "driver_specific": { 00:10:18.456 "raid": { 00:10:18.456 "uuid": "1cd0032f-8f5a-4107-80ea-00ee92c58b13", 00:10:18.456 "strip_size_kb": 64, 00:10:18.456 "state": "online", 00:10:18.456 "raid_level": "raid0", 00:10:18.456 "superblock": true, 00:10:18.456 "num_base_bdevs": 4, 00:10:18.456 "num_base_bdevs_discovered": 4, 00:10:18.456 "num_base_bdevs_operational": 4, 00:10:18.456 "base_bdevs_list": [ 00:10:18.456 { 00:10:18.456 "name": "pt1", 00:10:18.456 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:18.456 "is_configured": true, 00:10:18.456 "data_offset": 2048, 00:10:18.456 "data_size": 63488 00:10:18.456 }, 00:10:18.456 { 00:10:18.456 "name": "pt2", 00:10:18.456 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:18.456 "is_configured": true, 00:10:18.456 "data_offset": 2048, 00:10:18.456 "data_size": 63488 00:10:18.456 }, 00:10:18.456 { 00:10:18.456 "name": "pt3", 00:10:18.456 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:18.456 "is_configured": true, 00:10:18.456 "data_offset": 2048, 00:10:18.456 "data_size": 63488 00:10:18.456 }, 00:10:18.456 { 00:10:18.456 "name": "pt4", 00:10:18.456 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:18.456 "is_configured": true, 00:10:18.456 "data_offset": 2048, 00:10:18.456 "data_size": 63488 00:10:18.456 } 00:10:18.456 ] 00:10:18.456 } 00:10:18.456 } 00:10:18.456 }' 00:10:18.456 17:31:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:18.456 17:31:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:10:18.456 pt2 00:10:18.456 pt3 00:10:18.456 pt4' 00:10:18.456 17:31:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:18.456 17:31:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:18.456 17:31:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:18.457 17:31:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:18.457 17:31:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:10:18.457 17:31:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:18.457 17:31:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:18.457 17:31:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:18.717 17:31:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:18.717 17:31:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:18.717 17:31:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:18.717 17:31:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:10:18.717 17:31:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:18.717 17:31:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:18.717 17:31:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:18.717 17:31:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:18.717 17:31:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:18.717 17:31:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:18.717 17:31:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:18.717 17:31:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:10:18.717 17:31:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:18.717 17:31:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:18.717 17:31:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:18.717 17:31:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:18.717 17:31:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:18.717 17:31:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:18.717 17:31:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:18.717 17:31:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:18.717 17:31:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt4 00:10:18.717 17:31:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:18.717 17:31:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:18.717 17:31:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:18.717 17:31:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:18.717 17:31:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:18.717 17:31:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:10:18.717 17:31:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:10:18.717 17:31:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:18.717 17:31:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:18.717 [2024-11-27 17:31:49.777448] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:18.717 17:31:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:18.717 17:31:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=1cd0032f-8f5a-4107-80ea-00ee92c58b13 00:10:18.717 17:31:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z 1cd0032f-8f5a-4107-80ea-00ee92c58b13 ']' 00:10:18.717 17:31:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:10:18.717 17:31:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:18.717 17:31:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:18.717 [2024-11-27 17:31:49.825087] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:10:18.717 [2024-11-27 17:31:49.825119] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:18.717 [2024-11-27 17:31:49.825223] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:18.717 [2024-11-27 17:31:49.825319] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:18.717 [2024-11-27 17:31:49.825337] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:10:18.717 17:31:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:18.717 17:31:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:18.717 17:31:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:18.717 17:31:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:10:18.717 17:31:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:18.717 17:31:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:18.717 17:31:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:10:18.717 17:31:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:10:18.717 17:31:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:10:18.717 17:31:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:10:18.717 17:31:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:18.717 17:31:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:18.717 17:31:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:18.717 17:31:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:10:18.717 17:31:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:10:18.717 17:31:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:18.717 17:31:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:18.717 17:31:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:18.717 17:31:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:10:18.717 17:31:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt3 00:10:18.717 17:31:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:18.717 17:31:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:18.977 17:31:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:18.977 17:31:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:10:18.977 17:31:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt4 00:10:18.977 17:31:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:18.977 17:31:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:18.977 17:31:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:18.977 17:31:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:10:18.977 17:31:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:10:18.977 17:31:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:18.977 17:31:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:18.977 17:31:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:18.977 17:31:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:10:18.978 17:31:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:10:18.978 17:31:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@650 -- # local es=0 00:10:18.978 17:31:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:10:18.978 17:31:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:10:18.978 17:31:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:10:18.978 17:31:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:10:18.978 17:31:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:10:18.978 17:31:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:10:18.978 17:31:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:18.978 17:31:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:18.978 [2024-11-27 17:31:49.984836] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:10:18.978 [2024-11-27 17:31:49.987067] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:10:18.978 [2024-11-27 17:31:49.987178] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc3 is claimed 00:10:18.978 [2024-11-27 17:31:49.987237] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc4 is claimed 00:10:18.978 [2024-11-27 17:31:49.987317] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:10:18.978 [2024-11-27 17:31:49.987400] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:10:18.978 [2024-11-27 17:31:49.987453] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc3 00:10:18.978 [2024-11-27 17:31:49.987499] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc4 00:10:18.978 [2024-11-27 17:31:49.987552] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:10:18.978 [2024-11-27 17:31:49.987590] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state configuring 00:10:18.978 request: 00:10:18.978 { 00:10:18.978 "name": "raid_bdev1", 00:10:18.978 "raid_level": "raid0", 00:10:18.978 "base_bdevs": [ 00:10:18.978 "malloc1", 00:10:18.978 "malloc2", 00:10:18.978 "malloc3", 00:10:18.978 "malloc4" 00:10:18.978 ], 00:10:18.978 "strip_size_kb": 64, 00:10:18.978 "superblock": false, 00:10:18.978 "method": "bdev_raid_create", 00:10:18.978 "req_id": 1 00:10:18.978 } 00:10:18.978 Got JSON-RPC error response 00:10:18.978 response: 00:10:18.978 { 00:10:18.978 "code": -17, 00:10:18.978 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:10:18.978 } 00:10:18.978 17:31:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:10:18.978 17:31:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@653 -- # es=1 00:10:18.978 17:31:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:10:18.978 17:31:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:10:18.978 17:31:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:10:18.978 17:31:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:10:18.978 17:31:49 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:18.978 17:31:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:18.978 17:31:49 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:18.978 17:31:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:18.978 17:31:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:10:18.978 17:31:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:10:18.978 17:31:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:10:18.978 17:31:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:18.978 17:31:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:18.978 [2024-11-27 17:31:50.048686] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:10:18.978 [2024-11-27 17:31:50.048768] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:18.978 [2024-11-27 17:31:50.048830] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:10:18.978 [2024-11-27 17:31:50.048861] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:18.978 [2024-11-27 17:31:50.051414] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:18.978 [2024-11-27 17:31:50.051480] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:10:18.978 [2024-11-27 17:31:50.051595] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:10:18.978 [2024-11-27 17:31:50.051654] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:10:18.978 pt1 00:10:18.978 17:31:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:18.978 17:31:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid0 64 4 00:10:18.978 17:31:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:18.978 17:31:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:18.978 17:31:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:18.978 17:31:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:18.978 17:31:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:18.978 17:31:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:18.978 17:31:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:18.978 17:31:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:18.978 17:31:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:18.978 17:31:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:18.978 17:31:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:18.978 17:31:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:18.978 17:31:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:18.978 17:31:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:18.978 17:31:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:18.978 "name": "raid_bdev1", 00:10:18.978 "uuid": "1cd0032f-8f5a-4107-80ea-00ee92c58b13", 00:10:18.978 "strip_size_kb": 64, 00:10:18.978 "state": "configuring", 00:10:18.978 "raid_level": "raid0", 00:10:18.978 "superblock": true, 00:10:18.978 "num_base_bdevs": 4, 00:10:18.978 "num_base_bdevs_discovered": 1, 00:10:18.978 "num_base_bdevs_operational": 4, 00:10:18.978 "base_bdevs_list": [ 00:10:18.978 { 00:10:18.978 "name": "pt1", 00:10:18.978 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:18.978 "is_configured": true, 00:10:18.978 "data_offset": 2048, 00:10:18.978 "data_size": 63488 00:10:18.978 }, 00:10:18.978 { 00:10:18.978 "name": null, 00:10:18.978 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:18.978 "is_configured": false, 00:10:18.978 "data_offset": 2048, 00:10:18.978 "data_size": 63488 00:10:18.978 }, 00:10:18.978 { 00:10:18.978 "name": null, 00:10:18.978 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:18.978 "is_configured": false, 00:10:18.978 "data_offset": 2048, 00:10:18.978 "data_size": 63488 00:10:18.978 }, 00:10:18.978 { 00:10:18.978 "name": null, 00:10:18.978 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:18.978 "is_configured": false, 00:10:18.978 "data_offset": 2048, 00:10:18.978 "data_size": 63488 00:10:18.978 } 00:10:18.978 ] 00:10:18.978 }' 00:10:18.978 17:31:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:18.978 17:31:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:19.549 17:31:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 4 -gt 2 ']' 00:10:19.549 17:31:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@472 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:10:19.549 17:31:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:19.549 17:31:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:19.549 [2024-11-27 17:31:50.467965] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:10:19.549 [2024-11-27 17:31:50.468015] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:19.549 [2024-11-27 17:31:50.468052] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009c80 00:10:19.549 [2024-11-27 17:31:50.468061] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:19.549 [2024-11-27 17:31:50.468486] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:19.549 [2024-11-27 17:31:50.468515] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:10:19.549 [2024-11-27 17:31:50.468581] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:10:19.549 [2024-11-27 17:31:50.468610] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:10:19.549 pt2 00:10:19.549 17:31:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:19.549 17:31:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@473 -- # rpc_cmd bdev_passthru_delete pt2 00:10:19.549 17:31:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:19.549 17:31:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:19.549 [2024-11-27 17:31:50.475996] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: pt2 00:10:19.549 17:31:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:19.549 17:31:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@474 -- # verify_raid_bdev_state raid_bdev1 configuring raid0 64 4 00:10:19.549 17:31:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:19.549 17:31:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:19.549 17:31:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:19.549 17:31:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:19.549 17:31:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:19.549 17:31:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:19.549 17:31:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:19.549 17:31:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:19.549 17:31:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:19.549 17:31:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:19.549 17:31:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:19.549 17:31:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:19.549 17:31:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:19.549 17:31:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:19.549 17:31:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:19.549 "name": "raid_bdev1", 00:10:19.549 "uuid": "1cd0032f-8f5a-4107-80ea-00ee92c58b13", 00:10:19.549 "strip_size_kb": 64, 00:10:19.549 "state": "configuring", 00:10:19.549 "raid_level": "raid0", 00:10:19.549 "superblock": true, 00:10:19.549 "num_base_bdevs": 4, 00:10:19.549 "num_base_bdevs_discovered": 1, 00:10:19.549 "num_base_bdevs_operational": 4, 00:10:19.549 "base_bdevs_list": [ 00:10:19.549 { 00:10:19.549 "name": "pt1", 00:10:19.549 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:19.549 "is_configured": true, 00:10:19.549 "data_offset": 2048, 00:10:19.549 "data_size": 63488 00:10:19.549 }, 00:10:19.549 { 00:10:19.549 "name": null, 00:10:19.549 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:19.549 "is_configured": false, 00:10:19.549 "data_offset": 0, 00:10:19.549 "data_size": 63488 00:10:19.549 }, 00:10:19.549 { 00:10:19.549 "name": null, 00:10:19.549 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:19.549 "is_configured": false, 00:10:19.549 "data_offset": 2048, 00:10:19.549 "data_size": 63488 00:10:19.549 }, 00:10:19.549 { 00:10:19.549 "name": null, 00:10:19.549 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:19.549 "is_configured": false, 00:10:19.549 "data_offset": 2048, 00:10:19.549 "data_size": 63488 00:10:19.549 } 00:10:19.549 ] 00:10:19.549 }' 00:10:19.549 17:31:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:19.549 17:31:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:19.810 17:31:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:10:19.810 17:31:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:10:19.810 17:31:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:10:19.810 17:31:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:19.810 17:31:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:19.810 [2024-11-27 17:31:50.859316] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:10:19.810 [2024-11-27 17:31:50.859428] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:19.810 [2024-11-27 17:31:50.859462] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009f80 00:10:19.810 [2024-11-27 17:31:50.859492] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:19.810 [2024-11-27 17:31:50.859926] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:19.810 [2024-11-27 17:31:50.859990] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:10:19.810 [2024-11-27 17:31:50.860083] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:10:19.810 [2024-11-27 17:31:50.860136] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:10:19.810 pt2 00:10:19.810 17:31:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:19.810 17:31:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:10:19.810 17:31:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:10:19.810 17:31:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:10:19.810 17:31:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:19.810 17:31:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:19.810 [2024-11-27 17:31:50.871263] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:10:19.810 [2024-11-27 17:31:50.871350] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:19.810 [2024-11-27 17:31:50.871382] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:10:19.810 [2024-11-27 17:31:50.871423] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:19.810 [2024-11-27 17:31:50.871787] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:19.810 [2024-11-27 17:31:50.871848] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:10:19.810 [2024-11-27 17:31:50.871947] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:10:19.810 [2024-11-27 17:31:50.871999] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:10:19.810 pt3 00:10:19.810 17:31:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:19.810 17:31:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:10:19.810 17:31:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:10:19.810 17:31:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:10:19.810 17:31:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:19.810 17:31:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:19.810 [2024-11-27 17:31:50.883274] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:10:19.810 [2024-11-27 17:31:50.883353] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:19.810 [2024-11-27 17:31:50.883398] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a580 00:10:19.810 [2024-11-27 17:31:50.883426] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:19.810 [2024-11-27 17:31:50.883739] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:19.810 [2024-11-27 17:31:50.883798] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:10:19.810 [2024-11-27 17:31:50.883871] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt4 00:10:19.810 [2024-11-27 17:31:50.883914] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:10:19.810 [2024-11-27 17:31:50.884027] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:10:19.810 [2024-11-27 17:31:50.884066] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:10:19.810 [2024-11-27 17:31:50.884352] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:10:19.810 [2024-11-27 17:31:50.884503] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:10:19.810 [2024-11-27 17:31:50.884538] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:10:19.810 [2024-11-27 17:31:50.884640] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:19.810 pt4 00:10:19.810 17:31:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:19.810 17:31:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:10:19.810 17:31:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:10:19.810 17:31:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 4 00:10:19.810 17:31:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:19.810 17:31:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:19.810 17:31:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:19.810 17:31:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:19.810 17:31:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:19.810 17:31:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:19.810 17:31:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:19.810 17:31:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:19.810 17:31:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:19.810 17:31:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:19.810 17:31:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:19.810 17:31:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:19.810 17:31:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:19.810 17:31:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:19.810 17:31:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:19.810 "name": "raid_bdev1", 00:10:19.810 "uuid": "1cd0032f-8f5a-4107-80ea-00ee92c58b13", 00:10:19.810 "strip_size_kb": 64, 00:10:19.810 "state": "online", 00:10:19.810 "raid_level": "raid0", 00:10:19.810 "superblock": true, 00:10:19.810 "num_base_bdevs": 4, 00:10:19.810 "num_base_bdevs_discovered": 4, 00:10:19.810 "num_base_bdevs_operational": 4, 00:10:19.810 "base_bdevs_list": [ 00:10:19.810 { 00:10:19.810 "name": "pt1", 00:10:19.810 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:19.810 "is_configured": true, 00:10:19.810 "data_offset": 2048, 00:10:19.810 "data_size": 63488 00:10:19.810 }, 00:10:19.810 { 00:10:19.810 "name": "pt2", 00:10:19.810 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:19.810 "is_configured": true, 00:10:19.810 "data_offset": 2048, 00:10:19.810 "data_size": 63488 00:10:19.810 }, 00:10:19.810 { 00:10:19.810 "name": "pt3", 00:10:19.810 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:19.810 "is_configured": true, 00:10:19.810 "data_offset": 2048, 00:10:19.810 "data_size": 63488 00:10:19.810 }, 00:10:19.810 { 00:10:19.810 "name": "pt4", 00:10:19.810 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:19.810 "is_configured": true, 00:10:19.810 "data_offset": 2048, 00:10:19.810 "data_size": 63488 00:10:19.810 } 00:10:19.810 ] 00:10:19.810 }' 00:10:19.810 17:31:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:19.810 17:31:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:20.381 17:31:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:10:20.381 17:31:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:10:20.381 17:31:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:10:20.381 17:31:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:10:20.381 17:31:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:10:20.381 17:31:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:10:20.381 17:31:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:10:20.381 17:31:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:10:20.381 17:31:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:20.381 17:31:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:20.381 [2024-11-27 17:31:51.310930] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:20.381 17:31:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:20.381 17:31:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:20.381 "name": "raid_bdev1", 00:10:20.381 "aliases": [ 00:10:20.381 "1cd0032f-8f5a-4107-80ea-00ee92c58b13" 00:10:20.381 ], 00:10:20.381 "product_name": "Raid Volume", 00:10:20.381 "block_size": 512, 00:10:20.381 "num_blocks": 253952, 00:10:20.381 "uuid": "1cd0032f-8f5a-4107-80ea-00ee92c58b13", 00:10:20.381 "assigned_rate_limits": { 00:10:20.381 "rw_ios_per_sec": 0, 00:10:20.381 "rw_mbytes_per_sec": 0, 00:10:20.381 "r_mbytes_per_sec": 0, 00:10:20.381 "w_mbytes_per_sec": 0 00:10:20.381 }, 00:10:20.381 "claimed": false, 00:10:20.381 "zoned": false, 00:10:20.381 "supported_io_types": { 00:10:20.381 "read": true, 00:10:20.381 "write": true, 00:10:20.381 "unmap": true, 00:10:20.381 "flush": true, 00:10:20.381 "reset": true, 00:10:20.381 "nvme_admin": false, 00:10:20.381 "nvme_io": false, 00:10:20.381 "nvme_io_md": false, 00:10:20.381 "write_zeroes": true, 00:10:20.381 "zcopy": false, 00:10:20.381 "get_zone_info": false, 00:10:20.381 "zone_management": false, 00:10:20.381 "zone_append": false, 00:10:20.381 "compare": false, 00:10:20.381 "compare_and_write": false, 00:10:20.381 "abort": false, 00:10:20.381 "seek_hole": false, 00:10:20.381 "seek_data": false, 00:10:20.381 "copy": false, 00:10:20.381 "nvme_iov_md": false 00:10:20.381 }, 00:10:20.381 "memory_domains": [ 00:10:20.381 { 00:10:20.381 "dma_device_id": "system", 00:10:20.381 "dma_device_type": 1 00:10:20.381 }, 00:10:20.381 { 00:10:20.381 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:20.381 "dma_device_type": 2 00:10:20.381 }, 00:10:20.381 { 00:10:20.381 "dma_device_id": "system", 00:10:20.381 "dma_device_type": 1 00:10:20.381 }, 00:10:20.381 { 00:10:20.381 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:20.381 "dma_device_type": 2 00:10:20.381 }, 00:10:20.381 { 00:10:20.381 "dma_device_id": "system", 00:10:20.381 "dma_device_type": 1 00:10:20.381 }, 00:10:20.381 { 00:10:20.381 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:20.381 "dma_device_type": 2 00:10:20.381 }, 00:10:20.381 { 00:10:20.382 "dma_device_id": "system", 00:10:20.382 "dma_device_type": 1 00:10:20.382 }, 00:10:20.382 { 00:10:20.382 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:20.382 "dma_device_type": 2 00:10:20.382 } 00:10:20.382 ], 00:10:20.382 "driver_specific": { 00:10:20.382 "raid": { 00:10:20.382 "uuid": "1cd0032f-8f5a-4107-80ea-00ee92c58b13", 00:10:20.382 "strip_size_kb": 64, 00:10:20.382 "state": "online", 00:10:20.382 "raid_level": "raid0", 00:10:20.382 "superblock": true, 00:10:20.382 "num_base_bdevs": 4, 00:10:20.382 "num_base_bdevs_discovered": 4, 00:10:20.382 "num_base_bdevs_operational": 4, 00:10:20.382 "base_bdevs_list": [ 00:10:20.382 { 00:10:20.382 "name": "pt1", 00:10:20.382 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:20.382 "is_configured": true, 00:10:20.382 "data_offset": 2048, 00:10:20.382 "data_size": 63488 00:10:20.382 }, 00:10:20.382 { 00:10:20.382 "name": "pt2", 00:10:20.382 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:20.382 "is_configured": true, 00:10:20.382 "data_offset": 2048, 00:10:20.382 "data_size": 63488 00:10:20.382 }, 00:10:20.382 { 00:10:20.382 "name": "pt3", 00:10:20.382 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:20.382 "is_configured": true, 00:10:20.382 "data_offset": 2048, 00:10:20.382 "data_size": 63488 00:10:20.382 }, 00:10:20.382 { 00:10:20.382 "name": "pt4", 00:10:20.382 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:20.382 "is_configured": true, 00:10:20.382 "data_offset": 2048, 00:10:20.382 "data_size": 63488 00:10:20.382 } 00:10:20.382 ] 00:10:20.382 } 00:10:20.382 } 00:10:20.382 }' 00:10:20.382 17:31:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:20.382 17:31:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:10:20.382 pt2 00:10:20.382 pt3 00:10:20.382 pt4' 00:10:20.382 17:31:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:20.382 17:31:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:20.382 17:31:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:20.382 17:31:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:20.382 17:31:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:10:20.382 17:31:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:20.382 17:31:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:20.382 17:31:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:20.382 17:31:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:20.382 17:31:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:20.382 17:31:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:20.382 17:31:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:20.382 17:31:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:10:20.382 17:31:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:20.382 17:31:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:20.382 17:31:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:20.382 17:31:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:20.382 17:31:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:20.382 17:31:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:20.382 17:31:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:10:20.382 17:31:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:20.382 17:31:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:20.382 17:31:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:20.382 17:31:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:20.382 17:31:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:20.382 17:31:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:20.382 17:31:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:20.642 17:31:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:20.642 17:31:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt4 00:10:20.642 17:31:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:20.642 17:31:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:20.642 17:31:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:20.642 17:31:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:20.642 17:31:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:20.642 17:31:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:10:20.642 17:31:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:20.642 17:31:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:10:20.642 17:31:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:20.642 [2024-11-27 17:31:51.622337] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:20.642 17:31:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:20.642 17:31:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' 1cd0032f-8f5a-4107-80ea-00ee92c58b13 '!=' 1cd0032f-8f5a-4107-80ea-00ee92c58b13 ']' 00:10:20.642 17:31:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy raid0 00:10:20.642 17:31:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:10:20.642 17:31:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # return 1 00:10:20.642 17:31:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 81510 00:10:20.642 17:31:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@950 -- # '[' -z 81510 ']' 00:10:20.642 17:31:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # kill -0 81510 00:10:20.642 17:31:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@955 -- # uname 00:10:20.642 17:31:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:10:20.643 17:31:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 81510 00:10:20.643 killing process with pid 81510 00:10:20.643 17:31:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:10:20.643 17:31:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:10:20.643 17:31:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 81510' 00:10:20.643 17:31:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@969 -- # kill 81510 00:10:20.643 [2024-11-27 17:31:51.697853] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:10:20.643 [2024-11-27 17:31:51.697961] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:20.643 17:31:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@974 -- # wait 81510 00:10:20.643 [2024-11-27 17:31:51.698032] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:20.643 [2024-11-27 17:31:51.698045] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:10:20.643 [2024-11-27 17:31:51.778736] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:10:21.214 ************************************ 00:10:21.214 END TEST raid_superblock_test 00:10:21.214 ************************************ 00:10:21.214 17:31:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:10:21.214 00:10:21.214 real 0m4.213s 00:10:21.214 user 0m6.368s 00:10:21.214 sys 0m1.035s 00:10:21.214 17:31:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:10:21.214 17:31:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:21.214 17:31:52 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test raid0 4 read 00:10:21.214 17:31:52 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:10:21.214 17:31:52 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:10:21.214 17:31:52 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:10:21.214 ************************************ 00:10:21.214 START TEST raid_read_error_test 00:10:21.214 ************************************ 00:10:21.214 17:31:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1125 -- # raid_io_error_test raid0 4 read 00:10:21.214 17:31:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid0 00:10:21.214 17:31:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=4 00:10:21.214 17:31:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:10:21.214 17:31:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:10:21.214 17:31:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:21.214 17:31:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:10:21.214 17:31:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:21.214 17:31:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:21.214 17:31:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:10:21.214 17:31:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:21.214 17:31:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:21.214 17:31:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:10:21.214 17:31:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:21.214 17:31:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:21.214 17:31:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev4 00:10:21.214 17:31:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:21.214 17:31:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:21.214 17:31:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:10:21.214 17:31:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:10:21.214 17:31:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:10:21.214 17:31:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:10:21.214 17:31:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:10:21.214 17:31:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:10:21.214 17:31:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:10:21.214 17:31:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid0 '!=' raid1 ']' 00:10:21.214 17:31:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:10:21.214 17:31:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:10:21.214 17:31:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:10:21.214 17:31:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.I8pzdVHp0N 00:10:21.214 17:31:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=81758 00:10:21.214 17:31:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 81758 00:10:21.214 17:31:52 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:10:21.214 17:31:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@831 -- # '[' -z 81758 ']' 00:10:21.214 17:31:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:21.214 17:31:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:10:21.214 17:31:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:21.214 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:21.214 17:31:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:10:21.214 17:31:52 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:21.214 [2024-11-27 17:31:52.324069] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:10:21.214 [2024-11-27 17:31:52.324310] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid81758 ] 00:10:21.475 [2024-11-27 17:31:52.471241] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:21.475 [2024-11-27 17:31:52.539596] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:10:21.475 [2024-11-27 17:31:52.615494] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:21.475 [2024-11-27 17:31:52.615530] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:22.045 17:31:53 bdev_raid.raid_read_error_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:10:22.045 17:31:53 bdev_raid.raid_read_error_test -- common/autotest_common.sh@864 -- # return 0 00:10:22.045 17:31:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:22.045 17:31:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:10:22.045 17:31:53 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:22.045 17:31:53 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:22.045 BaseBdev1_malloc 00:10:22.045 17:31:53 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:22.045 17:31:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:10:22.045 17:31:53 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:22.045 17:31:53 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:22.045 true 00:10:22.045 17:31:53 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:22.045 17:31:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:10:22.045 17:31:53 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:22.045 17:31:53 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:22.045 [2024-11-27 17:31:53.190180] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:10:22.045 [2024-11-27 17:31:53.190327] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:22.045 [2024-11-27 17:31:53.190384] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:10:22.045 [2024-11-27 17:31:53.190438] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:22.045 [2024-11-27 17:31:53.192906] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:22.045 [2024-11-27 17:31:53.192980] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:10:22.045 BaseBdev1 00:10:22.045 17:31:53 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:22.045 17:31:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:22.045 17:31:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:10:22.045 17:31:53 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:22.045 17:31:53 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:22.045 BaseBdev2_malloc 00:10:22.045 17:31:53 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:22.045 17:31:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:10:22.045 17:31:53 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:22.045 17:31:53 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:22.306 true 00:10:22.306 17:31:53 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:22.306 17:31:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:10:22.306 17:31:53 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:22.306 17:31:53 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:22.306 [2024-11-27 17:31:53.246969] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:10:22.306 [2024-11-27 17:31:53.247122] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:22.306 [2024-11-27 17:31:53.247217] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:10:22.306 [2024-11-27 17:31:53.247269] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:22.306 [2024-11-27 17:31:53.250892] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:22.306 [2024-11-27 17:31:53.250995] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:10:22.306 BaseBdev2 00:10:22.306 17:31:53 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:22.306 17:31:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:22.306 17:31:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:10:22.306 17:31:53 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:22.306 17:31:53 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:22.306 BaseBdev3_malloc 00:10:22.306 17:31:53 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:22.307 17:31:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:10:22.307 17:31:53 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:22.307 17:31:53 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:22.307 true 00:10:22.307 17:31:53 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:22.307 17:31:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:10:22.307 17:31:53 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:22.307 17:31:53 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:22.307 [2024-11-27 17:31:53.293749] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:10:22.307 [2024-11-27 17:31:53.293844] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:22.307 [2024-11-27 17:31:53.293884] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008780 00:10:22.307 [2024-11-27 17:31:53.293894] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:22.307 [2024-11-27 17:31:53.296321] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:22.307 [2024-11-27 17:31:53.296362] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:10:22.307 BaseBdev3 00:10:22.307 17:31:53 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:22.307 17:31:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:22.307 17:31:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:10:22.307 17:31:53 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:22.307 17:31:53 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:22.307 BaseBdev4_malloc 00:10:22.307 17:31:53 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:22.307 17:31:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev4_malloc 00:10:22.307 17:31:53 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:22.307 17:31:53 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:22.307 true 00:10:22.307 17:31:53 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:22.307 17:31:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev4_malloc -p BaseBdev4 00:10:22.307 17:31:53 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:22.307 17:31:53 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:22.307 [2024-11-27 17:31:53.340267] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev4_malloc 00:10:22.307 [2024-11-27 17:31:53.340314] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:22.307 [2024-11-27 17:31:53.340338] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:10:22.307 [2024-11-27 17:31:53.340346] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:22.307 [2024-11-27 17:31:53.342758] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:22.307 [2024-11-27 17:31:53.342828] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:10:22.307 BaseBdev4 00:10:22.307 17:31:53 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:22.307 17:31:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 -s 00:10:22.307 17:31:53 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:22.307 17:31:53 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:22.307 [2024-11-27 17:31:53.352364] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:22.307 [2024-11-27 17:31:53.354410] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:22.307 [2024-11-27 17:31:53.354487] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:22.307 [2024-11-27 17:31:53.354548] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:10:22.307 [2024-11-27 17:31:53.354753] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000002000 00:10:22.307 [2024-11-27 17:31:53.354765] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:10:22.307 [2024-11-27 17:31:53.355032] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:10:22.307 [2024-11-27 17:31:53.355195] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000002000 00:10:22.307 [2024-11-27 17:31:53.355209] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000002000 00:10:22.307 [2024-11-27 17:31:53.355342] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:22.307 17:31:53 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:22.307 17:31:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 4 00:10:22.307 17:31:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:22.307 17:31:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:22.307 17:31:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:22.307 17:31:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:22.307 17:31:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:22.307 17:31:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:22.307 17:31:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:22.307 17:31:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:22.307 17:31:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:22.307 17:31:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:22.307 17:31:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:22.307 17:31:53 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:22.307 17:31:53 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:22.307 17:31:53 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:22.307 17:31:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:22.307 "name": "raid_bdev1", 00:10:22.307 "uuid": "7ddf3e30-d2ec-4ebd-8d64-e5d0a29c82dc", 00:10:22.307 "strip_size_kb": 64, 00:10:22.307 "state": "online", 00:10:22.307 "raid_level": "raid0", 00:10:22.307 "superblock": true, 00:10:22.307 "num_base_bdevs": 4, 00:10:22.307 "num_base_bdevs_discovered": 4, 00:10:22.307 "num_base_bdevs_operational": 4, 00:10:22.307 "base_bdevs_list": [ 00:10:22.307 { 00:10:22.307 "name": "BaseBdev1", 00:10:22.307 "uuid": "b766da74-2b28-5dad-8243-8a0697cd0849", 00:10:22.307 "is_configured": true, 00:10:22.307 "data_offset": 2048, 00:10:22.307 "data_size": 63488 00:10:22.307 }, 00:10:22.307 { 00:10:22.307 "name": "BaseBdev2", 00:10:22.307 "uuid": "aeae383c-7f1f-54f6-b054-adc68af54a19", 00:10:22.307 "is_configured": true, 00:10:22.307 "data_offset": 2048, 00:10:22.307 "data_size": 63488 00:10:22.307 }, 00:10:22.307 { 00:10:22.307 "name": "BaseBdev3", 00:10:22.307 "uuid": "7743f6f5-b150-5ec0-95cc-4f03e21156dd", 00:10:22.307 "is_configured": true, 00:10:22.307 "data_offset": 2048, 00:10:22.307 "data_size": 63488 00:10:22.307 }, 00:10:22.307 { 00:10:22.307 "name": "BaseBdev4", 00:10:22.307 "uuid": "982b7359-852f-5db9-b607-1ef890d84f2e", 00:10:22.307 "is_configured": true, 00:10:22.307 "data_offset": 2048, 00:10:22.307 "data_size": 63488 00:10:22.307 } 00:10:22.307 ] 00:10:22.307 }' 00:10:22.307 17:31:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:22.307 17:31:53 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:22.876 17:31:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:10:22.876 17:31:53 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:10:22.876 [2024-11-27 17:31:53.871992] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000026d0 00:10:23.826 17:31:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:10:23.826 17:31:54 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:23.826 17:31:54 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:23.826 17:31:54 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:23.826 17:31:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:10:23.826 17:31:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid0 = \r\a\i\d\1 ]] 00:10:23.826 17:31:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=4 00:10:23.826 17:31:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 4 00:10:23.826 17:31:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:23.826 17:31:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:23.826 17:31:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:23.826 17:31:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:23.826 17:31:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:23.826 17:31:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:23.826 17:31:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:23.826 17:31:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:23.826 17:31:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:23.826 17:31:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:23.826 17:31:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:23.826 17:31:54 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:23.826 17:31:54 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:23.826 17:31:54 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:23.826 17:31:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:23.826 "name": "raid_bdev1", 00:10:23.826 "uuid": "7ddf3e30-d2ec-4ebd-8d64-e5d0a29c82dc", 00:10:23.826 "strip_size_kb": 64, 00:10:23.826 "state": "online", 00:10:23.826 "raid_level": "raid0", 00:10:23.826 "superblock": true, 00:10:23.826 "num_base_bdevs": 4, 00:10:23.826 "num_base_bdevs_discovered": 4, 00:10:23.826 "num_base_bdevs_operational": 4, 00:10:23.826 "base_bdevs_list": [ 00:10:23.826 { 00:10:23.826 "name": "BaseBdev1", 00:10:23.826 "uuid": "b766da74-2b28-5dad-8243-8a0697cd0849", 00:10:23.826 "is_configured": true, 00:10:23.826 "data_offset": 2048, 00:10:23.826 "data_size": 63488 00:10:23.826 }, 00:10:23.826 { 00:10:23.826 "name": "BaseBdev2", 00:10:23.826 "uuid": "aeae383c-7f1f-54f6-b054-adc68af54a19", 00:10:23.826 "is_configured": true, 00:10:23.826 "data_offset": 2048, 00:10:23.826 "data_size": 63488 00:10:23.826 }, 00:10:23.826 { 00:10:23.826 "name": "BaseBdev3", 00:10:23.826 "uuid": "7743f6f5-b150-5ec0-95cc-4f03e21156dd", 00:10:23.826 "is_configured": true, 00:10:23.826 "data_offset": 2048, 00:10:23.826 "data_size": 63488 00:10:23.826 }, 00:10:23.826 { 00:10:23.826 "name": "BaseBdev4", 00:10:23.826 "uuid": "982b7359-852f-5db9-b607-1ef890d84f2e", 00:10:23.826 "is_configured": true, 00:10:23.826 "data_offset": 2048, 00:10:23.826 "data_size": 63488 00:10:23.826 } 00:10:23.826 ] 00:10:23.826 }' 00:10:23.826 17:31:54 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:23.826 17:31:54 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:24.085 17:31:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:10:24.085 17:31:55 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:24.085 17:31:55 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:24.085 [2024-11-27 17:31:55.268878] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:10:24.085 [2024-11-27 17:31:55.268917] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:24.085 [2024-11-27 17:31:55.271484] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:24.086 [2024-11-27 17:31:55.271582] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:24.086 [2024-11-27 17:31:55.271668] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:24.086 [2024-11-27 17:31:55.271723] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002000 name raid_bdev1, state offline 00:10:24.086 { 00:10:24.086 "results": [ 00:10:24.086 { 00:10:24.086 "job": "raid_bdev1", 00:10:24.086 "core_mask": "0x1", 00:10:24.086 "workload": "randrw", 00:10:24.086 "percentage": 50, 00:10:24.086 "status": "finished", 00:10:24.086 "queue_depth": 1, 00:10:24.086 "io_size": 131072, 00:10:24.086 "runtime": 1.397435, 00:10:24.086 "iops": 14726.98193475904, 00:10:24.086 "mibps": 1840.87274184488, 00:10:24.086 "io_failed": 1, 00:10:24.086 "io_timeout": 0, 00:10:24.086 "avg_latency_us": 95.42163368129634, 00:10:24.086 "min_latency_us": 24.705676855895195, 00:10:24.086 "max_latency_us": 1409.4532751091704 00:10:24.086 } 00:10:24.086 ], 00:10:24.086 "core_count": 1 00:10:24.086 } 00:10:24.086 17:31:55 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:24.086 17:31:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 81758 00:10:24.086 17:31:55 bdev_raid.raid_read_error_test -- common/autotest_common.sh@950 -- # '[' -z 81758 ']' 00:10:24.086 17:31:55 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # kill -0 81758 00:10:24.344 17:31:55 bdev_raid.raid_read_error_test -- common/autotest_common.sh@955 -- # uname 00:10:24.344 17:31:55 bdev_raid.raid_read_error_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:10:24.344 17:31:55 bdev_raid.raid_read_error_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 81758 00:10:24.345 17:31:55 bdev_raid.raid_read_error_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:10:24.345 17:31:55 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:10:24.345 killing process with pid 81758 00:10:24.345 17:31:55 bdev_raid.raid_read_error_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 81758' 00:10:24.345 17:31:55 bdev_raid.raid_read_error_test -- common/autotest_common.sh@969 -- # kill 81758 00:10:24.345 [2024-11-27 17:31:55.318180] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:10:24.345 17:31:55 bdev_raid.raid_read_error_test -- common/autotest_common.sh@974 -- # wait 81758 00:10:24.345 [2024-11-27 17:31:55.386691] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:10:24.604 17:31:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.I8pzdVHp0N 00:10:24.604 17:31:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:10:24.604 17:31:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:10:24.604 17:31:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.72 00:10:24.604 17:31:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid0 00:10:24.604 ************************************ 00:10:24.604 END TEST raid_read_error_test 00:10:24.604 ************************************ 00:10:24.604 17:31:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:10:24.604 17:31:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:10:24.604 17:31:55 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.72 != \0\.\0\0 ]] 00:10:24.604 00:10:24.604 real 0m3.540s 00:10:24.604 user 0m4.290s 00:10:24.604 sys 0m0.667s 00:10:24.604 17:31:55 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:10:24.604 17:31:55 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:24.864 17:31:55 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test raid0 4 write 00:10:24.864 17:31:55 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:10:24.864 17:31:55 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:10:24.864 17:31:55 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:10:24.864 ************************************ 00:10:24.864 START TEST raid_write_error_test 00:10:24.864 ************************************ 00:10:24.864 17:31:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1125 -- # raid_io_error_test raid0 4 write 00:10:24.864 17:31:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid0 00:10:24.864 17:31:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=4 00:10:24.864 17:31:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:10:24.864 17:31:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:10:24.864 17:31:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:24.864 17:31:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:10:24.864 17:31:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:24.864 17:31:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:24.864 17:31:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:10:24.864 17:31:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:24.864 17:31:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:24.864 17:31:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:10:24.864 17:31:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:24.864 17:31:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:24.864 17:31:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev4 00:10:24.864 17:31:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:24.864 17:31:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:24.864 17:31:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:10:24.864 17:31:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:10:24.864 17:31:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:10:24.864 17:31:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:10:24.864 17:31:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:10:24.864 17:31:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:10:24.864 17:31:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:10:24.864 17:31:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid0 '!=' raid1 ']' 00:10:24.864 17:31:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:10:24.864 17:31:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:10:24.864 17:31:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:10:24.864 17:31:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.HBlON6o7rz 00:10:24.864 17:31:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=81893 00:10:24.864 17:31:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:10:24.864 17:31:55 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 81893 00:10:24.864 17:31:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@831 -- # '[' -z 81893 ']' 00:10:24.864 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:24.864 17:31:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:24.864 17:31:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:10:24.864 17:31:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:24.864 17:31:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:10:24.864 17:31:55 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:24.864 [2024-11-27 17:31:55.940431] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:10:24.864 [2024-11-27 17:31:55.940554] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid81893 ] 00:10:25.124 [2024-11-27 17:31:56.086134] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:25.124 [2024-11-27 17:31:56.154245] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:10:25.124 [2024-11-27 17:31:56.229897] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:25.124 [2024-11-27 17:31:56.229946] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:25.693 17:31:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:10:25.693 17:31:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@864 -- # return 0 00:10:25.693 17:31:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:25.693 17:31:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:10:25.693 17:31:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:25.693 17:31:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:25.693 BaseBdev1_malloc 00:10:25.693 17:31:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:25.693 17:31:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:10:25.693 17:31:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:25.693 17:31:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:25.693 true 00:10:25.693 17:31:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:25.693 17:31:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:10:25.693 17:31:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:25.693 17:31:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:25.693 [2024-11-27 17:31:56.803708] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:10:25.693 [2024-11-27 17:31:56.803783] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:25.693 [2024-11-27 17:31:56.803808] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:10:25.693 [2024-11-27 17:31:56.803817] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:25.693 [2024-11-27 17:31:56.806328] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:25.693 [2024-11-27 17:31:56.806448] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:10:25.693 BaseBdev1 00:10:25.693 17:31:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:25.693 17:31:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:25.693 17:31:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:10:25.693 17:31:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:25.693 17:31:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:25.693 BaseBdev2_malloc 00:10:25.693 17:31:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:25.693 17:31:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:10:25.693 17:31:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:25.693 17:31:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:25.693 true 00:10:25.693 17:31:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:25.693 17:31:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:10:25.693 17:31:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:25.693 17:31:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:25.693 [2024-11-27 17:31:56.867802] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:10:25.693 [2024-11-27 17:31:56.867893] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:25.694 [2024-11-27 17:31:56.867926] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:10:25.694 [2024-11-27 17:31:56.867940] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:25.694 [2024-11-27 17:31:56.871456] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:25.694 [2024-11-27 17:31:56.871496] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:10:25.694 BaseBdev2 00:10:25.694 17:31:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:25.694 17:31:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:25.694 17:31:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:10:25.694 17:31:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:25.694 17:31:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:25.954 BaseBdev3_malloc 00:10:25.954 17:31:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:25.954 17:31:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:10:25.954 17:31:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:25.954 17:31:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:25.954 true 00:10:25.954 17:31:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:25.954 17:31:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:10:25.954 17:31:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:25.954 17:31:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:25.954 [2024-11-27 17:31:56.914544] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:10:25.954 [2024-11-27 17:31:56.914686] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:25.954 [2024-11-27 17:31:56.914713] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008780 00:10:25.954 [2024-11-27 17:31:56.914722] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:25.954 [2024-11-27 17:31:56.917125] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:25.954 [2024-11-27 17:31:56.917172] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:10:25.954 BaseBdev3 00:10:25.954 17:31:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:25.954 17:31:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:25.954 17:31:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:10:25.954 17:31:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:25.954 17:31:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:25.954 BaseBdev4_malloc 00:10:25.954 17:31:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:25.954 17:31:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev4_malloc 00:10:25.954 17:31:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:25.954 17:31:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:25.954 true 00:10:25.954 17:31:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:25.954 17:31:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev4_malloc -p BaseBdev4 00:10:25.954 17:31:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:25.954 17:31:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:25.954 [2024-11-27 17:31:56.960941] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev4_malloc 00:10:25.954 [2024-11-27 17:31:56.960988] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:25.954 [2024-11-27 17:31:56.961027] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:10:25.954 [2024-11-27 17:31:56.961036] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:25.954 [2024-11-27 17:31:56.963374] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:25.954 [2024-11-27 17:31:56.963408] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:10:25.954 BaseBdev4 00:10:25.954 17:31:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:25.954 17:31:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r raid0 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 -s 00:10:25.954 17:31:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:25.954 17:31:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:25.954 [2024-11-27 17:31:56.972980] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:25.954 [2024-11-27 17:31:56.975130] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:25.954 [2024-11-27 17:31:56.975216] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:25.954 [2024-11-27 17:31:56.975280] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:10:25.954 [2024-11-27 17:31:56.975472] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000002000 00:10:25.954 [2024-11-27 17:31:56.975482] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:10:25.954 [2024-11-27 17:31:56.975730] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:10:25.954 [2024-11-27 17:31:56.975865] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000002000 00:10:25.954 [2024-11-27 17:31:56.975879] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000002000 00:10:25.954 [2024-11-27 17:31:56.976006] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:25.954 17:31:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:25.954 17:31:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 4 00:10:25.954 17:31:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:25.954 17:31:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:25.954 17:31:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:25.954 17:31:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:25.954 17:31:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:25.954 17:31:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:25.954 17:31:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:25.954 17:31:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:25.954 17:31:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:25.954 17:31:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:25.954 17:31:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:25.954 17:31:56 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:25.954 17:31:56 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:25.954 17:31:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:25.954 17:31:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:25.954 "name": "raid_bdev1", 00:10:25.954 "uuid": "80e7ee4b-76dd-40ed-a5f1-d6e65f2b0758", 00:10:25.954 "strip_size_kb": 64, 00:10:25.954 "state": "online", 00:10:25.954 "raid_level": "raid0", 00:10:25.954 "superblock": true, 00:10:25.954 "num_base_bdevs": 4, 00:10:25.954 "num_base_bdevs_discovered": 4, 00:10:25.954 "num_base_bdevs_operational": 4, 00:10:25.954 "base_bdevs_list": [ 00:10:25.954 { 00:10:25.954 "name": "BaseBdev1", 00:10:25.954 "uuid": "4a36a766-8be3-57e0-917e-12b259c2ea91", 00:10:25.954 "is_configured": true, 00:10:25.954 "data_offset": 2048, 00:10:25.954 "data_size": 63488 00:10:25.954 }, 00:10:25.954 { 00:10:25.954 "name": "BaseBdev2", 00:10:25.954 "uuid": "2459daf1-869f-5188-83eb-943e42bb3123", 00:10:25.954 "is_configured": true, 00:10:25.954 "data_offset": 2048, 00:10:25.954 "data_size": 63488 00:10:25.954 }, 00:10:25.954 { 00:10:25.954 "name": "BaseBdev3", 00:10:25.954 "uuid": "d4f249dc-973b-5195-92c2-7cb76ce1888d", 00:10:25.954 "is_configured": true, 00:10:25.954 "data_offset": 2048, 00:10:25.954 "data_size": 63488 00:10:25.954 }, 00:10:25.954 { 00:10:25.954 "name": "BaseBdev4", 00:10:25.955 "uuid": "56cdb2b4-281e-539e-8062-71e9beaaa77f", 00:10:25.955 "is_configured": true, 00:10:25.955 "data_offset": 2048, 00:10:25.955 "data_size": 63488 00:10:25.955 } 00:10:25.955 ] 00:10:25.955 }' 00:10:25.955 17:31:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:25.955 17:31:57 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:26.523 17:31:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:10:26.523 17:31:57 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:10:26.523 [2024-11-27 17:31:57.528539] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000026d0 00:10:27.462 17:31:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:10:27.462 17:31:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:27.462 17:31:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:27.462 17:31:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:27.462 17:31:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:10:27.462 17:31:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid0 = \r\a\i\d\1 ]] 00:10:27.462 17:31:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=4 00:10:27.462 17:31:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid0 64 4 00:10:27.462 17:31:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:27.462 17:31:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:27.462 17:31:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid0 00:10:27.462 17:31:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:27.462 17:31:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:27.462 17:31:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:27.462 17:31:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:27.462 17:31:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:27.462 17:31:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:27.462 17:31:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:27.462 17:31:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:27.462 17:31:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:27.462 17:31:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:27.462 17:31:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:27.462 17:31:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:27.462 "name": "raid_bdev1", 00:10:27.462 "uuid": "80e7ee4b-76dd-40ed-a5f1-d6e65f2b0758", 00:10:27.462 "strip_size_kb": 64, 00:10:27.462 "state": "online", 00:10:27.462 "raid_level": "raid0", 00:10:27.462 "superblock": true, 00:10:27.462 "num_base_bdevs": 4, 00:10:27.462 "num_base_bdevs_discovered": 4, 00:10:27.462 "num_base_bdevs_operational": 4, 00:10:27.462 "base_bdevs_list": [ 00:10:27.462 { 00:10:27.462 "name": "BaseBdev1", 00:10:27.462 "uuid": "4a36a766-8be3-57e0-917e-12b259c2ea91", 00:10:27.462 "is_configured": true, 00:10:27.462 "data_offset": 2048, 00:10:27.462 "data_size": 63488 00:10:27.462 }, 00:10:27.462 { 00:10:27.462 "name": "BaseBdev2", 00:10:27.462 "uuid": "2459daf1-869f-5188-83eb-943e42bb3123", 00:10:27.462 "is_configured": true, 00:10:27.462 "data_offset": 2048, 00:10:27.462 "data_size": 63488 00:10:27.462 }, 00:10:27.462 { 00:10:27.462 "name": "BaseBdev3", 00:10:27.462 "uuid": "d4f249dc-973b-5195-92c2-7cb76ce1888d", 00:10:27.462 "is_configured": true, 00:10:27.462 "data_offset": 2048, 00:10:27.462 "data_size": 63488 00:10:27.462 }, 00:10:27.462 { 00:10:27.462 "name": "BaseBdev4", 00:10:27.462 "uuid": "56cdb2b4-281e-539e-8062-71e9beaaa77f", 00:10:27.462 "is_configured": true, 00:10:27.462 "data_offset": 2048, 00:10:27.462 "data_size": 63488 00:10:27.462 } 00:10:27.463 ] 00:10:27.463 }' 00:10:27.463 17:31:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:27.463 17:31:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:28.035 17:31:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:10:28.035 17:31:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:28.035 17:31:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:28.035 [2024-11-27 17:31:58.921268] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:10:28.035 [2024-11-27 17:31:58.921383] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:28.035 [2024-11-27 17:31:58.923952] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:28.035 [2024-11-27 17:31:58.924047] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:28.035 [2024-11-27 17:31:58.924131] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:28.035 [2024-11-27 17:31:58.924194] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002000 name raid_bdev1, state offline 00:10:28.035 { 00:10:28.035 "results": [ 00:10:28.035 { 00:10:28.035 "job": "raid_bdev1", 00:10:28.035 "core_mask": "0x1", 00:10:28.035 "workload": "randrw", 00:10:28.035 "percentage": 50, 00:10:28.035 "status": "finished", 00:10:28.035 "queue_depth": 1, 00:10:28.035 "io_size": 131072, 00:10:28.035 "runtime": 1.393453, 00:10:28.035 "iops": 14708.066938748561, 00:10:28.035 "mibps": 1838.5083673435702, 00:10:28.035 "io_failed": 1, 00:10:28.035 "io_timeout": 0, 00:10:28.035 "avg_latency_us": 95.49436336922913, 00:10:28.035 "min_latency_us": 24.705676855895195, 00:10:28.035 "max_latency_us": 1366.5257641921398 00:10:28.035 } 00:10:28.035 ], 00:10:28.035 "core_count": 1 00:10:28.035 } 00:10:28.035 17:31:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:28.035 17:31:58 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 81893 00:10:28.035 17:31:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@950 -- # '[' -z 81893 ']' 00:10:28.035 17:31:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # kill -0 81893 00:10:28.035 17:31:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@955 -- # uname 00:10:28.035 17:31:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:10:28.035 17:31:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 81893 00:10:28.035 17:31:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:10:28.035 17:31:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:10:28.035 17:31:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 81893' 00:10:28.035 killing process with pid 81893 00:10:28.035 17:31:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@969 -- # kill 81893 00:10:28.035 [2024-11-27 17:31:58.959774] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:10:28.035 17:31:58 bdev_raid.raid_write_error_test -- common/autotest_common.sh@974 -- # wait 81893 00:10:28.035 [2024-11-27 17:31:59.023613] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:10:28.295 17:31:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.HBlON6o7rz 00:10:28.295 17:31:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:10:28.295 17:31:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:10:28.295 17:31:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.72 00:10:28.295 17:31:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid0 00:10:28.295 17:31:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:10:28.295 17:31:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:10:28.295 17:31:59 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.72 != \0\.\0\0 ]] 00:10:28.295 00:10:28.295 real 0m3.562s 00:10:28.295 user 0m4.350s 00:10:28.295 sys 0m0.641s 00:10:28.295 ************************************ 00:10:28.295 END TEST raid_write_error_test 00:10:28.295 ************************************ 00:10:28.295 17:31:59 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:10:28.295 17:31:59 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:28.295 17:31:59 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:10:28.295 17:31:59 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test concat 4 false 00:10:28.295 17:31:59 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:10:28.295 17:31:59 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:10:28.295 17:31:59 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:10:28.295 ************************************ 00:10:28.295 START TEST raid_state_function_test 00:10:28.295 ************************************ 00:10:28.295 17:31:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1125 -- # raid_state_function_test concat 4 false 00:10:28.295 17:31:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=concat 00:10:28.295 17:31:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=4 00:10:28.295 17:31:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:10:28.295 17:31:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:10:28.295 17:31:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:10:28.295 17:31:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:28.295 17:31:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:10:28.295 17:31:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:28.295 17:31:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:28.295 17:31:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:10:28.295 17:31:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:28.295 17:31:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:28.295 17:31:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:10:28.295 17:31:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:28.295 17:31:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:28.295 17:31:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev4 00:10:28.295 17:31:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:28.295 17:31:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:28.295 17:31:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:10:28.295 17:31:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:10:28.295 17:31:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:10:28.295 17:31:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:10:28.295 17:31:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:10:28.295 17:31:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:10:28.295 17:31:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' concat '!=' raid1 ']' 00:10:28.295 17:31:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:10:28.295 17:31:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:10:28.295 17:31:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:10:28.295 17:31:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:10:28.555 Process raid pid: 82020 00:10:28.555 17:31:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=82020 00:10:28.555 17:31:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:10:28.555 17:31:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 82020' 00:10:28.555 17:31:59 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 82020 00:10:28.555 17:31:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@831 -- # '[' -z 82020 ']' 00:10:28.555 17:31:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:28.555 17:31:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:10:28.555 17:31:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:28.555 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:28.555 17:31:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:10:28.555 17:31:59 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:28.555 [2024-11-27 17:31:59.568242] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:10:28.555 [2024-11-27 17:31:59.568470] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:10:28.555 [2024-11-27 17:31:59.696619] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:28.815 [2024-11-27 17:31:59.767505] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:10:28.815 [2024-11-27 17:31:59.846626] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:28.815 [2024-11-27 17:31:59.846661] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:29.387 17:32:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:10:29.387 17:32:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@864 -- # return 0 00:10:29.387 17:32:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:10:29.387 17:32:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:29.387 17:32:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:29.387 [2024-11-27 17:32:00.395187] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:10:29.387 [2024-11-27 17:32:00.395335] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:10:29.387 [2024-11-27 17:32:00.395354] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:10:29.387 [2024-11-27 17:32:00.395364] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:10:29.387 [2024-11-27 17:32:00.395370] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:10:29.387 [2024-11-27 17:32:00.395384] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:10:29.387 [2024-11-27 17:32:00.395390] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:10:29.387 [2024-11-27 17:32:00.395399] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:10:29.387 17:32:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:29.387 17:32:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:29.387 17:32:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:29.387 17:32:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:29.387 17:32:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:29.387 17:32:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:29.387 17:32:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:29.387 17:32:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:29.387 17:32:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:29.387 17:32:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:29.387 17:32:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:29.387 17:32:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:29.387 17:32:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:29.387 17:32:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:29.387 17:32:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:29.387 17:32:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:29.387 17:32:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:29.387 "name": "Existed_Raid", 00:10:29.387 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:29.387 "strip_size_kb": 64, 00:10:29.387 "state": "configuring", 00:10:29.387 "raid_level": "concat", 00:10:29.387 "superblock": false, 00:10:29.387 "num_base_bdevs": 4, 00:10:29.387 "num_base_bdevs_discovered": 0, 00:10:29.387 "num_base_bdevs_operational": 4, 00:10:29.387 "base_bdevs_list": [ 00:10:29.387 { 00:10:29.387 "name": "BaseBdev1", 00:10:29.387 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:29.387 "is_configured": false, 00:10:29.387 "data_offset": 0, 00:10:29.387 "data_size": 0 00:10:29.387 }, 00:10:29.387 { 00:10:29.387 "name": "BaseBdev2", 00:10:29.387 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:29.387 "is_configured": false, 00:10:29.387 "data_offset": 0, 00:10:29.387 "data_size": 0 00:10:29.387 }, 00:10:29.387 { 00:10:29.387 "name": "BaseBdev3", 00:10:29.387 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:29.387 "is_configured": false, 00:10:29.387 "data_offset": 0, 00:10:29.387 "data_size": 0 00:10:29.387 }, 00:10:29.387 { 00:10:29.387 "name": "BaseBdev4", 00:10:29.387 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:29.387 "is_configured": false, 00:10:29.387 "data_offset": 0, 00:10:29.387 "data_size": 0 00:10:29.387 } 00:10:29.387 ] 00:10:29.387 }' 00:10:29.387 17:32:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:29.387 17:32:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:29.648 17:32:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:10:29.648 17:32:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:29.648 17:32:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:29.908 [2024-11-27 17:32:00.838300] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:10:29.908 [2024-11-27 17:32:00.838397] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:10:29.908 17:32:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:29.908 17:32:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:10:29.908 17:32:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:29.908 17:32:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:29.908 [2024-11-27 17:32:00.850290] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:10:29.908 [2024-11-27 17:32:00.850371] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:10:29.908 [2024-11-27 17:32:00.850414] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:10:29.908 [2024-11-27 17:32:00.850445] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:10:29.908 [2024-11-27 17:32:00.850462] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:10:29.908 [2024-11-27 17:32:00.850483] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:10:29.908 [2024-11-27 17:32:00.850501] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:10:29.908 [2024-11-27 17:32:00.850522] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:10:29.908 17:32:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:29.908 17:32:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:10:29.908 17:32:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:29.908 17:32:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:29.908 [2024-11-27 17:32:00.877852] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:29.908 BaseBdev1 00:10:29.908 17:32:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:29.908 17:32:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:10:29.908 17:32:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:10:29.908 17:32:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:10:29.908 17:32:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:10:29.908 17:32:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:10:29.908 17:32:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:10:29.908 17:32:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:10:29.908 17:32:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:29.908 17:32:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:29.908 17:32:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:29.908 17:32:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:10:29.908 17:32:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:29.908 17:32:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:29.908 [ 00:10:29.908 { 00:10:29.908 "name": "BaseBdev1", 00:10:29.908 "aliases": [ 00:10:29.908 "862e4113-2785-4be6-a749-5477a18aa000" 00:10:29.908 ], 00:10:29.908 "product_name": "Malloc disk", 00:10:29.908 "block_size": 512, 00:10:29.908 "num_blocks": 65536, 00:10:29.908 "uuid": "862e4113-2785-4be6-a749-5477a18aa000", 00:10:29.908 "assigned_rate_limits": { 00:10:29.908 "rw_ios_per_sec": 0, 00:10:29.908 "rw_mbytes_per_sec": 0, 00:10:29.908 "r_mbytes_per_sec": 0, 00:10:29.908 "w_mbytes_per_sec": 0 00:10:29.908 }, 00:10:29.908 "claimed": true, 00:10:29.908 "claim_type": "exclusive_write", 00:10:29.908 "zoned": false, 00:10:29.908 "supported_io_types": { 00:10:29.908 "read": true, 00:10:29.908 "write": true, 00:10:29.908 "unmap": true, 00:10:29.908 "flush": true, 00:10:29.908 "reset": true, 00:10:29.908 "nvme_admin": false, 00:10:29.908 "nvme_io": false, 00:10:29.908 "nvme_io_md": false, 00:10:29.908 "write_zeroes": true, 00:10:29.908 "zcopy": true, 00:10:29.908 "get_zone_info": false, 00:10:29.908 "zone_management": false, 00:10:29.908 "zone_append": false, 00:10:29.908 "compare": false, 00:10:29.908 "compare_and_write": false, 00:10:29.908 "abort": true, 00:10:29.908 "seek_hole": false, 00:10:29.908 "seek_data": false, 00:10:29.908 "copy": true, 00:10:29.908 "nvme_iov_md": false 00:10:29.908 }, 00:10:29.908 "memory_domains": [ 00:10:29.908 { 00:10:29.908 "dma_device_id": "system", 00:10:29.908 "dma_device_type": 1 00:10:29.908 }, 00:10:29.908 { 00:10:29.908 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:29.908 "dma_device_type": 2 00:10:29.908 } 00:10:29.908 ], 00:10:29.908 "driver_specific": {} 00:10:29.908 } 00:10:29.908 ] 00:10:29.908 17:32:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:29.909 17:32:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:10:29.909 17:32:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:29.909 17:32:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:29.909 17:32:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:29.909 17:32:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:29.909 17:32:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:29.909 17:32:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:29.909 17:32:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:29.909 17:32:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:29.909 17:32:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:29.909 17:32:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:29.909 17:32:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:29.909 17:32:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:29.909 17:32:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:29.909 17:32:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:29.909 17:32:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:29.909 17:32:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:29.909 "name": "Existed_Raid", 00:10:29.909 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:29.909 "strip_size_kb": 64, 00:10:29.909 "state": "configuring", 00:10:29.909 "raid_level": "concat", 00:10:29.909 "superblock": false, 00:10:29.909 "num_base_bdevs": 4, 00:10:29.909 "num_base_bdevs_discovered": 1, 00:10:29.909 "num_base_bdevs_operational": 4, 00:10:29.909 "base_bdevs_list": [ 00:10:29.909 { 00:10:29.909 "name": "BaseBdev1", 00:10:29.909 "uuid": "862e4113-2785-4be6-a749-5477a18aa000", 00:10:29.909 "is_configured": true, 00:10:29.909 "data_offset": 0, 00:10:29.909 "data_size": 65536 00:10:29.909 }, 00:10:29.909 { 00:10:29.909 "name": "BaseBdev2", 00:10:29.909 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:29.909 "is_configured": false, 00:10:29.909 "data_offset": 0, 00:10:29.909 "data_size": 0 00:10:29.909 }, 00:10:29.909 { 00:10:29.909 "name": "BaseBdev3", 00:10:29.909 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:29.909 "is_configured": false, 00:10:29.909 "data_offset": 0, 00:10:29.909 "data_size": 0 00:10:29.909 }, 00:10:29.909 { 00:10:29.909 "name": "BaseBdev4", 00:10:29.909 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:29.909 "is_configured": false, 00:10:29.909 "data_offset": 0, 00:10:29.909 "data_size": 0 00:10:29.909 } 00:10:29.909 ] 00:10:29.909 }' 00:10:29.909 17:32:00 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:29.909 17:32:00 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:30.169 17:32:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:10:30.169 17:32:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:30.169 17:32:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:30.429 [2024-11-27 17:32:01.361013] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:10:30.429 [2024-11-27 17:32:01.361063] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:10:30.429 17:32:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:30.429 17:32:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:10:30.429 17:32:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:30.429 17:32:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:30.429 [2024-11-27 17:32:01.373062] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:30.429 [2024-11-27 17:32:01.375241] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:10:30.429 [2024-11-27 17:32:01.375315] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:10:30.429 [2024-11-27 17:32:01.375343] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:10:30.429 [2024-11-27 17:32:01.375394] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:10:30.429 [2024-11-27 17:32:01.375412] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:10:30.429 [2024-11-27 17:32:01.375432] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:10:30.429 17:32:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:30.429 17:32:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:10:30.429 17:32:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:30.429 17:32:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:30.429 17:32:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:30.429 17:32:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:30.429 17:32:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:30.429 17:32:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:30.429 17:32:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:30.429 17:32:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:30.429 17:32:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:30.429 17:32:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:30.429 17:32:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:30.429 17:32:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:30.429 17:32:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:30.429 17:32:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:30.429 17:32:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:30.429 17:32:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:30.429 17:32:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:30.429 "name": "Existed_Raid", 00:10:30.429 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:30.429 "strip_size_kb": 64, 00:10:30.429 "state": "configuring", 00:10:30.429 "raid_level": "concat", 00:10:30.429 "superblock": false, 00:10:30.429 "num_base_bdevs": 4, 00:10:30.429 "num_base_bdevs_discovered": 1, 00:10:30.429 "num_base_bdevs_operational": 4, 00:10:30.429 "base_bdevs_list": [ 00:10:30.429 { 00:10:30.429 "name": "BaseBdev1", 00:10:30.429 "uuid": "862e4113-2785-4be6-a749-5477a18aa000", 00:10:30.429 "is_configured": true, 00:10:30.429 "data_offset": 0, 00:10:30.429 "data_size": 65536 00:10:30.429 }, 00:10:30.429 { 00:10:30.429 "name": "BaseBdev2", 00:10:30.429 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:30.429 "is_configured": false, 00:10:30.429 "data_offset": 0, 00:10:30.429 "data_size": 0 00:10:30.429 }, 00:10:30.429 { 00:10:30.429 "name": "BaseBdev3", 00:10:30.429 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:30.429 "is_configured": false, 00:10:30.429 "data_offset": 0, 00:10:30.429 "data_size": 0 00:10:30.429 }, 00:10:30.429 { 00:10:30.429 "name": "BaseBdev4", 00:10:30.429 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:30.429 "is_configured": false, 00:10:30.429 "data_offset": 0, 00:10:30.429 "data_size": 0 00:10:30.429 } 00:10:30.429 ] 00:10:30.429 }' 00:10:30.429 17:32:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:30.429 17:32:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:30.690 17:32:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:10:30.690 17:32:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:30.690 17:32:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:30.690 [2024-11-27 17:32:01.811154] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:30.690 BaseBdev2 00:10:30.690 17:32:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:30.690 17:32:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:10:30.690 17:32:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:10:30.690 17:32:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:10:30.690 17:32:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:10:30.690 17:32:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:10:30.690 17:32:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:10:30.690 17:32:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:10:30.690 17:32:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:30.690 17:32:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:30.690 17:32:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:30.690 17:32:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:10:30.690 17:32:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:30.690 17:32:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:30.690 [ 00:10:30.690 { 00:10:30.690 "name": "BaseBdev2", 00:10:30.690 "aliases": [ 00:10:30.690 "c5a98eb1-f908-4e70-9538-a95c29933019" 00:10:30.690 ], 00:10:30.690 "product_name": "Malloc disk", 00:10:30.690 "block_size": 512, 00:10:30.690 "num_blocks": 65536, 00:10:30.690 "uuid": "c5a98eb1-f908-4e70-9538-a95c29933019", 00:10:30.690 "assigned_rate_limits": { 00:10:30.690 "rw_ios_per_sec": 0, 00:10:30.690 "rw_mbytes_per_sec": 0, 00:10:30.690 "r_mbytes_per_sec": 0, 00:10:30.690 "w_mbytes_per_sec": 0 00:10:30.690 }, 00:10:30.690 "claimed": true, 00:10:30.690 "claim_type": "exclusive_write", 00:10:30.690 "zoned": false, 00:10:30.690 "supported_io_types": { 00:10:30.690 "read": true, 00:10:30.691 "write": true, 00:10:30.691 "unmap": true, 00:10:30.691 "flush": true, 00:10:30.691 "reset": true, 00:10:30.691 "nvme_admin": false, 00:10:30.691 "nvme_io": false, 00:10:30.691 "nvme_io_md": false, 00:10:30.691 "write_zeroes": true, 00:10:30.691 "zcopy": true, 00:10:30.691 "get_zone_info": false, 00:10:30.691 "zone_management": false, 00:10:30.691 "zone_append": false, 00:10:30.691 "compare": false, 00:10:30.691 "compare_and_write": false, 00:10:30.691 "abort": true, 00:10:30.691 "seek_hole": false, 00:10:30.691 "seek_data": false, 00:10:30.691 "copy": true, 00:10:30.691 "nvme_iov_md": false 00:10:30.691 }, 00:10:30.691 "memory_domains": [ 00:10:30.691 { 00:10:30.691 "dma_device_id": "system", 00:10:30.691 "dma_device_type": 1 00:10:30.691 }, 00:10:30.691 { 00:10:30.691 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:30.691 "dma_device_type": 2 00:10:30.691 } 00:10:30.691 ], 00:10:30.691 "driver_specific": {} 00:10:30.691 } 00:10:30.691 ] 00:10:30.691 17:32:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:30.691 17:32:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:10:30.691 17:32:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:10:30.691 17:32:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:30.691 17:32:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:30.691 17:32:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:30.691 17:32:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:30.691 17:32:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:30.691 17:32:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:30.691 17:32:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:30.691 17:32:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:30.691 17:32:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:30.691 17:32:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:30.691 17:32:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:30.691 17:32:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:30.691 17:32:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:30.691 17:32:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:30.691 17:32:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:30.691 17:32:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:30.951 17:32:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:30.951 "name": "Existed_Raid", 00:10:30.951 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:30.951 "strip_size_kb": 64, 00:10:30.951 "state": "configuring", 00:10:30.951 "raid_level": "concat", 00:10:30.951 "superblock": false, 00:10:30.951 "num_base_bdevs": 4, 00:10:30.951 "num_base_bdevs_discovered": 2, 00:10:30.951 "num_base_bdevs_operational": 4, 00:10:30.951 "base_bdevs_list": [ 00:10:30.951 { 00:10:30.951 "name": "BaseBdev1", 00:10:30.951 "uuid": "862e4113-2785-4be6-a749-5477a18aa000", 00:10:30.951 "is_configured": true, 00:10:30.951 "data_offset": 0, 00:10:30.951 "data_size": 65536 00:10:30.951 }, 00:10:30.951 { 00:10:30.951 "name": "BaseBdev2", 00:10:30.951 "uuid": "c5a98eb1-f908-4e70-9538-a95c29933019", 00:10:30.951 "is_configured": true, 00:10:30.951 "data_offset": 0, 00:10:30.951 "data_size": 65536 00:10:30.951 }, 00:10:30.951 { 00:10:30.951 "name": "BaseBdev3", 00:10:30.951 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:30.951 "is_configured": false, 00:10:30.951 "data_offset": 0, 00:10:30.951 "data_size": 0 00:10:30.951 }, 00:10:30.951 { 00:10:30.951 "name": "BaseBdev4", 00:10:30.951 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:30.951 "is_configured": false, 00:10:30.951 "data_offset": 0, 00:10:30.951 "data_size": 0 00:10:30.951 } 00:10:30.951 ] 00:10:30.951 }' 00:10:30.951 17:32:01 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:30.951 17:32:01 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:31.212 17:32:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:10:31.212 17:32:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:31.212 17:32:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:31.212 [2024-11-27 17:32:02.323319] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:31.212 BaseBdev3 00:10:31.212 17:32:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:31.212 17:32:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:10:31.212 17:32:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:10:31.212 17:32:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:10:31.212 17:32:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:10:31.212 17:32:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:10:31.212 17:32:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:10:31.212 17:32:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:10:31.212 17:32:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:31.212 17:32:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:31.212 17:32:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:31.212 17:32:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:10:31.212 17:32:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:31.212 17:32:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:31.212 [ 00:10:31.212 { 00:10:31.212 "name": "BaseBdev3", 00:10:31.212 "aliases": [ 00:10:31.212 "d0de4cf0-5a2b-4b2e-b49f-fab88d122f67" 00:10:31.212 ], 00:10:31.212 "product_name": "Malloc disk", 00:10:31.212 "block_size": 512, 00:10:31.212 "num_blocks": 65536, 00:10:31.212 "uuid": "d0de4cf0-5a2b-4b2e-b49f-fab88d122f67", 00:10:31.212 "assigned_rate_limits": { 00:10:31.212 "rw_ios_per_sec": 0, 00:10:31.212 "rw_mbytes_per_sec": 0, 00:10:31.212 "r_mbytes_per_sec": 0, 00:10:31.212 "w_mbytes_per_sec": 0 00:10:31.212 }, 00:10:31.212 "claimed": true, 00:10:31.212 "claim_type": "exclusive_write", 00:10:31.212 "zoned": false, 00:10:31.212 "supported_io_types": { 00:10:31.212 "read": true, 00:10:31.212 "write": true, 00:10:31.212 "unmap": true, 00:10:31.212 "flush": true, 00:10:31.212 "reset": true, 00:10:31.212 "nvme_admin": false, 00:10:31.212 "nvme_io": false, 00:10:31.212 "nvme_io_md": false, 00:10:31.212 "write_zeroes": true, 00:10:31.212 "zcopy": true, 00:10:31.212 "get_zone_info": false, 00:10:31.212 "zone_management": false, 00:10:31.212 "zone_append": false, 00:10:31.212 "compare": false, 00:10:31.212 "compare_and_write": false, 00:10:31.212 "abort": true, 00:10:31.212 "seek_hole": false, 00:10:31.212 "seek_data": false, 00:10:31.212 "copy": true, 00:10:31.212 "nvme_iov_md": false 00:10:31.212 }, 00:10:31.212 "memory_domains": [ 00:10:31.212 { 00:10:31.212 "dma_device_id": "system", 00:10:31.212 "dma_device_type": 1 00:10:31.212 }, 00:10:31.212 { 00:10:31.212 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:31.212 "dma_device_type": 2 00:10:31.212 } 00:10:31.212 ], 00:10:31.212 "driver_specific": {} 00:10:31.212 } 00:10:31.212 ] 00:10:31.212 17:32:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:31.212 17:32:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:10:31.212 17:32:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:10:31.212 17:32:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:31.212 17:32:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:31.212 17:32:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:31.212 17:32:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:31.212 17:32:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:31.212 17:32:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:31.212 17:32:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:31.212 17:32:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:31.212 17:32:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:31.212 17:32:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:31.212 17:32:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:31.212 17:32:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:31.212 17:32:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:31.212 17:32:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:31.212 17:32:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:31.212 17:32:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:31.472 17:32:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:31.472 "name": "Existed_Raid", 00:10:31.472 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:31.472 "strip_size_kb": 64, 00:10:31.472 "state": "configuring", 00:10:31.472 "raid_level": "concat", 00:10:31.472 "superblock": false, 00:10:31.472 "num_base_bdevs": 4, 00:10:31.472 "num_base_bdevs_discovered": 3, 00:10:31.472 "num_base_bdevs_operational": 4, 00:10:31.472 "base_bdevs_list": [ 00:10:31.472 { 00:10:31.472 "name": "BaseBdev1", 00:10:31.472 "uuid": "862e4113-2785-4be6-a749-5477a18aa000", 00:10:31.472 "is_configured": true, 00:10:31.472 "data_offset": 0, 00:10:31.472 "data_size": 65536 00:10:31.472 }, 00:10:31.472 { 00:10:31.472 "name": "BaseBdev2", 00:10:31.472 "uuid": "c5a98eb1-f908-4e70-9538-a95c29933019", 00:10:31.472 "is_configured": true, 00:10:31.472 "data_offset": 0, 00:10:31.472 "data_size": 65536 00:10:31.472 }, 00:10:31.472 { 00:10:31.472 "name": "BaseBdev3", 00:10:31.472 "uuid": "d0de4cf0-5a2b-4b2e-b49f-fab88d122f67", 00:10:31.472 "is_configured": true, 00:10:31.472 "data_offset": 0, 00:10:31.472 "data_size": 65536 00:10:31.472 }, 00:10:31.472 { 00:10:31.472 "name": "BaseBdev4", 00:10:31.472 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:31.472 "is_configured": false, 00:10:31.472 "data_offset": 0, 00:10:31.472 "data_size": 0 00:10:31.472 } 00:10:31.472 ] 00:10:31.472 }' 00:10:31.472 17:32:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:31.472 17:32:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:31.733 17:32:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:10:31.733 17:32:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:31.733 17:32:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:31.733 [2024-11-27 17:32:02.751571] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:10:31.733 [2024-11-27 17:32:02.751704] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:10:31.733 [2024-11-27 17:32:02.751719] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 262144, blocklen 512 00:10:31.733 [2024-11-27 17:32:02.752058] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:10:31.733 [2024-11-27 17:32:02.752243] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:10:31.733 [2024-11-27 17:32:02.752270] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:10:31.733 [2024-11-27 17:32:02.752500] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:31.733 BaseBdev4 00:10:31.733 17:32:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:31.733 17:32:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev4 00:10:31.733 17:32:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev4 00:10:31.733 17:32:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:10:31.733 17:32:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:10:31.733 17:32:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:10:31.733 17:32:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:10:31.733 17:32:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:10:31.733 17:32:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:31.733 17:32:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:31.733 17:32:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:31.733 17:32:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:10:31.733 17:32:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:31.733 17:32:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:31.733 [ 00:10:31.733 { 00:10:31.733 "name": "BaseBdev4", 00:10:31.733 "aliases": [ 00:10:31.733 "c6c989da-f006-4684-a5b9-0d19718a177a" 00:10:31.733 ], 00:10:31.733 "product_name": "Malloc disk", 00:10:31.733 "block_size": 512, 00:10:31.733 "num_blocks": 65536, 00:10:31.733 "uuid": "c6c989da-f006-4684-a5b9-0d19718a177a", 00:10:31.733 "assigned_rate_limits": { 00:10:31.733 "rw_ios_per_sec": 0, 00:10:31.733 "rw_mbytes_per_sec": 0, 00:10:31.733 "r_mbytes_per_sec": 0, 00:10:31.733 "w_mbytes_per_sec": 0 00:10:31.733 }, 00:10:31.733 "claimed": true, 00:10:31.733 "claim_type": "exclusive_write", 00:10:31.733 "zoned": false, 00:10:31.733 "supported_io_types": { 00:10:31.733 "read": true, 00:10:31.733 "write": true, 00:10:31.733 "unmap": true, 00:10:31.733 "flush": true, 00:10:31.733 "reset": true, 00:10:31.733 "nvme_admin": false, 00:10:31.733 "nvme_io": false, 00:10:31.733 "nvme_io_md": false, 00:10:31.733 "write_zeroes": true, 00:10:31.733 "zcopy": true, 00:10:31.733 "get_zone_info": false, 00:10:31.733 "zone_management": false, 00:10:31.733 "zone_append": false, 00:10:31.733 "compare": false, 00:10:31.733 "compare_and_write": false, 00:10:31.733 "abort": true, 00:10:31.733 "seek_hole": false, 00:10:31.733 "seek_data": false, 00:10:31.733 "copy": true, 00:10:31.733 "nvme_iov_md": false 00:10:31.733 }, 00:10:31.733 "memory_domains": [ 00:10:31.733 { 00:10:31.733 "dma_device_id": "system", 00:10:31.733 "dma_device_type": 1 00:10:31.733 }, 00:10:31.733 { 00:10:31.733 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:31.733 "dma_device_type": 2 00:10:31.733 } 00:10:31.733 ], 00:10:31.733 "driver_specific": {} 00:10:31.733 } 00:10:31.733 ] 00:10:31.733 17:32:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:31.733 17:32:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:10:31.733 17:32:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:10:31.733 17:32:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:31.733 17:32:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online concat 64 4 00:10:31.733 17:32:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:31.733 17:32:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:31.734 17:32:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:31.734 17:32:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:31.734 17:32:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:31.734 17:32:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:31.734 17:32:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:31.734 17:32:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:31.734 17:32:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:31.734 17:32:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:31.734 17:32:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:31.734 17:32:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:31.734 17:32:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:31.734 17:32:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:31.734 17:32:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:31.734 "name": "Existed_Raid", 00:10:31.734 "uuid": "6861168e-107c-47c8-b43a-e43f82713b21", 00:10:31.734 "strip_size_kb": 64, 00:10:31.734 "state": "online", 00:10:31.734 "raid_level": "concat", 00:10:31.734 "superblock": false, 00:10:31.734 "num_base_bdevs": 4, 00:10:31.734 "num_base_bdevs_discovered": 4, 00:10:31.734 "num_base_bdevs_operational": 4, 00:10:31.734 "base_bdevs_list": [ 00:10:31.734 { 00:10:31.734 "name": "BaseBdev1", 00:10:31.734 "uuid": "862e4113-2785-4be6-a749-5477a18aa000", 00:10:31.734 "is_configured": true, 00:10:31.734 "data_offset": 0, 00:10:31.734 "data_size": 65536 00:10:31.734 }, 00:10:31.734 { 00:10:31.734 "name": "BaseBdev2", 00:10:31.734 "uuid": "c5a98eb1-f908-4e70-9538-a95c29933019", 00:10:31.734 "is_configured": true, 00:10:31.734 "data_offset": 0, 00:10:31.734 "data_size": 65536 00:10:31.734 }, 00:10:31.734 { 00:10:31.734 "name": "BaseBdev3", 00:10:31.734 "uuid": "d0de4cf0-5a2b-4b2e-b49f-fab88d122f67", 00:10:31.734 "is_configured": true, 00:10:31.734 "data_offset": 0, 00:10:31.734 "data_size": 65536 00:10:31.734 }, 00:10:31.734 { 00:10:31.734 "name": "BaseBdev4", 00:10:31.734 "uuid": "c6c989da-f006-4684-a5b9-0d19718a177a", 00:10:31.734 "is_configured": true, 00:10:31.734 "data_offset": 0, 00:10:31.734 "data_size": 65536 00:10:31.734 } 00:10:31.734 ] 00:10:31.734 }' 00:10:31.734 17:32:02 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:31.734 17:32:02 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:32.304 17:32:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:10:32.304 17:32:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:10:32.304 17:32:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:10:32.304 17:32:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:10:32.304 17:32:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:10:32.304 17:32:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:10:32.304 17:32:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:10:32.304 17:32:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:10:32.304 17:32:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:32.304 17:32:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:32.304 [2024-11-27 17:32:03.251117] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:32.304 17:32:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:32.304 17:32:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:32.304 "name": "Existed_Raid", 00:10:32.304 "aliases": [ 00:10:32.304 "6861168e-107c-47c8-b43a-e43f82713b21" 00:10:32.304 ], 00:10:32.304 "product_name": "Raid Volume", 00:10:32.304 "block_size": 512, 00:10:32.304 "num_blocks": 262144, 00:10:32.304 "uuid": "6861168e-107c-47c8-b43a-e43f82713b21", 00:10:32.304 "assigned_rate_limits": { 00:10:32.304 "rw_ios_per_sec": 0, 00:10:32.304 "rw_mbytes_per_sec": 0, 00:10:32.304 "r_mbytes_per_sec": 0, 00:10:32.304 "w_mbytes_per_sec": 0 00:10:32.304 }, 00:10:32.304 "claimed": false, 00:10:32.304 "zoned": false, 00:10:32.304 "supported_io_types": { 00:10:32.304 "read": true, 00:10:32.304 "write": true, 00:10:32.304 "unmap": true, 00:10:32.304 "flush": true, 00:10:32.304 "reset": true, 00:10:32.304 "nvme_admin": false, 00:10:32.304 "nvme_io": false, 00:10:32.304 "nvme_io_md": false, 00:10:32.304 "write_zeroes": true, 00:10:32.304 "zcopy": false, 00:10:32.304 "get_zone_info": false, 00:10:32.304 "zone_management": false, 00:10:32.304 "zone_append": false, 00:10:32.304 "compare": false, 00:10:32.304 "compare_and_write": false, 00:10:32.304 "abort": false, 00:10:32.304 "seek_hole": false, 00:10:32.305 "seek_data": false, 00:10:32.305 "copy": false, 00:10:32.305 "nvme_iov_md": false 00:10:32.305 }, 00:10:32.305 "memory_domains": [ 00:10:32.305 { 00:10:32.305 "dma_device_id": "system", 00:10:32.305 "dma_device_type": 1 00:10:32.305 }, 00:10:32.305 { 00:10:32.305 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:32.305 "dma_device_type": 2 00:10:32.305 }, 00:10:32.305 { 00:10:32.305 "dma_device_id": "system", 00:10:32.305 "dma_device_type": 1 00:10:32.305 }, 00:10:32.305 { 00:10:32.305 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:32.305 "dma_device_type": 2 00:10:32.305 }, 00:10:32.305 { 00:10:32.305 "dma_device_id": "system", 00:10:32.305 "dma_device_type": 1 00:10:32.305 }, 00:10:32.305 { 00:10:32.305 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:32.305 "dma_device_type": 2 00:10:32.305 }, 00:10:32.305 { 00:10:32.305 "dma_device_id": "system", 00:10:32.305 "dma_device_type": 1 00:10:32.305 }, 00:10:32.305 { 00:10:32.305 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:32.305 "dma_device_type": 2 00:10:32.305 } 00:10:32.305 ], 00:10:32.305 "driver_specific": { 00:10:32.305 "raid": { 00:10:32.305 "uuid": "6861168e-107c-47c8-b43a-e43f82713b21", 00:10:32.305 "strip_size_kb": 64, 00:10:32.305 "state": "online", 00:10:32.305 "raid_level": "concat", 00:10:32.305 "superblock": false, 00:10:32.305 "num_base_bdevs": 4, 00:10:32.305 "num_base_bdevs_discovered": 4, 00:10:32.305 "num_base_bdevs_operational": 4, 00:10:32.305 "base_bdevs_list": [ 00:10:32.305 { 00:10:32.305 "name": "BaseBdev1", 00:10:32.305 "uuid": "862e4113-2785-4be6-a749-5477a18aa000", 00:10:32.305 "is_configured": true, 00:10:32.305 "data_offset": 0, 00:10:32.305 "data_size": 65536 00:10:32.305 }, 00:10:32.305 { 00:10:32.305 "name": "BaseBdev2", 00:10:32.305 "uuid": "c5a98eb1-f908-4e70-9538-a95c29933019", 00:10:32.305 "is_configured": true, 00:10:32.305 "data_offset": 0, 00:10:32.305 "data_size": 65536 00:10:32.305 }, 00:10:32.305 { 00:10:32.305 "name": "BaseBdev3", 00:10:32.305 "uuid": "d0de4cf0-5a2b-4b2e-b49f-fab88d122f67", 00:10:32.305 "is_configured": true, 00:10:32.305 "data_offset": 0, 00:10:32.305 "data_size": 65536 00:10:32.305 }, 00:10:32.305 { 00:10:32.305 "name": "BaseBdev4", 00:10:32.305 "uuid": "c6c989da-f006-4684-a5b9-0d19718a177a", 00:10:32.305 "is_configured": true, 00:10:32.305 "data_offset": 0, 00:10:32.305 "data_size": 65536 00:10:32.305 } 00:10:32.305 ] 00:10:32.305 } 00:10:32.305 } 00:10:32.305 }' 00:10:32.305 17:32:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:32.305 17:32:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:10:32.305 BaseBdev2 00:10:32.305 BaseBdev3 00:10:32.305 BaseBdev4' 00:10:32.305 17:32:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:32.305 17:32:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:32.305 17:32:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:32.305 17:32:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:10:32.305 17:32:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:32.305 17:32:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:32.305 17:32:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:32.305 17:32:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:32.305 17:32:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:32.305 17:32:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:32.305 17:32:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:32.305 17:32:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:32.305 17:32:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:10:32.305 17:32:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:32.305 17:32:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:32.305 17:32:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:32.305 17:32:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:32.305 17:32:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:32.305 17:32:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:32.305 17:32:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:32.305 17:32:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:10:32.305 17:32:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:32.305 17:32:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:32.305 17:32:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:32.305 17:32:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:32.305 17:32:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:32.305 17:32:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:32.305 17:32:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:32.305 17:32:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:10:32.305 17:32:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:32.305 17:32:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:32.565 17:32:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:32.565 17:32:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:32.566 17:32:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:32.566 17:32:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:10:32.566 17:32:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:32.566 17:32:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:32.566 [2024-11-27 17:32:03.510355] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:10:32.566 [2024-11-27 17:32:03.510387] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:32.566 [2024-11-27 17:32:03.510447] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:32.566 17:32:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:32.566 17:32:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:10:32.566 17:32:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy concat 00:10:32.566 17:32:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:10:32.566 17:32:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@200 -- # return 1 00:10:32.566 17:32:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:10:32.566 17:32:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline concat 64 3 00:10:32.566 17:32:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:32.566 17:32:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:10:32.566 17:32:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:32.566 17:32:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:32.566 17:32:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:32.566 17:32:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:32.566 17:32:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:32.566 17:32:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:32.566 17:32:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:32.566 17:32:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:32.566 17:32:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:32.566 17:32:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:32.566 17:32:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:32.566 17:32:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:32.566 17:32:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:32.566 "name": "Existed_Raid", 00:10:32.566 "uuid": "6861168e-107c-47c8-b43a-e43f82713b21", 00:10:32.566 "strip_size_kb": 64, 00:10:32.566 "state": "offline", 00:10:32.566 "raid_level": "concat", 00:10:32.566 "superblock": false, 00:10:32.566 "num_base_bdevs": 4, 00:10:32.566 "num_base_bdevs_discovered": 3, 00:10:32.566 "num_base_bdevs_operational": 3, 00:10:32.566 "base_bdevs_list": [ 00:10:32.566 { 00:10:32.566 "name": null, 00:10:32.566 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:32.566 "is_configured": false, 00:10:32.566 "data_offset": 0, 00:10:32.566 "data_size": 65536 00:10:32.566 }, 00:10:32.566 { 00:10:32.566 "name": "BaseBdev2", 00:10:32.566 "uuid": "c5a98eb1-f908-4e70-9538-a95c29933019", 00:10:32.566 "is_configured": true, 00:10:32.566 "data_offset": 0, 00:10:32.566 "data_size": 65536 00:10:32.566 }, 00:10:32.566 { 00:10:32.566 "name": "BaseBdev3", 00:10:32.566 "uuid": "d0de4cf0-5a2b-4b2e-b49f-fab88d122f67", 00:10:32.566 "is_configured": true, 00:10:32.566 "data_offset": 0, 00:10:32.566 "data_size": 65536 00:10:32.566 }, 00:10:32.566 { 00:10:32.566 "name": "BaseBdev4", 00:10:32.566 "uuid": "c6c989da-f006-4684-a5b9-0d19718a177a", 00:10:32.566 "is_configured": true, 00:10:32.566 "data_offset": 0, 00:10:32.566 "data_size": 65536 00:10:32.566 } 00:10:32.566 ] 00:10:32.566 }' 00:10:32.566 17:32:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:32.566 17:32:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:32.826 17:32:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:10:32.826 17:32:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:32.826 17:32:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:32.826 17:32:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:32.826 17:32:03 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:10:32.826 17:32:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:32.826 17:32:03 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:32.826 17:32:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:10:32.826 17:32:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:10:32.826 17:32:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:10:32.826 17:32:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:32.826 17:32:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:32.826 [2024-11-27 17:32:04.014323] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:10:33.087 17:32:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:33.087 17:32:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:10:33.087 17:32:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:33.087 17:32:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:33.087 17:32:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:10:33.087 17:32:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:33.087 17:32:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:33.087 17:32:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:33.087 17:32:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:10:33.087 17:32:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:10:33.087 17:32:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:10:33.087 17:32:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:33.087 17:32:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:33.087 [2024-11-27 17:32:04.094718] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:10:33.087 17:32:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:33.087 17:32:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:10:33.087 17:32:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:33.087 17:32:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:33.087 17:32:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:10:33.087 17:32:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:33.087 17:32:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:33.087 17:32:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:33.087 17:32:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:10:33.087 17:32:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:10:33.087 17:32:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev4 00:10:33.087 17:32:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:33.087 17:32:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:33.087 [2024-11-27 17:32:04.171114] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev4 00:10:33.087 [2024-11-27 17:32:04.171226] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:10:33.087 17:32:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:33.087 17:32:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:10:33.087 17:32:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:33.087 17:32:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:33.087 17:32:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:10:33.087 17:32:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:33.087 17:32:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:33.087 17:32:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:33.087 17:32:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:10:33.087 17:32:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:10:33.087 17:32:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 4 -gt 2 ']' 00:10:33.087 17:32:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:10:33.087 17:32:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:33.087 17:32:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:10:33.087 17:32:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:33.087 17:32:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:33.087 BaseBdev2 00:10:33.087 17:32:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:33.087 17:32:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:10:33.087 17:32:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:10:33.087 17:32:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:10:33.087 17:32:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:10:33.087 17:32:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:10:33.087 17:32:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:10:33.087 17:32:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:10:33.087 17:32:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:33.087 17:32:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:33.087 17:32:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:33.087 17:32:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:10:33.087 17:32:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:33.348 17:32:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:33.348 [ 00:10:33.348 { 00:10:33.348 "name": "BaseBdev2", 00:10:33.348 "aliases": [ 00:10:33.348 "fd9d5026-b139-4e3f-af02-58d0de0516fd" 00:10:33.348 ], 00:10:33.348 "product_name": "Malloc disk", 00:10:33.348 "block_size": 512, 00:10:33.348 "num_blocks": 65536, 00:10:33.348 "uuid": "fd9d5026-b139-4e3f-af02-58d0de0516fd", 00:10:33.348 "assigned_rate_limits": { 00:10:33.348 "rw_ios_per_sec": 0, 00:10:33.348 "rw_mbytes_per_sec": 0, 00:10:33.348 "r_mbytes_per_sec": 0, 00:10:33.348 "w_mbytes_per_sec": 0 00:10:33.348 }, 00:10:33.348 "claimed": false, 00:10:33.348 "zoned": false, 00:10:33.348 "supported_io_types": { 00:10:33.348 "read": true, 00:10:33.348 "write": true, 00:10:33.348 "unmap": true, 00:10:33.348 "flush": true, 00:10:33.348 "reset": true, 00:10:33.348 "nvme_admin": false, 00:10:33.348 "nvme_io": false, 00:10:33.348 "nvme_io_md": false, 00:10:33.348 "write_zeroes": true, 00:10:33.348 "zcopy": true, 00:10:33.348 "get_zone_info": false, 00:10:33.348 "zone_management": false, 00:10:33.348 "zone_append": false, 00:10:33.348 "compare": false, 00:10:33.348 "compare_and_write": false, 00:10:33.348 "abort": true, 00:10:33.348 "seek_hole": false, 00:10:33.348 "seek_data": false, 00:10:33.348 "copy": true, 00:10:33.348 "nvme_iov_md": false 00:10:33.348 }, 00:10:33.348 "memory_domains": [ 00:10:33.348 { 00:10:33.348 "dma_device_id": "system", 00:10:33.348 "dma_device_type": 1 00:10:33.348 }, 00:10:33.348 { 00:10:33.348 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:33.348 "dma_device_type": 2 00:10:33.348 } 00:10:33.348 ], 00:10:33.348 "driver_specific": {} 00:10:33.348 } 00:10:33.348 ] 00:10:33.348 17:32:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:33.348 17:32:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:10:33.348 17:32:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:10:33.348 17:32:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:33.348 17:32:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:10:33.348 17:32:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:33.348 17:32:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:33.348 BaseBdev3 00:10:33.348 17:32:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:33.348 17:32:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:10:33.348 17:32:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:10:33.348 17:32:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:10:33.348 17:32:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:10:33.348 17:32:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:10:33.348 17:32:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:10:33.348 17:32:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:10:33.348 17:32:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:33.348 17:32:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:33.348 17:32:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:33.348 17:32:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:10:33.348 17:32:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:33.348 17:32:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:33.348 [ 00:10:33.348 { 00:10:33.348 "name": "BaseBdev3", 00:10:33.348 "aliases": [ 00:10:33.348 "2ff244e8-b27a-499f-a1a1-c74fd587554e" 00:10:33.348 ], 00:10:33.348 "product_name": "Malloc disk", 00:10:33.348 "block_size": 512, 00:10:33.348 "num_blocks": 65536, 00:10:33.348 "uuid": "2ff244e8-b27a-499f-a1a1-c74fd587554e", 00:10:33.348 "assigned_rate_limits": { 00:10:33.348 "rw_ios_per_sec": 0, 00:10:33.348 "rw_mbytes_per_sec": 0, 00:10:33.348 "r_mbytes_per_sec": 0, 00:10:33.348 "w_mbytes_per_sec": 0 00:10:33.348 }, 00:10:33.348 "claimed": false, 00:10:33.348 "zoned": false, 00:10:33.348 "supported_io_types": { 00:10:33.348 "read": true, 00:10:33.349 "write": true, 00:10:33.349 "unmap": true, 00:10:33.349 "flush": true, 00:10:33.349 "reset": true, 00:10:33.349 "nvme_admin": false, 00:10:33.349 "nvme_io": false, 00:10:33.349 "nvme_io_md": false, 00:10:33.349 "write_zeroes": true, 00:10:33.349 "zcopy": true, 00:10:33.349 "get_zone_info": false, 00:10:33.349 "zone_management": false, 00:10:33.349 "zone_append": false, 00:10:33.349 "compare": false, 00:10:33.349 "compare_and_write": false, 00:10:33.349 "abort": true, 00:10:33.349 "seek_hole": false, 00:10:33.349 "seek_data": false, 00:10:33.349 "copy": true, 00:10:33.349 "nvme_iov_md": false 00:10:33.349 }, 00:10:33.349 "memory_domains": [ 00:10:33.349 { 00:10:33.349 "dma_device_id": "system", 00:10:33.349 "dma_device_type": 1 00:10:33.349 }, 00:10:33.349 { 00:10:33.349 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:33.349 "dma_device_type": 2 00:10:33.349 } 00:10:33.349 ], 00:10:33.349 "driver_specific": {} 00:10:33.349 } 00:10:33.349 ] 00:10:33.349 17:32:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:33.349 17:32:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:10:33.349 17:32:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:10:33.349 17:32:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:33.349 17:32:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:10:33.349 17:32:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:33.349 17:32:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:33.349 BaseBdev4 00:10:33.349 17:32:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:33.349 17:32:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev4 00:10:33.349 17:32:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev4 00:10:33.349 17:32:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:10:33.349 17:32:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:10:33.349 17:32:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:10:33.349 17:32:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:10:33.349 17:32:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:10:33.349 17:32:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:33.349 17:32:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:33.349 17:32:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:33.349 17:32:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:10:33.349 17:32:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:33.349 17:32:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:33.349 [ 00:10:33.349 { 00:10:33.349 "name": "BaseBdev4", 00:10:33.349 "aliases": [ 00:10:33.349 "6981d8d6-7db7-4a2f-8d66-e15d009f92db" 00:10:33.349 ], 00:10:33.349 "product_name": "Malloc disk", 00:10:33.349 "block_size": 512, 00:10:33.349 "num_blocks": 65536, 00:10:33.349 "uuid": "6981d8d6-7db7-4a2f-8d66-e15d009f92db", 00:10:33.349 "assigned_rate_limits": { 00:10:33.349 "rw_ios_per_sec": 0, 00:10:33.349 "rw_mbytes_per_sec": 0, 00:10:33.349 "r_mbytes_per_sec": 0, 00:10:33.349 "w_mbytes_per_sec": 0 00:10:33.349 }, 00:10:33.349 "claimed": false, 00:10:33.349 "zoned": false, 00:10:33.349 "supported_io_types": { 00:10:33.349 "read": true, 00:10:33.349 "write": true, 00:10:33.349 "unmap": true, 00:10:33.349 "flush": true, 00:10:33.349 "reset": true, 00:10:33.349 "nvme_admin": false, 00:10:33.349 "nvme_io": false, 00:10:33.349 "nvme_io_md": false, 00:10:33.349 "write_zeroes": true, 00:10:33.349 "zcopy": true, 00:10:33.349 "get_zone_info": false, 00:10:33.349 "zone_management": false, 00:10:33.349 "zone_append": false, 00:10:33.349 "compare": false, 00:10:33.349 "compare_and_write": false, 00:10:33.349 "abort": true, 00:10:33.349 "seek_hole": false, 00:10:33.349 "seek_data": false, 00:10:33.349 "copy": true, 00:10:33.349 "nvme_iov_md": false 00:10:33.349 }, 00:10:33.349 "memory_domains": [ 00:10:33.349 { 00:10:33.349 "dma_device_id": "system", 00:10:33.349 "dma_device_type": 1 00:10:33.349 }, 00:10:33.349 { 00:10:33.349 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:33.349 "dma_device_type": 2 00:10:33.349 } 00:10:33.349 ], 00:10:33.349 "driver_specific": {} 00:10:33.349 } 00:10:33.349 ] 00:10:33.349 17:32:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:33.349 17:32:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:10:33.349 17:32:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:10:33.349 17:32:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:33.349 17:32:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:10:33.349 17:32:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:33.349 17:32:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:33.349 [2024-11-27 17:32:04.423498] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:10:33.349 [2024-11-27 17:32:04.423616] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:10:33.349 [2024-11-27 17:32:04.423675] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:33.349 [2024-11-27 17:32:04.425728] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:33.349 [2024-11-27 17:32:04.425828] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:10:33.349 17:32:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:33.349 17:32:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:33.349 17:32:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:33.349 17:32:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:33.349 17:32:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:33.350 17:32:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:33.350 17:32:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:33.350 17:32:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:33.350 17:32:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:33.350 17:32:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:33.350 17:32:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:33.350 17:32:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:33.350 17:32:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:33.350 17:32:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:33.350 17:32:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:33.350 17:32:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:33.350 17:32:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:33.350 "name": "Existed_Raid", 00:10:33.350 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:33.350 "strip_size_kb": 64, 00:10:33.350 "state": "configuring", 00:10:33.350 "raid_level": "concat", 00:10:33.350 "superblock": false, 00:10:33.350 "num_base_bdevs": 4, 00:10:33.350 "num_base_bdevs_discovered": 3, 00:10:33.350 "num_base_bdevs_operational": 4, 00:10:33.350 "base_bdevs_list": [ 00:10:33.350 { 00:10:33.350 "name": "BaseBdev1", 00:10:33.350 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:33.350 "is_configured": false, 00:10:33.350 "data_offset": 0, 00:10:33.350 "data_size": 0 00:10:33.350 }, 00:10:33.350 { 00:10:33.350 "name": "BaseBdev2", 00:10:33.350 "uuid": "fd9d5026-b139-4e3f-af02-58d0de0516fd", 00:10:33.350 "is_configured": true, 00:10:33.350 "data_offset": 0, 00:10:33.350 "data_size": 65536 00:10:33.350 }, 00:10:33.350 { 00:10:33.350 "name": "BaseBdev3", 00:10:33.350 "uuid": "2ff244e8-b27a-499f-a1a1-c74fd587554e", 00:10:33.350 "is_configured": true, 00:10:33.350 "data_offset": 0, 00:10:33.350 "data_size": 65536 00:10:33.350 }, 00:10:33.350 { 00:10:33.350 "name": "BaseBdev4", 00:10:33.350 "uuid": "6981d8d6-7db7-4a2f-8d66-e15d009f92db", 00:10:33.350 "is_configured": true, 00:10:33.350 "data_offset": 0, 00:10:33.350 "data_size": 65536 00:10:33.350 } 00:10:33.350 ] 00:10:33.350 }' 00:10:33.350 17:32:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:33.350 17:32:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:33.920 17:32:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:10:33.920 17:32:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:33.920 17:32:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:33.920 [2024-11-27 17:32:04.894690] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:10:33.920 17:32:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:33.920 17:32:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:33.920 17:32:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:33.920 17:32:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:33.920 17:32:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:33.920 17:32:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:33.920 17:32:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:33.920 17:32:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:33.920 17:32:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:33.920 17:32:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:33.920 17:32:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:33.920 17:32:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:33.920 17:32:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:33.920 17:32:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:33.920 17:32:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:33.920 17:32:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:33.920 17:32:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:33.920 "name": "Existed_Raid", 00:10:33.920 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:33.920 "strip_size_kb": 64, 00:10:33.920 "state": "configuring", 00:10:33.920 "raid_level": "concat", 00:10:33.920 "superblock": false, 00:10:33.920 "num_base_bdevs": 4, 00:10:33.920 "num_base_bdevs_discovered": 2, 00:10:33.920 "num_base_bdevs_operational": 4, 00:10:33.920 "base_bdevs_list": [ 00:10:33.920 { 00:10:33.920 "name": "BaseBdev1", 00:10:33.920 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:33.920 "is_configured": false, 00:10:33.920 "data_offset": 0, 00:10:33.920 "data_size": 0 00:10:33.920 }, 00:10:33.920 { 00:10:33.920 "name": null, 00:10:33.920 "uuid": "fd9d5026-b139-4e3f-af02-58d0de0516fd", 00:10:33.920 "is_configured": false, 00:10:33.920 "data_offset": 0, 00:10:33.920 "data_size": 65536 00:10:33.920 }, 00:10:33.920 { 00:10:33.920 "name": "BaseBdev3", 00:10:33.920 "uuid": "2ff244e8-b27a-499f-a1a1-c74fd587554e", 00:10:33.920 "is_configured": true, 00:10:33.920 "data_offset": 0, 00:10:33.920 "data_size": 65536 00:10:33.920 }, 00:10:33.920 { 00:10:33.920 "name": "BaseBdev4", 00:10:33.920 "uuid": "6981d8d6-7db7-4a2f-8d66-e15d009f92db", 00:10:33.920 "is_configured": true, 00:10:33.920 "data_offset": 0, 00:10:33.920 "data_size": 65536 00:10:33.920 } 00:10:33.920 ] 00:10:33.920 }' 00:10:33.920 17:32:04 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:33.920 17:32:04 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:34.180 17:32:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:34.180 17:32:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:10:34.180 17:32:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:34.180 17:32:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:34.181 17:32:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:34.181 17:32:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:10:34.181 17:32:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:10:34.181 17:32:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:34.181 17:32:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:34.181 [2024-11-27 17:32:05.354711] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:34.181 BaseBdev1 00:10:34.181 17:32:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:34.181 17:32:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:10:34.181 17:32:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:10:34.181 17:32:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:10:34.181 17:32:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:10:34.181 17:32:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:10:34.181 17:32:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:10:34.181 17:32:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:10:34.181 17:32:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:34.181 17:32:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:34.181 17:32:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:34.181 17:32:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:10:34.181 17:32:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:34.181 17:32:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:34.441 [ 00:10:34.441 { 00:10:34.441 "name": "BaseBdev1", 00:10:34.441 "aliases": [ 00:10:34.441 "b0d9a64b-d334-487b-b3cf-076573d6bcf9" 00:10:34.441 ], 00:10:34.441 "product_name": "Malloc disk", 00:10:34.441 "block_size": 512, 00:10:34.441 "num_blocks": 65536, 00:10:34.441 "uuid": "b0d9a64b-d334-487b-b3cf-076573d6bcf9", 00:10:34.441 "assigned_rate_limits": { 00:10:34.441 "rw_ios_per_sec": 0, 00:10:34.441 "rw_mbytes_per_sec": 0, 00:10:34.441 "r_mbytes_per_sec": 0, 00:10:34.441 "w_mbytes_per_sec": 0 00:10:34.441 }, 00:10:34.441 "claimed": true, 00:10:34.441 "claim_type": "exclusive_write", 00:10:34.441 "zoned": false, 00:10:34.441 "supported_io_types": { 00:10:34.441 "read": true, 00:10:34.441 "write": true, 00:10:34.441 "unmap": true, 00:10:34.441 "flush": true, 00:10:34.441 "reset": true, 00:10:34.441 "nvme_admin": false, 00:10:34.441 "nvme_io": false, 00:10:34.441 "nvme_io_md": false, 00:10:34.441 "write_zeroes": true, 00:10:34.441 "zcopy": true, 00:10:34.441 "get_zone_info": false, 00:10:34.441 "zone_management": false, 00:10:34.441 "zone_append": false, 00:10:34.441 "compare": false, 00:10:34.441 "compare_and_write": false, 00:10:34.441 "abort": true, 00:10:34.441 "seek_hole": false, 00:10:34.441 "seek_data": false, 00:10:34.441 "copy": true, 00:10:34.441 "nvme_iov_md": false 00:10:34.441 }, 00:10:34.441 "memory_domains": [ 00:10:34.441 { 00:10:34.441 "dma_device_id": "system", 00:10:34.441 "dma_device_type": 1 00:10:34.441 }, 00:10:34.441 { 00:10:34.441 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:34.441 "dma_device_type": 2 00:10:34.441 } 00:10:34.441 ], 00:10:34.441 "driver_specific": {} 00:10:34.441 } 00:10:34.441 ] 00:10:34.441 17:32:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:34.441 17:32:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:10:34.441 17:32:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:34.441 17:32:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:34.441 17:32:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:34.441 17:32:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:34.441 17:32:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:34.441 17:32:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:34.441 17:32:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:34.441 17:32:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:34.441 17:32:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:34.441 17:32:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:34.441 17:32:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:34.441 17:32:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:34.441 17:32:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:34.441 17:32:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:34.441 17:32:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:34.441 17:32:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:34.441 "name": "Existed_Raid", 00:10:34.441 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:34.441 "strip_size_kb": 64, 00:10:34.441 "state": "configuring", 00:10:34.441 "raid_level": "concat", 00:10:34.441 "superblock": false, 00:10:34.441 "num_base_bdevs": 4, 00:10:34.441 "num_base_bdevs_discovered": 3, 00:10:34.441 "num_base_bdevs_operational": 4, 00:10:34.441 "base_bdevs_list": [ 00:10:34.441 { 00:10:34.441 "name": "BaseBdev1", 00:10:34.441 "uuid": "b0d9a64b-d334-487b-b3cf-076573d6bcf9", 00:10:34.441 "is_configured": true, 00:10:34.441 "data_offset": 0, 00:10:34.441 "data_size": 65536 00:10:34.441 }, 00:10:34.441 { 00:10:34.441 "name": null, 00:10:34.441 "uuid": "fd9d5026-b139-4e3f-af02-58d0de0516fd", 00:10:34.441 "is_configured": false, 00:10:34.441 "data_offset": 0, 00:10:34.441 "data_size": 65536 00:10:34.441 }, 00:10:34.441 { 00:10:34.441 "name": "BaseBdev3", 00:10:34.441 "uuid": "2ff244e8-b27a-499f-a1a1-c74fd587554e", 00:10:34.441 "is_configured": true, 00:10:34.441 "data_offset": 0, 00:10:34.441 "data_size": 65536 00:10:34.441 }, 00:10:34.441 { 00:10:34.441 "name": "BaseBdev4", 00:10:34.441 "uuid": "6981d8d6-7db7-4a2f-8d66-e15d009f92db", 00:10:34.441 "is_configured": true, 00:10:34.441 "data_offset": 0, 00:10:34.441 "data_size": 65536 00:10:34.441 } 00:10:34.441 ] 00:10:34.441 }' 00:10:34.441 17:32:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:34.441 17:32:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:34.701 17:32:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:34.701 17:32:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:34.701 17:32:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:34.702 17:32:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:10:34.702 17:32:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:34.702 17:32:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:10:34.702 17:32:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:10:34.702 17:32:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:34.702 17:32:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:34.962 [2024-11-27 17:32:05.893789] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:10:34.962 17:32:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:34.962 17:32:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:34.962 17:32:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:34.962 17:32:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:34.962 17:32:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:34.962 17:32:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:34.962 17:32:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:34.962 17:32:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:34.962 17:32:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:34.962 17:32:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:34.962 17:32:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:34.962 17:32:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:34.962 17:32:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:34.962 17:32:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:34.962 17:32:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:34.962 17:32:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:34.962 17:32:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:34.962 "name": "Existed_Raid", 00:10:34.962 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:34.962 "strip_size_kb": 64, 00:10:34.962 "state": "configuring", 00:10:34.962 "raid_level": "concat", 00:10:34.962 "superblock": false, 00:10:34.962 "num_base_bdevs": 4, 00:10:34.962 "num_base_bdevs_discovered": 2, 00:10:34.962 "num_base_bdevs_operational": 4, 00:10:34.962 "base_bdevs_list": [ 00:10:34.962 { 00:10:34.962 "name": "BaseBdev1", 00:10:34.962 "uuid": "b0d9a64b-d334-487b-b3cf-076573d6bcf9", 00:10:34.962 "is_configured": true, 00:10:34.962 "data_offset": 0, 00:10:34.962 "data_size": 65536 00:10:34.962 }, 00:10:34.962 { 00:10:34.962 "name": null, 00:10:34.962 "uuid": "fd9d5026-b139-4e3f-af02-58d0de0516fd", 00:10:34.962 "is_configured": false, 00:10:34.962 "data_offset": 0, 00:10:34.962 "data_size": 65536 00:10:34.962 }, 00:10:34.962 { 00:10:34.962 "name": null, 00:10:34.962 "uuid": "2ff244e8-b27a-499f-a1a1-c74fd587554e", 00:10:34.962 "is_configured": false, 00:10:34.962 "data_offset": 0, 00:10:34.962 "data_size": 65536 00:10:34.962 }, 00:10:34.962 { 00:10:34.962 "name": "BaseBdev4", 00:10:34.962 "uuid": "6981d8d6-7db7-4a2f-8d66-e15d009f92db", 00:10:34.962 "is_configured": true, 00:10:34.962 "data_offset": 0, 00:10:34.962 "data_size": 65536 00:10:34.962 } 00:10:34.962 ] 00:10:34.962 }' 00:10:34.962 17:32:05 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:34.962 17:32:05 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:35.222 17:32:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:10:35.222 17:32:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:35.222 17:32:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:35.222 17:32:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:35.222 17:32:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:35.222 17:32:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:10:35.222 17:32:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:10:35.222 17:32:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:35.222 17:32:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:35.222 [2024-11-27 17:32:06.389008] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:35.222 17:32:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:35.222 17:32:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:35.222 17:32:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:35.222 17:32:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:35.222 17:32:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:35.222 17:32:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:35.222 17:32:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:35.222 17:32:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:35.222 17:32:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:35.222 17:32:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:35.222 17:32:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:35.222 17:32:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:35.222 17:32:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:35.222 17:32:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:35.222 17:32:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:35.482 17:32:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:35.482 17:32:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:35.482 "name": "Existed_Raid", 00:10:35.482 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:35.482 "strip_size_kb": 64, 00:10:35.482 "state": "configuring", 00:10:35.482 "raid_level": "concat", 00:10:35.482 "superblock": false, 00:10:35.482 "num_base_bdevs": 4, 00:10:35.482 "num_base_bdevs_discovered": 3, 00:10:35.482 "num_base_bdevs_operational": 4, 00:10:35.482 "base_bdevs_list": [ 00:10:35.482 { 00:10:35.482 "name": "BaseBdev1", 00:10:35.482 "uuid": "b0d9a64b-d334-487b-b3cf-076573d6bcf9", 00:10:35.482 "is_configured": true, 00:10:35.482 "data_offset": 0, 00:10:35.482 "data_size": 65536 00:10:35.482 }, 00:10:35.482 { 00:10:35.482 "name": null, 00:10:35.482 "uuid": "fd9d5026-b139-4e3f-af02-58d0de0516fd", 00:10:35.482 "is_configured": false, 00:10:35.482 "data_offset": 0, 00:10:35.482 "data_size": 65536 00:10:35.482 }, 00:10:35.482 { 00:10:35.482 "name": "BaseBdev3", 00:10:35.482 "uuid": "2ff244e8-b27a-499f-a1a1-c74fd587554e", 00:10:35.482 "is_configured": true, 00:10:35.482 "data_offset": 0, 00:10:35.482 "data_size": 65536 00:10:35.482 }, 00:10:35.482 { 00:10:35.482 "name": "BaseBdev4", 00:10:35.482 "uuid": "6981d8d6-7db7-4a2f-8d66-e15d009f92db", 00:10:35.482 "is_configured": true, 00:10:35.482 "data_offset": 0, 00:10:35.482 "data_size": 65536 00:10:35.482 } 00:10:35.482 ] 00:10:35.482 }' 00:10:35.482 17:32:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:35.482 17:32:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:35.742 17:32:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:10:35.742 17:32:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:35.742 17:32:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:35.742 17:32:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:35.742 17:32:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:35.742 17:32:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:10:35.742 17:32:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:10:35.742 17:32:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:35.742 17:32:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:35.742 [2024-11-27 17:32:06.860202] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:10:35.742 17:32:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:35.742 17:32:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:35.742 17:32:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:35.742 17:32:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:35.742 17:32:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:35.742 17:32:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:35.742 17:32:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:35.742 17:32:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:35.742 17:32:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:35.742 17:32:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:35.742 17:32:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:35.742 17:32:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:35.742 17:32:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:35.742 17:32:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:35.742 17:32:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:35.742 17:32:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:36.002 17:32:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:36.002 "name": "Existed_Raid", 00:10:36.002 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:36.002 "strip_size_kb": 64, 00:10:36.002 "state": "configuring", 00:10:36.002 "raid_level": "concat", 00:10:36.002 "superblock": false, 00:10:36.002 "num_base_bdevs": 4, 00:10:36.002 "num_base_bdevs_discovered": 2, 00:10:36.002 "num_base_bdevs_operational": 4, 00:10:36.002 "base_bdevs_list": [ 00:10:36.002 { 00:10:36.002 "name": null, 00:10:36.002 "uuid": "b0d9a64b-d334-487b-b3cf-076573d6bcf9", 00:10:36.002 "is_configured": false, 00:10:36.002 "data_offset": 0, 00:10:36.002 "data_size": 65536 00:10:36.002 }, 00:10:36.002 { 00:10:36.002 "name": null, 00:10:36.002 "uuid": "fd9d5026-b139-4e3f-af02-58d0de0516fd", 00:10:36.002 "is_configured": false, 00:10:36.002 "data_offset": 0, 00:10:36.002 "data_size": 65536 00:10:36.002 }, 00:10:36.002 { 00:10:36.002 "name": "BaseBdev3", 00:10:36.002 "uuid": "2ff244e8-b27a-499f-a1a1-c74fd587554e", 00:10:36.002 "is_configured": true, 00:10:36.002 "data_offset": 0, 00:10:36.002 "data_size": 65536 00:10:36.002 }, 00:10:36.002 { 00:10:36.002 "name": "BaseBdev4", 00:10:36.002 "uuid": "6981d8d6-7db7-4a2f-8d66-e15d009f92db", 00:10:36.002 "is_configured": true, 00:10:36.002 "data_offset": 0, 00:10:36.002 "data_size": 65536 00:10:36.002 } 00:10:36.002 ] 00:10:36.002 }' 00:10:36.002 17:32:06 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:36.002 17:32:06 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:36.262 17:32:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:36.262 17:32:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:36.262 17:32:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:36.262 17:32:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:10:36.262 17:32:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:36.262 17:32:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:10:36.262 17:32:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:10:36.262 17:32:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:36.262 17:32:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:36.262 [2024-11-27 17:32:07.383125] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:36.262 17:32:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:36.262 17:32:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:36.262 17:32:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:36.262 17:32:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:36.262 17:32:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:36.262 17:32:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:36.262 17:32:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:36.262 17:32:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:36.262 17:32:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:36.262 17:32:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:36.262 17:32:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:36.262 17:32:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:36.262 17:32:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:36.262 17:32:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:36.262 17:32:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:36.262 17:32:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:36.262 17:32:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:36.262 "name": "Existed_Raid", 00:10:36.262 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:36.262 "strip_size_kb": 64, 00:10:36.262 "state": "configuring", 00:10:36.262 "raid_level": "concat", 00:10:36.262 "superblock": false, 00:10:36.262 "num_base_bdevs": 4, 00:10:36.262 "num_base_bdevs_discovered": 3, 00:10:36.262 "num_base_bdevs_operational": 4, 00:10:36.262 "base_bdevs_list": [ 00:10:36.262 { 00:10:36.262 "name": null, 00:10:36.262 "uuid": "b0d9a64b-d334-487b-b3cf-076573d6bcf9", 00:10:36.262 "is_configured": false, 00:10:36.262 "data_offset": 0, 00:10:36.262 "data_size": 65536 00:10:36.262 }, 00:10:36.262 { 00:10:36.262 "name": "BaseBdev2", 00:10:36.262 "uuid": "fd9d5026-b139-4e3f-af02-58d0de0516fd", 00:10:36.262 "is_configured": true, 00:10:36.262 "data_offset": 0, 00:10:36.262 "data_size": 65536 00:10:36.262 }, 00:10:36.262 { 00:10:36.262 "name": "BaseBdev3", 00:10:36.262 "uuid": "2ff244e8-b27a-499f-a1a1-c74fd587554e", 00:10:36.262 "is_configured": true, 00:10:36.262 "data_offset": 0, 00:10:36.262 "data_size": 65536 00:10:36.262 }, 00:10:36.262 { 00:10:36.262 "name": "BaseBdev4", 00:10:36.262 "uuid": "6981d8d6-7db7-4a2f-8d66-e15d009f92db", 00:10:36.262 "is_configured": true, 00:10:36.262 "data_offset": 0, 00:10:36.262 "data_size": 65536 00:10:36.262 } 00:10:36.262 ] 00:10:36.262 }' 00:10:36.262 17:32:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:36.262 17:32:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:36.831 17:32:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:36.831 17:32:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:36.831 17:32:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:10:36.831 17:32:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:36.831 17:32:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:36.831 17:32:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:10:36.831 17:32:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:10:36.831 17:32:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:36.831 17:32:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:36.831 17:32:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:36.831 17:32:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:36.831 17:32:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u b0d9a64b-d334-487b-b3cf-076573d6bcf9 00:10:36.831 17:32:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:36.831 17:32:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:36.831 [2024-11-27 17:32:07.951087] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:10:36.831 [2024-11-27 17:32:07.951139] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:10:36.831 [2024-11-27 17:32:07.951147] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 262144, blocklen 512 00:10:36.831 [2024-11-27 17:32:07.951463] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002a10 00:10:36.831 [2024-11-27 17:32:07.951618] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:10:36.831 [2024-11-27 17:32:07.951636] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001c80 00:10:36.831 [2024-11-27 17:32:07.951845] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:36.831 NewBaseBdev 00:10:36.831 17:32:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:36.831 17:32:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:10:36.831 17:32:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=NewBaseBdev 00:10:36.831 17:32:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:10:36.831 17:32:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:10:36.831 17:32:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:10:36.831 17:32:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:10:36.831 17:32:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:10:36.831 17:32:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:36.831 17:32:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:36.831 17:32:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:36.831 17:32:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:10:36.831 17:32:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:36.831 17:32:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:36.831 [ 00:10:36.831 { 00:10:36.831 "name": "NewBaseBdev", 00:10:36.831 "aliases": [ 00:10:36.831 "b0d9a64b-d334-487b-b3cf-076573d6bcf9" 00:10:36.831 ], 00:10:36.831 "product_name": "Malloc disk", 00:10:36.831 "block_size": 512, 00:10:36.831 "num_blocks": 65536, 00:10:36.831 "uuid": "b0d9a64b-d334-487b-b3cf-076573d6bcf9", 00:10:36.831 "assigned_rate_limits": { 00:10:36.831 "rw_ios_per_sec": 0, 00:10:36.831 "rw_mbytes_per_sec": 0, 00:10:36.831 "r_mbytes_per_sec": 0, 00:10:36.831 "w_mbytes_per_sec": 0 00:10:36.831 }, 00:10:36.831 "claimed": true, 00:10:36.831 "claim_type": "exclusive_write", 00:10:36.831 "zoned": false, 00:10:36.831 "supported_io_types": { 00:10:36.831 "read": true, 00:10:36.831 "write": true, 00:10:36.831 "unmap": true, 00:10:36.831 "flush": true, 00:10:36.831 "reset": true, 00:10:36.831 "nvme_admin": false, 00:10:36.831 "nvme_io": false, 00:10:36.831 "nvme_io_md": false, 00:10:36.831 "write_zeroes": true, 00:10:36.831 "zcopy": true, 00:10:36.831 "get_zone_info": false, 00:10:36.831 "zone_management": false, 00:10:36.831 "zone_append": false, 00:10:36.831 "compare": false, 00:10:36.831 "compare_and_write": false, 00:10:36.831 "abort": true, 00:10:36.831 "seek_hole": false, 00:10:36.831 "seek_data": false, 00:10:36.831 "copy": true, 00:10:36.831 "nvme_iov_md": false 00:10:36.831 }, 00:10:36.831 "memory_domains": [ 00:10:36.831 { 00:10:36.831 "dma_device_id": "system", 00:10:36.831 "dma_device_type": 1 00:10:36.831 }, 00:10:36.831 { 00:10:36.831 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:36.831 "dma_device_type": 2 00:10:36.831 } 00:10:36.831 ], 00:10:36.831 "driver_specific": {} 00:10:36.831 } 00:10:36.831 ] 00:10:36.831 17:32:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:36.831 17:32:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:10:36.831 17:32:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online concat 64 4 00:10:36.831 17:32:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:36.831 17:32:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:36.831 17:32:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:36.831 17:32:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:36.831 17:32:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:36.831 17:32:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:36.831 17:32:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:36.831 17:32:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:36.831 17:32:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:36.831 17:32:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:36.831 17:32:07 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:36.831 17:32:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:36.831 17:32:07 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:36.831 17:32:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:37.092 17:32:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:37.092 "name": "Existed_Raid", 00:10:37.092 "uuid": "7d96a431-31c7-4ee8-8114-5b7644e1e917", 00:10:37.092 "strip_size_kb": 64, 00:10:37.092 "state": "online", 00:10:37.092 "raid_level": "concat", 00:10:37.092 "superblock": false, 00:10:37.092 "num_base_bdevs": 4, 00:10:37.092 "num_base_bdevs_discovered": 4, 00:10:37.092 "num_base_bdevs_operational": 4, 00:10:37.092 "base_bdevs_list": [ 00:10:37.092 { 00:10:37.092 "name": "NewBaseBdev", 00:10:37.092 "uuid": "b0d9a64b-d334-487b-b3cf-076573d6bcf9", 00:10:37.092 "is_configured": true, 00:10:37.092 "data_offset": 0, 00:10:37.092 "data_size": 65536 00:10:37.092 }, 00:10:37.092 { 00:10:37.092 "name": "BaseBdev2", 00:10:37.092 "uuid": "fd9d5026-b139-4e3f-af02-58d0de0516fd", 00:10:37.092 "is_configured": true, 00:10:37.092 "data_offset": 0, 00:10:37.092 "data_size": 65536 00:10:37.092 }, 00:10:37.092 { 00:10:37.092 "name": "BaseBdev3", 00:10:37.092 "uuid": "2ff244e8-b27a-499f-a1a1-c74fd587554e", 00:10:37.092 "is_configured": true, 00:10:37.092 "data_offset": 0, 00:10:37.092 "data_size": 65536 00:10:37.092 }, 00:10:37.092 { 00:10:37.092 "name": "BaseBdev4", 00:10:37.092 "uuid": "6981d8d6-7db7-4a2f-8d66-e15d009f92db", 00:10:37.092 "is_configured": true, 00:10:37.092 "data_offset": 0, 00:10:37.092 "data_size": 65536 00:10:37.092 } 00:10:37.092 ] 00:10:37.092 }' 00:10:37.092 17:32:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:37.092 17:32:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:37.353 17:32:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:10:37.353 17:32:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:10:37.353 17:32:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:10:37.353 17:32:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:10:37.353 17:32:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:10:37.353 17:32:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:10:37.353 17:32:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:10:37.353 17:32:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:37.353 17:32:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:37.353 17:32:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:10:37.353 [2024-11-27 17:32:08.422712] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:37.353 17:32:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:37.353 17:32:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:37.353 "name": "Existed_Raid", 00:10:37.353 "aliases": [ 00:10:37.353 "7d96a431-31c7-4ee8-8114-5b7644e1e917" 00:10:37.353 ], 00:10:37.353 "product_name": "Raid Volume", 00:10:37.353 "block_size": 512, 00:10:37.353 "num_blocks": 262144, 00:10:37.353 "uuid": "7d96a431-31c7-4ee8-8114-5b7644e1e917", 00:10:37.353 "assigned_rate_limits": { 00:10:37.353 "rw_ios_per_sec": 0, 00:10:37.353 "rw_mbytes_per_sec": 0, 00:10:37.353 "r_mbytes_per_sec": 0, 00:10:37.353 "w_mbytes_per_sec": 0 00:10:37.353 }, 00:10:37.353 "claimed": false, 00:10:37.353 "zoned": false, 00:10:37.353 "supported_io_types": { 00:10:37.353 "read": true, 00:10:37.353 "write": true, 00:10:37.353 "unmap": true, 00:10:37.353 "flush": true, 00:10:37.353 "reset": true, 00:10:37.353 "nvme_admin": false, 00:10:37.353 "nvme_io": false, 00:10:37.353 "nvme_io_md": false, 00:10:37.353 "write_zeroes": true, 00:10:37.353 "zcopy": false, 00:10:37.353 "get_zone_info": false, 00:10:37.353 "zone_management": false, 00:10:37.353 "zone_append": false, 00:10:37.353 "compare": false, 00:10:37.353 "compare_and_write": false, 00:10:37.353 "abort": false, 00:10:37.353 "seek_hole": false, 00:10:37.353 "seek_data": false, 00:10:37.353 "copy": false, 00:10:37.353 "nvme_iov_md": false 00:10:37.353 }, 00:10:37.353 "memory_domains": [ 00:10:37.353 { 00:10:37.353 "dma_device_id": "system", 00:10:37.353 "dma_device_type": 1 00:10:37.353 }, 00:10:37.353 { 00:10:37.353 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:37.353 "dma_device_type": 2 00:10:37.353 }, 00:10:37.353 { 00:10:37.353 "dma_device_id": "system", 00:10:37.353 "dma_device_type": 1 00:10:37.353 }, 00:10:37.353 { 00:10:37.353 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:37.353 "dma_device_type": 2 00:10:37.353 }, 00:10:37.353 { 00:10:37.353 "dma_device_id": "system", 00:10:37.353 "dma_device_type": 1 00:10:37.353 }, 00:10:37.353 { 00:10:37.353 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:37.353 "dma_device_type": 2 00:10:37.353 }, 00:10:37.353 { 00:10:37.353 "dma_device_id": "system", 00:10:37.353 "dma_device_type": 1 00:10:37.353 }, 00:10:37.353 { 00:10:37.353 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:37.353 "dma_device_type": 2 00:10:37.353 } 00:10:37.353 ], 00:10:37.353 "driver_specific": { 00:10:37.353 "raid": { 00:10:37.353 "uuid": "7d96a431-31c7-4ee8-8114-5b7644e1e917", 00:10:37.353 "strip_size_kb": 64, 00:10:37.353 "state": "online", 00:10:37.353 "raid_level": "concat", 00:10:37.353 "superblock": false, 00:10:37.353 "num_base_bdevs": 4, 00:10:37.353 "num_base_bdevs_discovered": 4, 00:10:37.353 "num_base_bdevs_operational": 4, 00:10:37.353 "base_bdevs_list": [ 00:10:37.353 { 00:10:37.353 "name": "NewBaseBdev", 00:10:37.353 "uuid": "b0d9a64b-d334-487b-b3cf-076573d6bcf9", 00:10:37.353 "is_configured": true, 00:10:37.353 "data_offset": 0, 00:10:37.353 "data_size": 65536 00:10:37.353 }, 00:10:37.353 { 00:10:37.353 "name": "BaseBdev2", 00:10:37.353 "uuid": "fd9d5026-b139-4e3f-af02-58d0de0516fd", 00:10:37.353 "is_configured": true, 00:10:37.353 "data_offset": 0, 00:10:37.353 "data_size": 65536 00:10:37.353 }, 00:10:37.353 { 00:10:37.353 "name": "BaseBdev3", 00:10:37.353 "uuid": "2ff244e8-b27a-499f-a1a1-c74fd587554e", 00:10:37.353 "is_configured": true, 00:10:37.353 "data_offset": 0, 00:10:37.353 "data_size": 65536 00:10:37.353 }, 00:10:37.353 { 00:10:37.353 "name": "BaseBdev4", 00:10:37.353 "uuid": "6981d8d6-7db7-4a2f-8d66-e15d009f92db", 00:10:37.353 "is_configured": true, 00:10:37.353 "data_offset": 0, 00:10:37.353 "data_size": 65536 00:10:37.353 } 00:10:37.353 ] 00:10:37.353 } 00:10:37.353 } 00:10:37.353 }' 00:10:37.353 17:32:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:37.353 17:32:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:10:37.353 BaseBdev2 00:10:37.353 BaseBdev3 00:10:37.353 BaseBdev4' 00:10:37.353 17:32:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:37.614 17:32:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:37.614 17:32:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:37.614 17:32:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:10:37.614 17:32:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:37.614 17:32:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:37.614 17:32:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:37.614 17:32:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:37.614 17:32:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:37.614 17:32:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:37.614 17:32:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:37.614 17:32:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:37.614 17:32:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:10:37.614 17:32:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:37.614 17:32:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:37.614 17:32:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:37.614 17:32:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:37.614 17:32:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:37.614 17:32:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:37.614 17:32:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:10:37.614 17:32:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:37.614 17:32:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:37.614 17:32:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:37.614 17:32:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:37.614 17:32:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:37.614 17:32:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:37.614 17:32:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:37.614 17:32:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:37.614 17:32:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:10:37.614 17:32:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:37.614 17:32:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:37.614 17:32:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:37.614 17:32:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:37.614 17:32:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:37.614 17:32:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:10:37.614 17:32:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:37.614 17:32:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:37.614 [2024-11-27 17:32:08.757830] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:10:37.614 [2024-11-27 17:32:08.757860] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:37.614 [2024-11-27 17:32:08.757947] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:37.614 [2024-11-27 17:32:08.758023] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:37.614 [2024-11-27 17:32:08.758033] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name Existed_Raid, state offline 00:10:37.614 17:32:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:37.614 17:32:08 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 82020 00:10:37.614 17:32:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@950 -- # '[' -z 82020 ']' 00:10:37.614 17:32:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # kill -0 82020 00:10:37.614 17:32:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@955 -- # uname 00:10:37.614 17:32:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:10:37.614 17:32:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 82020 00:10:37.874 17:32:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:10:37.874 17:32:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:10:37.874 17:32:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 82020' 00:10:37.874 killing process with pid 82020 00:10:37.874 17:32:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@969 -- # kill 82020 00:10:37.874 [2024-11-27 17:32:08.808289] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:10:37.874 17:32:08 bdev_raid.raid_state_function_test -- common/autotest_common.sh@974 -- # wait 82020 00:10:37.874 [2024-11-27 17:32:08.883257] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:10:38.134 17:32:09 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:10:38.134 00:10:38.134 real 0m9.782s 00:10:38.134 user 0m16.426s 00:10:38.134 sys 0m2.099s 00:10:38.134 17:32:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:10:38.134 17:32:09 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:38.134 ************************************ 00:10:38.134 END TEST raid_state_function_test 00:10:38.134 ************************************ 00:10:38.134 17:32:09 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test concat 4 true 00:10:38.134 17:32:09 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:10:38.134 17:32:09 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:10:38.134 17:32:09 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:10:38.394 ************************************ 00:10:38.394 START TEST raid_state_function_test_sb 00:10:38.394 ************************************ 00:10:38.394 17:32:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1125 -- # raid_state_function_test concat 4 true 00:10:38.394 17:32:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=concat 00:10:38.394 17:32:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=4 00:10:38.394 17:32:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:10:38.394 17:32:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:10:38.394 17:32:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:10:38.394 17:32:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:38.394 17:32:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:10:38.394 17:32:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:38.394 17:32:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:38.394 17:32:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:10:38.394 17:32:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:38.394 17:32:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:38.394 17:32:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:10:38.394 17:32:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:38.394 17:32:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:38.394 17:32:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev4 00:10:38.394 17:32:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:38.395 17:32:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:38.395 17:32:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:10:38.395 17:32:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:10:38.395 17:32:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:10:38.395 17:32:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:10:38.395 17:32:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:10:38.395 17:32:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:10:38.395 17:32:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' concat '!=' raid1 ']' 00:10:38.395 17:32:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:10:38.395 17:32:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:10:38.395 17:32:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:10:38.395 17:32:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:10:38.395 17:32:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=82675 00:10:38.395 17:32:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:10:38.395 17:32:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 82675' 00:10:38.395 Process raid pid: 82675 00:10:38.395 17:32:09 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 82675 00:10:38.395 17:32:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@831 -- # '[' -z 82675 ']' 00:10:38.395 17:32:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:38.395 17:32:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@836 -- # local max_retries=100 00:10:38.395 17:32:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:38.395 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:38.395 17:32:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # xtrace_disable 00:10:38.395 17:32:09 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:38.395 [2024-11-27 17:32:09.430497] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:10:38.395 [2024-11-27 17:32:09.430696] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:10:38.395 [2024-11-27 17:32:09.560742] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:38.655 [2024-11-27 17:32:09.628732] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:10:38.655 [2024-11-27 17:32:09.704295] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:38.655 [2024-11-27 17:32:09.704427] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:39.225 17:32:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:10:39.225 17:32:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@864 -- # return 0 00:10:39.225 17:32:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:10:39.225 17:32:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:39.225 17:32:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:39.225 [2024-11-27 17:32:10.255415] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:10:39.225 [2024-11-27 17:32:10.255470] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:10:39.225 [2024-11-27 17:32:10.255507] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:10:39.225 [2024-11-27 17:32:10.255518] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:10:39.225 [2024-11-27 17:32:10.255524] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:10:39.225 [2024-11-27 17:32:10.255539] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:10:39.225 [2024-11-27 17:32:10.255545] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:10:39.225 [2024-11-27 17:32:10.255553] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:10:39.225 17:32:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:39.225 17:32:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:39.225 17:32:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:39.225 17:32:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:39.225 17:32:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:39.225 17:32:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:39.225 17:32:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:39.225 17:32:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:39.225 17:32:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:39.225 17:32:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:39.225 17:32:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:39.225 17:32:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:39.225 17:32:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:39.225 17:32:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:39.225 17:32:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:39.225 17:32:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:39.225 17:32:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:39.225 "name": "Existed_Raid", 00:10:39.225 "uuid": "42f5a85f-39ef-46b8-8883-b8db67f0babd", 00:10:39.225 "strip_size_kb": 64, 00:10:39.225 "state": "configuring", 00:10:39.225 "raid_level": "concat", 00:10:39.225 "superblock": true, 00:10:39.225 "num_base_bdevs": 4, 00:10:39.225 "num_base_bdevs_discovered": 0, 00:10:39.225 "num_base_bdevs_operational": 4, 00:10:39.225 "base_bdevs_list": [ 00:10:39.225 { 00:10:39.225 "name": "BaseBdev1", 00:10:39.225 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:39.225 "is_configured": false, 00:10:39.225 "data_offset": 0, 00:10:39.225 "data_size": 0 00:10:39.225 }, 00:10:39.225 { 00:10:39.225 "name": "BaseBdev2", 00:10:39.225 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:39.225 "is_configured": false, 00:10:39.225 "data_offset": 0, 00:10:39.225 "data_size": 0 00:10:39.225 }, 00:10:39.225 { 00:10:39.225 "name": "BaseBdev3", 00:10:39.225 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:39.225 "is_configured": false, 00:10:39.225 "data_offset": 0, 00:10:39.225 "data_size": 0 00:10:39.225 }, 00:10:39.225 { 00:10:39.225 "name": "BaseBdev4", 00:10:39.225 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:39.225 "is_configured": false, 00:10:39.225 "data_offset": 0, 00:10:39.225 "data_size": 0 00:10:39.225 } 00:10:39.225 ] 00:10:39.225 }' 00:10:39.225 17:32:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:39.225 17:32:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:39.484 17:32:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:10:39.484 17:32:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:39.484 17:32:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:39.484 [2024-11-27 17:32:10.670587] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:10:39.484 [2024-11-27 17:32:10.670694] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:10:39.743 17:32:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:39.743 17:32:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:10:39.743 17:32:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:39.743 17:32:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:39.743 [2024-11-27 17:32:10.682585] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:10:39.743 [2024-11-27 17:32:10.682684] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:10:39.743 [2024-11-27 17:32:10.682730] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:10:39.743 [2024-11-27 17:32:10.682752] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:10:39.743 [2024-11-27 17:32:10.682771] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:10:39.743 [2024-11-27 17:32:10.682845] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:10:39.743 [2024-11-27 17:32:10.682870] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:10:39.743 [2024-11-27 17:32:10.682893] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:10:39.743 17:32:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:39.743 17:32:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:10:39.743 17:32:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:39.743 17:32:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:39.743 [2024-11-27 17:32:10.709504] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:39.743 BaseBdev1 00:10:39.743 17:32:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:39.743 17:32:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:10:39.743 17:32:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:10:39.743 17:32:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:10:39.743 17:32:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:10:39.743 17:32:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:10:39.743 17:32:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:10:39.743 17:32:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:10:39.743 17:32:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:39.743 17:32:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:39.743 17:32:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:39.743 17:32:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:10:39.743 17:32:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:39.743 17:32:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:39.743 [ 00:10:39.743 { 00:10:39.743 "name": "BaseBdev1", 00:10:39.743 "aliases": [ 00:10:39.743 "e1669a7a-2d5d-492d-a330-0de087782ddc" 00:10:39.743 ], 00:10:39.743 "product_name": "Malloc disk", 00:10:39.743 "block_size": 512, 00:10:39.743 "num_blocks": 65536, 00:10:39.743 "uuid": "e1669a7a-2d5d-492d-a330-0de087782ddc", 00:10:39.743 "assigned_rate_limits": { 00:10:39.743 "rw_ios_per_sec": 0, 00:10:39.743 "rw_mbytes_per_sec": 0, 00:10:39.743 "r_mbytes_per_sec": 0, 00:10:39.743 "w_mbytes_per_sec": 0 00:10:39.743 }, 00:10:39.743 "claimed": true, 00:10:39.743 "claim_type": "exclusive_write", 00:10:39.743 "zoned": false, 00:10:39.743 "supported_io_types": { 00:10:39.743 "read": true, 00:10:39.743 "write": true, 00:10:39.743 "unmap": true, 00:10:39.743 "flush": true, 00:10:39.743 "reset": true, 00:10:39.743 "nvme_admin": false, 00:10:39.743 "nvme_io": false, 00:10:39.743 "nvme_io_md": false, 00:10:39.743 "write_zeroes": true, 00:10:39.743 "zcopy": true, 00:10:39.743 "get_zone_info": false, 00:10:39.743 "zone_management": false, 00:10:39.744 "zone_append": false, 00:10:39.744 "compare": false, 00:10:39.744 "compare_and_write": false, 00:10:39.744 "abort": true, 00:10:39.744 "seek_hole": false, 00:10:39.744 "seek_data": false, 00:10:39.744 "copy": true, 00:10:39.744 "nvme_iov_md": false 00:10:39.744 }, 00:10:39.744 "memory_domains": [ 00:10:39.744 { 00:10:39.744 "dma_device_id": "system", 00:10:39.744 "dma_device_type": 1 00:10:39.744 }, 00:10:39.744 { 00:10:39.744 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:39.744 "dma_device_type": 2 00:10:39.744 } 00:10:39.744 ], 00:10:39.744 "driver_specific": {} 00:10:39.744 } 00:10:39.744 ] 00:10:39.744 17:32:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:39.744 17:32:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:10:39.744 17:32:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:39.744 17:32:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:39.744 17:32:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:39.744 17:32:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:39.744 17:32:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:39.744 17:32:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:39.744 17:32:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:39.744 17:32:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:39.744 17:32:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:39.744 17:32:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:39.744 17:32:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:39.744 17:32:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:39.744 17:32:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:39.744 17:32:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:39.744 17:32:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:39.744 17:32:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:39.744 "name": "Existed_Raid", 00:10:39.744 "uuid": "0d999535-d40f-49f1-953d-e28cf91f8734", 00:10:39.744 "strip_size_kb": 64, 00:10:39.744 "state": "configuring", 00:10:39.744 "raid_level": "concat", 00:10:39.744 "superblock": true, 00:10:39.744 "num_base_bdevs": 4, 00:10:39.744 "num_base_bdevs_discovered": 1, 00:10:39.744 "num_base_bdevs_operational": 4, 00:10:39.744 "base_bdevs_list": [ 00:10:39.744 { 00:10:39.744 "name": "BaseBdev1", 00:10:39.744 "uuid": "e1669a7a-2d5d-492d-a330-0de087782ddc", 00:10:39.744 "is_configured": true, 00:10:39.744 "data_offset": 2048, 00:10:39.744 "data_size": 63488 00:10:39.744 }, 00:10:39.744 { 00:10:39.744 "name": "BaseBdev2", 00:10:39.744 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:39.744 "is_configured": false, 00:10:39.744 "data_offset": 0, 00:10:39.744 "data_size": 0 00:10:39.744 }, 00:10:39.744 { 00:10:39.744 "name": "BaseBdev3", 00:10:39.744 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:39.744 "is_configured": false, 00:10:39.744 "data_offset": 0, 00:10:39.744 "data_size": 0 00:10:39.744 }, 00:10:39.744 { 00:10:39.744 "name": "BaseBdev4", 00:10:39.744 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:39.744 "is_configured": false, 00:10:39.744 "data_offset": 0, 00:10:39.744 "data_size": 0 00:10:39.744 } 00:10:39.744 ] 00:10:39.744 }' 00:10:39.744 17:32:10 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:39.744 17:32:10 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:40.004 17:32:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:10:40.004 17:32:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:40.004 17:32:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:40.264 [2024-11-27 17:32:11.196704] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:10:40.264 [2024-11-27 17:32:11.196804] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:10:40.264 17:32:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:40.264 17:32:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:10:40.264 17:32:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:40.265 17:32:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:40.265 [2024-11-27 17:32:11.204766] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:40.265 [2024-11-27 17:32:11.206786] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:10:40.265 [2024-11-27 17:32:11.206821] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:10:40.265 [2024-11-27 17:32:11.206830] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:10:40.265 [2024-11-27 17:32:11.206855] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:10:40.265 [2024-11-27 17:32:11.206861] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:10:40.265 [2024-11-27 17:32:11.206869] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:10:40.265 17:32:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:40.265 17:32:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:10:40.265 17:32:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:40.265 17:32:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:40.265 17:32:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:40.265 17:32:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:40.265 17:32:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:40.265 17:32:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:40.265 17:32:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:40.265 17:32:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:40.265 17:32:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:40.265 17:32:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:40.265 17:32:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:40.265 17:32:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:40.265 17:32:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:40.265 17:32:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:40.265 17:32:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:40.265 17:32:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:40.265 17:32:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:40.265 "name": "Existed_Raid", 00:10:40.265 "uuid": "35f3f64a-f847-417e-8390-6b3cb0344a1a", 00:10:40.265 "strip_size_kb": 64, 00:10:40.265 "state": "configuring", 00:10:40.265 "raid_level": "concat", 00:10:40.265 "superblock": true, 00:10:40.265 "num_base_bdevs": 4, 00:10:40.265 "num_base_bdevs_discovered": 1, 00:10:40.265 "num_base_bdevs_operational": 4, 00:10:40.265 "base_bdevs_list": [ 00:10:40.265 { 00:10:40.265 "name": "BaseBdev1", 00:10:40.265 "uuid": "e1669a7a-2d5d-492d-a330-0de087782ddc", 00:10:40.265 "is_configured": true, 00:10:40.265 "data_offset": 2048, 00:10:40.265 "data_size": 63488 00:10:40.265 }, 00:10:40.265 { 00:10:40.265 "name": "BaseBdev2", 00:10:40.265 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:40.265 "is_configured": false, 00:10:40.265 "data_offset": 0, 00:10:40.265 "data_size": 0 00:10:40.265 }, 00:10:40.265 { 00:10:40.265 "name": "BaseBdev3", 00:10:40.265 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:40.265 "is_configured": false, 00:10:40.265 "data_offset": 0, 00:10:40.265 "data_size": 0 00:10:40.265 }, 00:10:40.265 { 00:10:40.265 "name": "BaseBdev4", 00:10:40.265 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:40.265 "is_configured": false, 00:10:40.265 "data_offset": 0, 00:10:40.265 "data_size": 0 00:10:40.265 } 00:10:40.265 ] 00:10:40.265 }' 00:10:40.265 17:32:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:40.265 17:32:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:40.525 17:32:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:10:40.525 17:32:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:40.525 17:32:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:40.525 [2024-11-27 17:32:11.670603] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:40.525 BaseBdev2 00:10:40.525 17:32:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:40.525 17:32:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:10:40.525 17:32:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:10:40.525 17:32:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:10:40.525 17:32:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:10:40.525 17:32:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:10:40.525 17:32:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:10:40.525 17:32:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:10:40.525 17:32:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:40.525 17:32:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:40.525 17:32:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:40.525 17:32:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:10:40.525 17:32:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:40.525 17:32:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:40.525 [ 00:10:40.525 { 00:10:40.525 "name": "BaseBdev2", 00:10:40.525 "aliases": [ 00:10:40.525 "40be7ced-be09-4f96-a6ef-c8afc35869e5" 00:10:40.525 ], 00:10:40.525 "product_name": "Malloc disk", 00:10:40.525 "block_size": 512, 00:10:40.525 "num_blocks": 65536, 00:10:40.525 "uuid": "40be7ced-be09-4f96-a6ef-c8afc35869e5", 00:10:40.525 "assigned_rate_limits": { 00:10:40.525 "rw_ios_per_sec": 0, 00:10:40.525 "rw_mbytes_per_sec": 0, 00:10:40.525 "r_mbytes_per_sec": 0, 00:10:40.525 "w_mbytes_per_sec": 0 00:10:40.525 }, 00:10:40.525 "claimed": true, 00:10:40.525 "claim_type": "exclusive_write", 00:10:40.525 "zoned": false, 00:10:40.525 "supported_io_types": { 00:10:40.525 "read": true, 00:10:40.525 "write": true, 00:10:40.525 "unmap": true, 00:10:40.525 "flush": true, 00:10:40.525 "reset": true, 00:10:40.525 "nvme_admin": false, 00:10:40.525 "nvme_io": false, 00:10:40.525 "nvme_io_md": false, 00:10:40.525 "write_zeroes": true, 00:10:40.525 "zcopy": true, 00:10:40.525 "get_zone_info": false, 00:10:40.525 "zone_management": false, 00:10:40.525 "zone_append": false, 00:10:40.525 "compare": false, 00:10:40.525 "compare_and_write": false, 00:10:40.525 "abort": true, 00:10:40.525 "seek_hole": false, 00:10:40.525 "seek_data": false, 00:10:40.525 "copy": true, 00:10:40.525 "nvme_iov_md": false 00:10:40.525 }, 00:10:40.525 "memory_domains": [ 00:10:40.525 { 00:10:40.525 "dma_device_id": "system", 00:10:40.525 "dma_device_type": 1 00:10:40.525 }, 00:10:40.525 { 00:10:40.525 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:40.525 "dma_device_type": 2 00:10:40.525 } 00:10:40.525 ], 00:10:40.525 "driver_specific": {} 00:10:40.525 } 00:10:40.525 ] 00:10:40.525 17:32:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:40.526 17:32:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:10:40.526 17:32:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:10:40.526 17:32:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:40.526 17:32:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:40.526 17:32:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:40.526 17:32:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:40.526 17:32:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:40.526 17:32:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:40.526 17:32:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:40.526 17:32:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:40.526 17:32:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:40.526 17:32:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:40.526 17:32:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:40.526 17:32:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:40.526 17:32:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:40.790 17:32:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:40.790 17:32:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:40.790 17:32:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:40.790 17:32:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:40.790 "name": "Existed_Raid", 00:10:40.790 "uuid": "35f3f64a-f847-417e-8390-6b3cb0344a1a", 00:10:40.790 "strip_size_kb": 64, 00:10:40.790 "state": "configuring", 00:10:40.790 "raid_level": "concat", 00:10:40.790 "superblock": true, 00:10:40.790 "num_base_bdevs": 4, 00:10:40.790 "num_base_bdevs_discovered": 2, 00:10:40.790 "num_base_bdevs_operational": 4, 00:10:40.790 "base_bdevs_list": [ 00:10:40.790 { 00:10:40.790 "name": "BaseBdev1", 00:10:40.790 "uuid": "e1669a7a-2d5d-492d-a330-0de087782ddc", 00:10:40.790 "is_configured": true, 00:10:40.790 "data_offset": 2048, 00:10:40.791 "data_size": 63488 00:10:40.791 }, 00:10:40.791 { 00:10:40.791 "name": "BaseBdev2", 00:10:40.791 "uuid": "40be7ced-be09-4f96-a6ef-c8afc35869e5", 00:10:40.791 "is_configured": true, 00:10:40.791 "data_offset": 2048, 00:10:40.791 "data_size": 63488 00:10:40.791 }, 00:10:40.791 { 00:10:40.791 "name": "BaseBdev3", 00:10:40.791 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:40.791 "is_configured": false, 00:10:40.791 "data_offset": 0, 00:10:40.791 "data_size": 0 00:10:40.791 }, 00:10:40.791 { 00:10:40.791 "name": "BaseBdev4", 00:10:40.791 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:40.791 "is_configured": false, 00:10:40.791 "data_offset": 0, 00:10:40.791 "data_size": 0 00:10:40.791 } 00:10:40.791 ] 00:10:40.791 }' 00:10:40.791 17:32:11 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:40.791 17:32:11 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:41.066 17:32:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:10:41.066 17:32:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:41.066 17:32:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:41.066 [2024-11-27 17:32:12.150490] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:41.066 BaseBdev3 00:10:41.066 17:32:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:41.066 17:32:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:10:41.066 17:32:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:10:41.066 17:32:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:10:41.066 17:32:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:10:41.066 17:32:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:10:41.066 17:32:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:10:41.066 17:32:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:10:41.066 17:32:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:41.066 17:32:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:41.066 17:32:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:41.066 17:32:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:10:41.066 17:32:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:41.066 17:32:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:41.066 [ 00:10:41.066 { 00:10:41.066 "name": "BaseBdev3", 00:10:41.066 "aliases": [ 00:10:41.066 "2b67aea3-0e0d-48f1-99e1-377ae3b196f0" 00:10:41.066 ], 00:10:41.066 "product_name": "Malloc disk", 00:10:41.066 "block_size": 512, 00:10:41.066 "num_blocks": 65536, 00:10:41.066 "uuid": "2b67aea3-0e0d-48f1-99e1-377ae3b196f0", 00:10:41.066 "assigned_rate_limits": { 00:10:41.066 "rw_ios_per_sec": 0, 00:10:41.066 "rw_mbytes_per_sec": 0, 00:10:41.066 "r_mbytes_per_sec": 0, 00:10:41.066 "w_mbytes_per_sec": 0 00:10:41.066 }, 00:10:41.066 "claimed": true, 00:10:41.066 "claim_type": "exclusive_write", 00:10:41.066 "zoned": false, 00:10:41.066 "supported_io_types": { 00:10:41.066 "read": true, 00:10:41.066 "write": true, 00:10:41.066 "unmap": true, 00:10:41.066 "flush": true, 00:10:41.066 "reset": true, 00:10:41.066 "nvme_admin": false, 00:10:41.066 "nvme_io": false, 00:10:41.066 "nvme_io_md": false, 00:10:41.066 "write_zeroes": true, 00:10:41.066 "zcopy": true, 00:10:41.066 "get_zone_info": false, 00:10:41.066 "zone_management": false, 00:10:41.066 "zone_append": false, 00:10:41.066 "compare": false, 00:10:41.066 "compare_and_write": false, 00:10:41.066 "abort": true, 00:10:41.066 "seek_hole": false, 00:10:41.066 "seek_data": false, 00:10:41.066 "copy": true, 00:10:41.066 "nvme_iov_md": false 00:10:41.066 }, 00:10:41.066 "memory_domains": [ 00:10:41.066 { 00:10:41.066 "dma_device_id": "system", 00:10:41.066 "dma_device_type": 1 00:10:41.066 }, 00:10:41.066 { 00:10:41.066 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:41.066 "dma_device_type": 2 00:10:41.066 } 00:10:41.066 ], 00:10:41.066 "driver_specific": {} 00:10:41.066 } 00:10:41.066 ] 00:10:41.066 17:32:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:41.066 17:32:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:10:41.066 17:32:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:10:41.066 17:32:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:41.066 17:32:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:41.066 17:32:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:41.066 17:32:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:41.066 17:32:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:41.066 17:32:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:41.066 17:32:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:41.066 17:32:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:41.066 17:32:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:41.066 17:32:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:41.066 17:32:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:41.066 17:32:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:41.066 17:32:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:41.066 17:32:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:41.066 17:32:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:41.066 17:32:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:41.066 17:32:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:41.066 "name": "Existed_Raid", 00:10:41.066 "uuid": "35f3f64a-f847-417e-8390-6b3cb0344a1a", 00:10:41.066 "strip_size_kb": 64, 00:10:41.066 "state": "configuring", 00:10:41.066 "raid_level": "concat", 00:10:41.066 "superblock": true, 00:10:41.066 "num_base_bdevs": 4, 00:10:41.066 "num_base_bdevs_discovered": 3, 00:10:41.066 "num_base_bdevs_operational": 4, 00:10:41.066 "base_bdevs_list": [ 00:10:41.066 { 00:10:41.066 "name": "BaseBdev1", 00:10:41.066 "uuid": "e1669a7a-2d5d-492d-a330-0de087782ddc", 00:10:41.066 "is_configured": true, 00:10:41.066 "data_offset": 2048, 00:10:41.066 "data_size": 63488 00:10:41.066 }, 00:10:41.066 { 00:10:41.066 "name": "BaseBdev2", 00:10:41.066 "uuid": "40be7ced-be09-4f96-a6ef-c8afc35869e5", 00:10:41.066 "is_configured": true, 00:10:41.066 "data_offset": 2048, 00:10:41.066 "data_size": 63488 00:10:41.066 }, 00:10:41.066 { 00:10:41.066 "name": "BaseBdev3", 00:10:41.066 "uuid": "2b67aea3-0e0d-48f1-99e1-377ae3b196f0", 00:10:41.066 "is_configured": true, 00:10:41.066 "data_offset": 2048, 00:10:41.066 "data_size": 63488 00:10:41.066 }, 00:10:41.066 { 00:10:41.066 "name": "BaseBdev4", 00:10:41.066 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:41.066 "is_configured": false, 00:10:41.066 "data_offset": 0, 00:10:41.066 "data_size": 0 00:10:41.066 } 00:10:41.066 ] 00:10:41.066 }' 00:10:41.066 17:32:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:41.066 17:32:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:41.651 17:32:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:10:41.651 17:32:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:41.651 17:32:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:41.651 [2024-11-27 17:32:12.670524] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:10:41.651 [2024-11-27 17:32:12.670868] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:10:41.651 [2024-11-27 17:32:12.670929] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:10:41.651 [2024-11-27 17:32:12.671281] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:10:41.651 [2024-11-27 17:32:12.671485] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:10:41.651 [2024-11-27 17:32:12.671544] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:10:41.651 [2024-11-27 17:32:12.671704] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:41.651 BaseBdev4 00:10:41.651 17:32:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:41.651 17:32:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev4 00:10:41.651 17:32:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev4 00:10:41.651 17:32:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:10:41.651 17:32:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:10:41.651 17:32:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:10:41.651 17:32:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:10:41.651 17:32:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:10:41.651 17:32:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:41.651 17:32:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:41.651 17:32:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:41.651 17:32:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:10:41.651 17:32:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:41.651 17:32:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:41.651 [ 00:10:41.651 { 00:10:41.651 "name": "BaseBdev4", 00:10:41.651 "aliases": [ 00:10:41.651 "6b11d6c9-ac6b-49e8-93fa-85d68e9508cc" 00:10:41.651 ], 00:10:41.651 "product_name": "Malloc disk", 00:10:41.651 "block_size": 512, 00:10:41.651 "num_blocks": 65536, 00:10:41.651 "uuid": "6b11d6c9-ac6b-49e8-93fa-85d68e9508cc", 00:10:41.651 "assigned_rate_limits": { 00:10:41.651 "rw_ios_per_sec": 0, 00:10:41.651 "rw_mbytes_per_sec": 0, 00:10:41.651 "r_mbytes_per_sec": 0, 00:10:41.651 "w_mbytes_per_sec": 0 00:10:41.651 }, 00:10:41.651 "claimed": true, 00:10:41.651 "claim_type": "exclusive_write", 00:10:41.651 "zoned": false, 00:10:41.651 "supported_io_types": { 00:10:41.651 "read": true, 00:10:41.651 "write": true, 00:10:41.651 "unmap": true, 00:10:41.651 "flush": true, 00:10:41.651 "reset": true, 00:10:41.651 "nvme_admin": false, 00:10:41.651 "nvme_io": false, 00:10:41.651 "nvme_io_md": false, 00:10:41.651 "write_zeroes": true, 00:10:41.651 "zcopy": true, 00:10:41.651 "get_zone_info": false, 00:10:41.651 "zone_management": false, 00:10:41.651 "zone_append": false, 00:10:41.651 "compare": false, 00:10:41.652 "compare_and_write": false, 00:10:41.652 "abort": true, 00:10:41.652 "seek_hole": false, 00:10:41.652 "seek_data": false, 00:10:41.652 "copy": true, 00:10:41.652 "nvme_iov_md": false 00:10:41.652 }, 00:10:41.652 "memory_domains": [ 00:10:41.652 { 00:10:41.652 "dma_device_id": "system", 00:10:41.652 "dma_device_type": 1 00:10:41.652 }, 00:10:41.652 { 00:10:41.652 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:41.652 "dma_device_type": 2 00:10:41.652 } 00:10:41.652 ], 00:10:41.652 "driver_specific": {} 00:10:41.652 } 00:10:41.652 ] 00:10:41.652 17:32:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:41.652 17:32:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:10:41.652 17:32:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:10:41.652 17:32:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:10:41.652 17:32:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online concat 64 4 00:10:41.652 17:32:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:41.652 17:32:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:41.652 17:32:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:41.652 17:32:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:41.652 17:32:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:41.652 17:32:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:41.652 17:32:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:41.652 17:32:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:41.652 17:32:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:41.652 17:32:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:41.652 17:32:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:41.652 17:32:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:41.652 17:32:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:41.652 17:32:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:41.652 17:32:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:41.652 "name": "Existed_Raid", 00:10:41.652 "uuid": "35f3f64a-f847-417e-8390-6b3cb0344a1a", 00:10:41.652 "strip_size_kb": 64, 00:10:41.652 "state": "online", 00:10:41.652 "raid_level": "concat", 00:10:41.652 "superblock": true, 00:10:41.652 "num_base_bdevs": 4, 00:10:41.652 "num_base_bdevs_discovered": 4, 00:10:41.652 "num_base_bdevs_operational": 4, 00:10:41.652 "base_bdevs_list": [ 00:10:41.652 { 00:10:41.652 "name": "BaseBdev1", 00:10:41.652 "uuid": "e1669a7a-2d5d-492d-a330-0de087782ddc", 00:10:41.652 "is_configured": true, 00:10:41.652 "data_offset": 2048, 00:10:41.652 "data_size": 63488 00:10:41.652 }, 00:10:41.652 { 00:10:41.652 "name": "BaseBdev2", 00:10:41.652 "uuid": "40be7ced-be09-4f96-a6ef-c8afc35869e5", 00:10:41.652 "is_configured": true, 00:10:41.652 "data_offset": 2048, 00:10:41.652 "data_size": 63488 00:10:41.652 }, 00:10:41.652 { 00:10:41.652 "name": "BaseBdev3", 00:10:41.652 "uuid": "2b67aea3-0e0d-48f1-99e1-377ae3b196f0", 00:10:41.652 "is_configured": true, 00:10:41.652 "data_offset": 2048, 00:10:41.652 "data_size": 63488 00:10:41.652 }, 00:10:41.652 { 00:10:41.652 "name": "BaseBdev4", 00:10:41.652 "uuid": "6b11d6c9-ac6b-49e8-93fa-85d68e9508cc", 00:10:41.652 "is_configured": true, 00:10:41.652 "data_offset": 2048, 00:10:41.652 "data_size": 63488 00:10:41.652 } 00:10:41.652 ] 00:10:41.652 }' 00:10:41.652 17:32:12 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:41.652 17:32:12 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:42.222 17:32:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:10:42.222 17:32:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:10:42.222 17:32:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:10:42.222 17:32:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:10:42.222 17:32:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:10:42.222 17:32:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:10:42.222 17:32:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:10:42.222 17:32:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:42.222 17:32:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:42.222 17:32:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:10:42.222 [2024-11-27 17:32:13.146050] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:42.222 17:32:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:42.223 17:32:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:42.223 "name": "Existed_Raid", 00:10:42.223 "aliases": [ 00:10:42.223 "35f3f64a-f847-417e-8390-6b3cb0344a1a" 00:10:42.223 ], 00:10:42.223 "product_name": "Raid Volume", 00:10:42.223 "block_size": 512, 00:10:42.223 "num_blocks": 253952, 00:10:42.223 "uuid": "35f3f64a-f847-417e-8390-6b3cb0344a1a", 00:10:42.223 "assigned_rate_limits": { 00:10:42.223 "rw_ios_per_sec": 0, 00:10:42.223 "rw_mbytes_per_sec": 0, 00:10:42.223 "r_mbytes_per_sec": 0, 00:10:42.223 "w_mbytes_per_sec": 0 00:10:42.223 }, 00:10:42.223 "claimed": false, 00:10:42.223 "zoned": false, 00:10:42.223 "supported_io_types": { 00:10:42.223 "read": true, 00:10:42.223 "write": true, 00:10:42.223 "unmap": true, 00:10:42.223 "flush": true, 00:10:42.223 "reset": true, 00:10:42.223 "nvme_admin": false, 00:10:42.223 "nvme_io": false, 00:10:42.223 "nvme_io_md": false, 00:10:42.223 "write_zeroes": true, 00:10:42.223 "zcopy": false, 00:10:42.223 "get_zone_info": false, 00:10:42.223 "zone_management": false, 00:10:42.223 "zone_append": false, 00:10:42.223 "compare": false, 00:10:42.223 "compare_and_write": false, 00:10:42.223 "abort": false, 00:10:42.223 "seek_hole": false, 00:10:42.223 "seek_data": false, 00:10:42.223 "copy": false, 00:10:42.223 "nvme_iov_md": false 00:10:42.223 }, 00:10:42.223 "memory_domains": [ 00:10:42.223 { 00:10:42.223 "dma_device_id": "system", 00:10:42.223 "dma_device_type": 1 00:10:42.223 }, 00:10:42.223 { 00:10:42.223 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:42.223 "dma_device_type": 2 00:10:42.223 }, 00:10:42.223 { 00:10:42.223 "dma_device_id": "system", 00:10:42.223 "dma_device_type": 1 00:10:42.223 }, 00:10:42.223 { 00:10:42.223 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:42.223 "dma_device_type": 2 00:10:42.223 }, 00:10:42.223 { 00:10:42.223 "dma_device_id": "system", 00:10:42.223 "dma_device_type": 1 00:10:42.223 }, 00:10:42.223 { 00:10:42.223 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:42.223 "dma_device_type": 2 00:10:42.223 }, 00:10:42.223 { 00:10:42.223 "dma_device_id": "system", 00:10:42.223 "dma_device_type": 1 00:10:42.223 }, 00:10:42.223 { 00:10:42.223 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:42.223 "dma_device_type": 2 00:10:42.223 } 00:10:42.223 ], 00:10:42.223 "driver_specific": { 00:10:42.223 "raid": { 00:10:42.223 "uuid": "35f3f64a-f847-417e-8390-6b3cb0344a1a", 00:10:42.223 "strip_size_kb": 64, 00:10:42.223 "state": "online", 00:10:42.223 "raid_level": "concat", 00:10:42.223 "superblock": true, 00:10:42.223 "num_base_bdevs": 4, 00:10:42.223 "num_base_bdevs_discovered": 4, 00:10:42.223 "num_base_bdevs_operational": 4, 00:10:42.223 "base_bdevs_list": [ 00:10:42.223 { 00:10:42.223 "name": "BaseBdev1", 00:10:42.223 "uuid": "e1669a7a-2d5d-492d-a330-0de087782ddc", 00:10:42.223 "is_configured": true, 00:10:42.223 "data_offset": 2048, 00:10:42.223 "data_size": 63488 00:10:42.223 }, 00:10:42.223 { 00:10:42.223 "name": "BaseBdev2", 00:10:42.223 "uuid": "40be7ced-be09-4f96-a6ef-c8afc35869e5", 00:10:42.223 "is_configured": true, 00:10:42.223 "data_offset": 2048, 00:10:42.223 "data_size": 63488 00:10:42.223 }, 00:10:42.223 { 00:10:42.223 "name": "BaseBdev3", 00:10:42.223 "uuid": "2b67aea3-0e0d-48f1-99e1-377ae3b196f0", 00:10:42.223 "is_configured": true, 00:10:42.223 "data_offset": 2048, 00:10:42.223 "data_size": 63488 00:10:42.223 }, 00:10:42.223 { 00:10:42.223 "name": "BaseBdev4", 00:10:42.223 "uuid": "6b11d6c9-ac6b-49e8-93fa-85d68e9508cc", 00:10:42.223 "is_configured": true, 00:10:42.223 "data_offset": 2048, 00:10:42.223 "data_size": 63488 00:10:42.223 } 00:10:42.223 ] 00:10:42.223 } 00:10:42.223 } 00:10:42.223 }' 00:10:42.223 17:32:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:42.223 17:32:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:10:42.223 BaseBdev2 00:10:42.223 BaseBdev3 00:10:42.223 BaseBdev4' 00:10:42.223 17:32:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:42.223 17:32:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:42.223 17:32:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:42.223 17:32:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:42.223 17:32:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:10:42.223 17:32:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:42.223 17:32:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:42.223 17:32:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:42.223 17:32:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:42.223 17:32:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:42.223 17:32:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:42.223 17:32:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:42.223 17:32:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:10:42.223 17:32:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:42.223 17:32:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:42.223 17:32:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:42.223 17:32:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:42.223 17:32:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:42.223 17:32:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:42.223 17:32:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:42.223 17:32:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:10:42.223 17:32:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:42.223 17:32:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:42.223 17:32:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:42.223 17:32:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:42.223 17:32:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:42.223 17:32:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:42.223 17:32:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:10:42.224 17:32:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:42.224 17:32:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:42.224 17:32:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:42.224 17:32:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:42.224 17:32:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:42.224 17:32:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:42.224 17:32:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:10:42.224 17:32:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:42.224 17:32:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:42.224 [2024-11-27 17:32:13.393333] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:10:42.224 [2024-11-27 17:32:13.393363] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:42.224 [2024-11-27 17:32:13.393413] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:42.483 17:32:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:42.483 17:32:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:10:42.483 17:32:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy concat 00:10:42.483 17:32:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:10:42.483 17:32:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@200 -- # return 1 00:10:42.483 17:32:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@262 -- # expected_state=offline 00:10:42.483 17:32:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid offline concat 64 3 00:10:42.483 17:32:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:42.483 17:32:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=offline 00:10:42.483 17:32:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:42.484 17:32:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:42.484 17:32:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:10:42.484 17:32:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:42.484 17:32:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:42.484 17:32:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:42.484 17:32:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:42.484 17:32:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:42.484 17:32:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:42.484 17:32:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:42.484 17:32:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:42.484 17:32:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:42.484 17:32:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:42.484 "name": "Existed_Raid", 00:10:42.484 "uuid": "35f3f64a-f847-417e-8390-6b3cb0344a1a", 00:10:42.484 "strip_size_kb": 64, 00:10:42.484 "state": "offline", 00:10:42.484 "raid_level": "concat", 00:10:42.484 "superblock": true, 00:10:42.484 "num_base_bdevs": 4, 00:10:42.484 "num_base_bdevs_discovered": 3, 00:10:42.484 "num_base_bdevs_operational": 3, 00:10:42.484 "base_bdevs_list": [ 00:10:42.484 { 00:10:42.484 "name": null, 00:10:42.484 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:42.484 "is_configured": false, 00:10:42.484 "data_offset": 0, 00:10:42.484 "data_size": 63488 00:10:42.484 }, 00:10:42.484 { 00:10:42.484 "name": "BaseBdev2", 00:10:42.484 "uuid": "40be7ced-be09-4f96-a6ef-c8afc35869e5", 00:10:42.484 "is_configured": true, 00:10:42.484 "data_offset": 2048, 00:10:42.484 "data_size": 63488 00:10:42.484 }, 00:10:42.484 { 00:10:42.484 "name": "BaseBdev3", 00:10:42.484 "uuid": "2b67aea3-0e0d-48f1-99e1-377ae3b196f0", 00:10:42.484 "is_configured": true, 00:10:42.484 "data_offset": 2048, 00:10:42.484 "data_size": 63488 00:10:42.484 }, 00:10:42.484 { 00:10:42.484 "name": "BaseBdev4", 00:10:42.484 "uuid": "6b11d6c9-ac6b-49e8-93fa-85d68e9508cc", 00:10:42.484 "is_configured": true, 00:10:42.484 "data_offset": 2048, 00:10:42.484 "data_size": 63488 00:10:42.484 } 00:10:42.484 ] 00:10:42.484 }' 00:10:42.484 17:32:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:42.484 17:32:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:42.744 17:32:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:10:42.744 17:32:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:42.744 17:32:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:42.744 17:32:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:42.744 17:32:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:42.744 17:32:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:10:42.744 17:32:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:42.744 17:32:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:10:42.744 17:32:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:10:42.744 17:32:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:10:42.744 17:32:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:42.744 17:32:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:42.744 [2024-11-27 17:32:13.885299] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:10:42.744 17:32:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:42.744 17:32:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:10:42.744 17:32:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:42.744 17:32:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:42.744 17:32:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:42.744 17:32:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:42.744 17:32:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:10:42.744 17:32:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:43.004 17:32:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:10:43.004 17:32:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:10:43.004 17:32:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:10:43.004 17:32:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:43.004 17:32:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:43.004 [2024-11-27 17:32:13.953658] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:10:43.004 17:32:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:43.004 17:32:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:10:43.004 17:32:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:43.004 17:32:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:43.004 17:32:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:43.004 17:32:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:43.004 17:32:13 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:10:43.004 17:32:13 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:43.004 17:32:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:10:43.004 17:32:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:10:43.004 17:32:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev4 00:10:43.004 17:32:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:43.004 17:32:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:43.004 [2024-11-27 17:32:14.029282] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev4 00:10:43.004 [2024-11-27 17:32:14.029392] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:10:43.004 17:32:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:43.004 17:32:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:10:43.004 17:32:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:10:43.004 17:32:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:43.004 17:32:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:10:43.004 17:32:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:43.004 17:32:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:43.004 17:32:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:43.004 17:32:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:10:43.004 17:32:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:10:43.004 17:32:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 4 -gt 2 ']' 00:10:43.004 17:32:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:10:43.004 17:32:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:43.004 17:32:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:10:43.004 17:32:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:43.004 17:32:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:43.004 BaseBdev2 00:10:43.004 17:32:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:43.005 17:32:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:10:43.005 17:32:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:10:43.005 17:32:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:10:43.005 17:32:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:10:43.005 17:32:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:10:43.005 17:32:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:10:43.005 17:32:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:10:43.005 17:32:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:43.005 17:32:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:43.005 17:32:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:43.005 17:32:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:10:43.005 17:32:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:43.005 17:32:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:43.005 [ 00:10:43.005 { 00:10:43.005 "name": "BaseBdev2", 00:10:43.005 "aliases": [ 00:10:43.005 "6d03b07a-708d-48f9-9d9b-3c6aa9450d1f" 00:10:43.005 ], 00:10:43.005 "product_name": "Malloc disk", 00:10:43.005 "block_size": 512, 00:10:43.005 "num_blocks": 65536, 00:10:43.005 "uuid": "6d03b07a-708d-48f9-9d9b-3c6aa9450d1f", 00:10:43.005 "assigned_rate_limits": { 00:10:43.005 "rw_ios_per_sec": 0, 00:10:43.005 "rw_mbytes_per_sec": 0, 00:10:43.005 "r_mbytes_per_sec": 0, 00:10:43.005 "w_mbytes_per_sec": 0 00:10:43.005 }, 00:10:43.005 "claimed": false, 00:10:43.005 "zoned": false, 00:10:43.005 "supported_io_types": { 00:10:43.005 "read": true, 00:10:43.005 "write": true, 00:10:43.005 "unmap": true, 00:10:43.005 "flush": true, 00:10:43.005 "reset": true, 00:10:43.005 "nvme_admin": false, 00:10:43.005 "nvme_io": false, 00:10:43.005 "nvme_io_md": false, 00:10:43.005 "write_zeroes": true, 00:10:43.005 "zcopy": true, 00:10:43.005 "get_zone_info": false, 00:10:43.005 "zone_management": false, 00:10:43.005 "zone_append": false, 00:10:43.005 "compare": false, 00:10:43.005 "compare_and_write": false, 00:10:43.005 "abort": true, 00:10:43.005 "seek_hole": false, 00:10:43.005 "seek_data": false, 00:10:43.005 "copy": true, 00:10:43.005 "nvme_iov_md": false 00:10:43.005 }, 00:10:43.005 "memory_domains": [ 00:10:43.005 { 00:10:43.005 "dma_device_id": "system", 00:10:43.005 "dma_device_type": 1 00:10:43.005 }, 00:10:43.005 { 00:10:43.005 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:43.005 "dma_device_type": 2 00:10:43.005 } 00:10:43.005 ], 00:10:43.005 "driver_specific": {} 00:10:43.005 } 00:10:43.005 ] 00:10:43.005 17:32:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:43.005 17:32:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:10:43.005 17:32:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:10:43.005 17:32:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:43.005 17:32:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:10:43.005 17:32:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:43.005 17:32:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:43.005 BaseBdev3 00:10:43.005 17:32:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:43.005 17:32:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:10:43.005 17:32:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:10:43.005 17:32:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:10:43.005 17:32:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:10:43.005 17:32:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:10:43.005 17:32:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:10:43.005 17:32:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:10:43.005 17:32:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:43.005 17:32:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:43.265 17:32:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:43.265 17:32:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:10:43.265 17:32:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:43.265 17:32:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:43.265 [ 00:10:43.265 { 00:10:43.265 "name": "BaseBdev3", 00:10:43.265 "aliases": [ 00:10:43.265 "987a8497-3566-482e-9484-33509955d1d2" 00:10:43.265 ], 00:10:43.265 "product_name": "Malloc disk", 00:10:43.266 "block_size": 512, 00:10:43.266 "num_blocks": 65536, 00:10:43.266 "uuid": "987a8497-3566-482e-9484-33509955d1d2", 00:10:43.266 "assigned_rate_limits": { 00:10:43.266 "rw_ios_per_sec": 0, 00:10:43.266 "rw_mbytes_per_sec": 0, 00:10:43.266 "r_mbytes_per_sec": 0, 00:10:43.266 "w_mbytes_per_sec": 0 00:10:43.266 }, 00:10:43.266 "claimed": false, 00:10:43.266 "zoned": false, 00:10:43.266 "supported_io_types": { 00:10:43.266 "read": true, 00:10:43.266 "write": true, 00:10:43.266 "unmap": true, 00:10:43.266 "flush": true, 00:10:43.266 "reset": true, 00:10:43.266 "nvme_admin": false, 00:10:43.266 "nvme_io": false, 00:10:43.266 "nvme_io_md": false, 00:10:43.266 "write_zeroes": true, 00:10:43.266 "zcopy": true, 00:10:43.266 "get_zone_info": false, 00:10:43.266 "zone_management": false, 00:10:43.266 "zone_append": false, 00:10:43.266 "compare": false, 00:10:43.266 "compare_and_write": false, 00:10:43.266 "abort": true, 00:10:43.266 "seek_hole": false, 00:10:43.266 "seek_data": false, 00:10:43.266 "copy": true, 00:10:43.266 "nvme_iov_md": false 00:10:43.266 }, 00:10:43.266 "memory_domains": [ 00:10:43.266 { 00:10:43.266 "dma_device_id": "system", 00:10:43.266 "dma_device_type": 1 00:10:43.266 }, 00:10:43.266 { 00:10:43.266 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:43.266 "dma_device_type": 2 00:10:43.266 } 00:10:43.266 ], 00:10:43.266 "driver_specific": {} 00:10:43.266 } 00:10:43.266 ] 00:10:43.266 17:32:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:43.266 17:32:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:10:43.266 17:32:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:10:43.266 17:32:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:43.266 17:32:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:10:43.266 17:32:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:43.266 17:32:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:43.266 BaseBdev4 00:10:43.266 17:32:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:43.266 17:32:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev4 00:10:43.266 17:32:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev4 00:10:43.266 17:32:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:10:43.266 17:32:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:10:43.266 17:32:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:10:43.266 17:32:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:10:43.266 17:32:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:10:43.266 17:32:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:43.266 17:32:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:43.266 17:32:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:43.266 17:32:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:10:43.266 17:32:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:43.266 17:32:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:43.266 [ 00:10:43.266 { 00:10:43.266 "name": "BaseBdev4", 00:10:43.266 "aliases": [ 00:10:43.266 "110bc70b-7a7e-4d0b-9f43-2044486e7e8b" 00:10:43.266 ], 00:10:43.266 "product_name": "Malloc disk", 00:10:43.266 "block_size": 512, 00:10:43.266 "num_blocks": 65536, 00:10:43.266 "uuid": "110bc70b-7a7e-4d0b-9f43-2044486e7e8b", 00:10:43.266 "assigned_rate_limits": { 00:10:43.266 "rw_ios_per_sec": 0, 00:10:43.266 "rw_mbytes_per_sec": 0, 00:10:43.266 "r_mbytes_per_sec": 0, 00:10:43.266 "w_mbytes_per_sec": 0 00:10:43.266 }, 00:10:43.266 "claimed": false, 00:10:43.266 "zoned": false, 00:10:43.266 "supported_io_types": { 00:10:43.266 "read": true, 00:10:43.266 "write": true, 00:10:43.266 "unmap": true, 00:10:43.266 "flush": true, 00:10:43.266 "reset": true, 00:10:43.266 "nvme_admin": false, 00:10:43.266 "nvme_io": false, 00:10:43.266 "nvme_io_md": false, 00:10:43.266 "write_zeroes": true, 00:10:43.266 "zcopy": true, 00:10:43.266 "get_zone_info": false, 00:10:43.266 "zone_management": false, 00:10:43.266 "zone_append": false, 00:10:43.266 "compare": false, 00:10:43.266 "compare_and_write": false, 00:10:43.266 "abort": true, 00:10:43.266 "seek_hole": false, 00:10:43.266 "seek_data": false, 00:10:43.266 "copy": true, 00:10:43.266 "nvme_iov_md": false 00:10:43.266 }, 00:10:43.266 "memory_domains": [ 00:10:43.266 { 00:10:43.266 "dma_device_id": "system", 00:10:43.266 "dma_device_type": 1 00:10:43.266 }, 00:10:43.266 { 00:10:43.266 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:43.266 "dma_device_type": 2 00:10:43.266 } 00:10:43.266 ], 00:10:43.266 "driver_specific": {} 00:10:43.266 } 00:10:43.266 ] 00:10:43.266 17:32:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:43.266 17:32:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:10:43.266 17:32:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:10:43.266 17:32:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:10:43.266 17:32:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -s -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:10:43.266 17:32:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:43.266 17:32:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:43.266 [2024-11-27 17:32:14.289920] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:10:43.266 [2024-11-27 17:32:14.290020] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:10:43.266 [2024-11-27 17:32:14.290067] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:43.266 [2024-11-27 17:32:14.292204] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:43.266 [2024-11-27 17:32:14.292290] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:10:43.266 17:32:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:43.266 17:32:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:43.266 17:32:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:43.266 17:32:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:43.266 17:32:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:43.266 17:32:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:43.266 17:32:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:43.266 17:32:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:43.266 17:32:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:43.266 17:32:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:43.266 17:32:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:43.266 17:32:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:43.266 17:32:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:43.266 17:32:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:43.266 17:32:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:43.266 17:32:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:43.266 17:32:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:43.266 "name": "Existed_Raid", 00:10:43.266 "uuid": "4253209d-2d69-4c9d-98cf-3ba7ccd58600", 00:10:43.266 "strip_size_kb": 64, 00:10:43.266 "state": "configuring", 00:10:43.266 "raid_level": "concat", 00:10:43.266 "superblock": true, 00:10:43.266 "num_base_bdevs": 4, 00:10:43.266 "num_base_bdevs_discovered": 3, 00:10:43.266 "num_base_bdevs_operational": 4, 00:10:43.266 "base_bdevs_list": [ 00:10:43.266 { 00:10:43.266 "name": "BaseBdev1", 00:10:43.266 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:43.266 "is_configured": false, 00:10:43.266 "data_offset": 0, 00:10:43.266 "data_size": 0 00:10:43.266 }, 00:10:43.266 { 00:10:43.266 "name": "BaseBdev2", 00:10:43.266 "uuid": "6d03b07a-708d-48f9-9d9b-3c6aa9450d1f", 00:10:43.266 "is_configured": true, 00:10:43.266 "data_offset": 2048, 00:10:43.266 "data_size": 63488 00:10:43.266 }, 00:10:43.266 { 00:10:43.266 "name": "BaseBdev3", 00:10:43.266 "uuid": "987a8497-3566-482e-9484-33509955d1d2", 00:10:43.266 "is_configured": true, 00:10:43.266 "data_offset": 2048, 00:10:43.266 "data_size": 63488 00:10:43.266 }, 00:10:43.266 { 00:10:43.266 "name": "BaseBdev4", 00:10:43.266 "uuid": "110bc70b-7a7e-4d0b-9f43-2044486e7e8b", 00:10:43.266 "is_configured": true, 00:10:43.266 "data_offset": 2048, 00:10:43.266 "data_size": 63488 00:10:43.266 } 00:10:43.266 ] 00:10:43.266 }' 00:10:43.266 17:32:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:43.266 17:32:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:43.835 17:32:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:10:43.835 17:32:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:43.835 17:32:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:43.835 [2024-11-27 17:32:14.757125] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:10:43.835 17:32:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:43.835 17:32:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:43.835 17:32:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:43.835 17:32:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:43.835 17:32:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:43.835 17:32:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:43.835 17:32:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:43.835 17:32:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:43.835 17:32:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:43.835 17:32:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:43.835 17:32:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:43.835 17:32:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:43.835 17:32:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:43.835 17:32:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:43.835 17:32:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:43.835 17:32:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:43.835 17:32:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:43.835 "name": "Existed_Raid", 00:10:43.835 "uuid": "4253209d-2d69-4c9d-98cf-3ba7ccd58600", 00:10:43.835 "strip_size_kb": 64, 00:10:43.835 "state": "configuring", 00:10:43.835 "raid_level": "concat", 00:10:43.835 "superblock": true, 00:10:43.835 "num_base_bdevs": 4, 00:10:43.835 "num_base_bdevs_discovered": 2, 00:10:43.835 "num_base_bdevs_operational": 4, 00:10:43.835 "base_bdevs_list": [ 00:10:43.835 { 00:10:43.835 "name": "BaseBdev1", 00:10:43.835 "uuid": "00000000-0000-0000-0000-000000000000", 00:10:43.835 "is_configured": false, 00:10:43.835 "data_offset": 0, 00:10:43.835 "data_size": 0 00:10:43.835 }, 00:10:43.835 { 00:10:43.835 "name": null, 00:10:43.835 "uuid": "6d03b07a-708d-48f9-9d9b-3c6aa9450d1f", 00:10:43.835 "is_configured": false, 00:10:43.835 "data_offset": 0, 00:10:43.835 "data_size": 63488 00:10:43.835 }, 00:10:43.835 { 00:10:43.835 "name": "BaseBdev3", 00:10:43.835 "uuid": "987a8497-3566-482e-9484-33509955d1d2", 00:10:43.835 "is_configured": true, 00:10:43.835 "data_offset": 2048, 00:10:43.835 "data_size": 63488 00:10:43.835 }, 00:10:43.835 { 00:10:43.835 "name": "BaseBdev4", 00:10:43.835 "uuid": "110bc70b-7a7e-4d0b-9f43-2044486e7e8b", 00:10:43.835 "is_configured": true, 00:10:43.835 "data_offset": 2048, 00:10:43.835 "data_size": 63488 00:10:43.835 } 00:10:43.835 ] 00:10:43.835 }' 00:10:43.835 17:32:14 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:43.835 17:32:14 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:44.095 17:32:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:44.095 17:32:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:44.095 17:32:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:44.095 17:32:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:10:44.095 17:32:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:44.095 17:32:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:10:44.095 17:32:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:10:44.095 17:32:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:44.095 17:32:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:44.095 [2024-11-27 17:32:15.277048] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:44.095 BaseBdev1 00:10:44.095 17:32:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:44.095 17:32:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:10:44.096 17:32:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:10:44.096 17:32:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:10:44.096 17:32:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:10:44.096 17:32:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:10:44.096 17:32:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:10:44.096 17:32:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:10:44.096 17:32:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:44.096 17:32:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:44.356 17:32:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:44.356 17:32:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:10:44.356 17:32:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:44.356 17:32:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:44.356 [ 00:10:44.356 { 00:10:44.356 "name": "BaseBdev1", 00:10:44.356 "aliases": [ 00:10:44.356 "b91a97ef-5142-4af3-997a-d2c83bbdf055" 00:10:44.356 ], 00:10:44.356 "product_name": "Malloc disk", 00:10:44.356 "block_size": 512, 00:10:44.356 "num_blocks": 65536, 00:10:44.356 "uuid": "b91a97ef-5142-4af3-997a-d2c83bbdf055", 00:10:44.356 "assigned_rate_limits": { 00:10:44.356 "rw_ios_per_sec": 0, 00:10:44.356 "rw_mbytes_per_sec": 0, 00:10:44.356 "r_mbytes_per_sec": 0, 00:10:44.356 "w_mbytes_per_sec": 0 00:10:44.356 }, 00:10:44.356 "claimed": true, 00:10:44.356 "claim_type": "exclusive_write", 00:10:44.356 "zoned": false, 00:10:44.356 "supported_io_types": { 00:10:44.356 "read": true, 00:10:44.356 "write": true, 00:10:44.356 "unmap": true, 00:10:44.356 "flush": true, 00:10:44.356 "reset": true, 00:10:44.356 "nvme_admin": false, 00:10:44.356 "nvme_io": false, 00:10:44.356 "nvme_io_md": false, 00:10:44.356 "write_zeroes": true, 00:10:44.356 "zcopy": true, 00:10:44.356 "get_zone_info": false, 00:10:44.356 "zone_management": false, 00:10:44.356 "zone_append": false, 00:10:44.356 "compare": false, 00:10:44.356 "compare_and_write": false, 00:10:44.356 "abort": true, 00:10:44.356 "seek_hole": false, 00:10:44.356 "seek_data": false, 00:10:44.356 "copy": true, 00:10:44.356 "nvme_iov_md": false 00:10:44.356 }, 00:10:44.356 "memory_domains": [ 00:10:44.356 { 00:10:44.356 "dma_device_id": "system", 00:10:44.356 "dma_device_type": 1 00:10:44.356 }, 00:10:44.356 { 00:10:44.356 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:44.356 "dma_device_type": 2 00:10:44.356 } 00:10:44.356 ], 00:10:44.356 "driver_specific": {} 00:10:44.356 } 00:10:44.356 ] 00:10:44.356 17:32:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:44.356 17:32:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:10:44.356 17:32:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:44.356 17:32:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:44.356 17:32:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:44.356 17:32:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:44.356 17:32:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:44.356 17:32:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:44.356 17:32:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:44.356 17:32:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:44.356 17:32:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:44.356 17:32:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:44.356 17:32:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:44.356 17:32:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:44.356 17:32:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:44.356 17:32:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:44.356 17:32:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:44.356 17:32:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:44.356 "name": "Existed_Raid", 00:10:44.356 "uuid": "4253209d-2d69-4c9d-98cf-3ba7ccd58600", 00:10:44.356 "strip_size_kb": 64, 00:10:44.356 "state": "configuring", 00:10:44.356 "raid_level": "concat", 00:10:44.356 "superblock": true, 00:10:44.356 "num_base_bdevs": 4, 00:10:44.356 "num_base_bdevs_discovered": 3, 00:10:44.356 "num_base_bdevs_operational": 4, 00:10:44.356 "base_bdevs_list": [ 00:10:44.356 { 00:10:44.356 "name": "BaseBdev1", 00:10:44.356 "uuid": "b91a97ef-5142-4af3-997a-d2c83bbdf055", 00:10:44.356 "is_configured": true, 00:10:44.356 "data_offset": 2048, 00:10:44.356 "data_size": 63488 00:10:44.356 }, 00:10:44.356 { 00:10:44.356 "name": null, 00:10:44.356 "uuid": "6d03b07a-708d-48f9-9d9b-3c6aa9450d1f", 00:10:44.356 "is_configured": false, 00:10:44.356 "data_offset": 0, 00:10:44.356 "data_size": 63488 00:10:44.356 }, 00:10:44.356 { 00:10:44.356 "name": "BaseBdev3", 00:10:44.356 "uuid": "987a8497-3566-482e-9484-33509955d1d2", 00:10:44.356 "is_configured": true, 00:10:44.356 "data_offset": 2048, 00:10:44.356 "data_size": 63488 00:10:44.356 }, 00:10:44.356 { 00:10:44.356 "name": "BaseBdev4", 00:10:44.356 "uuid": "110bc70b-7a7e-4d0b-9f43-2044486e7e8b", 00:10:44.356 "is_configured": true, 00:10:44.356 "data_offset": 2048, 00:10:44.356 "data_size": 63488 00:10:44.356 } 00:10:44.356 ] 00:10:44.356 }' 00:10:44.356 17:32:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:44.356 17:32:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:44.926 17:32:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:44.926 17:32:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:44.926 17:32:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:44.926 17:32:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:10:44.926 17:32:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:44.926 17:32:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:10:44.926 17:32:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:10:44.926 17:32:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:44.926 17:32:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:44.926 [2024-11-27 17:32:15.860062] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:10:44.926 17:32:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:44.926 17:32:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:44.926 17:32:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:44.926 17:32:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:44.926 17:32:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:44.926 17:32:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:44.926 17:32:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:44.926 17:32:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:44.926 17:32:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:44.926 17:32:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:44.926 17:32:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:44.926 17:32:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:44.926 17:32:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:44.926 17:32:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:44.926 17:32:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:44.926 17:32:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:44.926 17:32:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:44.926 "name": "Existed_Raid", 00:10:44.926 "uuid": "4253209d-2d69-4c9d-98cf-3ba7ccd58600", 00:10:44.926 "strip_size_kb": 64, 00:10:44.926 "state": "configuring", 00:10:44.926 "raid_level": "concat", 00:10:44.926 "superblock": true, 00:10:44.926 "num_base_bdevs": 4, 00:10:44.926 "num_base_bdevs_discovered": 2, 00:10:44.926 "num_base_bdevs_operational": 4, 00:10:44.926 "base_bdevs_list": [ 00:10:44.926 { 00:10:44.926 "name": "BaseBdev1", 00:10:44.926 "uuid": "b91a97ef-5142-4af3-997a-d2c83bbdf055", 00:10:44.926 "is_configured": true, 00:10:44.926 "data_offset": 2048, 00:10:44.926 "data_size": 63488 00:10:44.926 }, 00:10:44.926 { 00:10:44.926 "name": null, 00:10:44.926 "uuid": "6d03b07a-708d-48f9-9d9b-3c6aa9450d1f", 00:10:44.926 "is_configured": false, 00:10:44.926 "data_offset": 0, 00:10:44.926 "data_size": 63488 00:10:44.926 }, 00:10:44.926 { 00:10:44.926 "name": null, 00:10:44.926 "uuid": "987a8497-3566-482e-9484-33509955d1d2", 00:10:44.926 "is_configured": false, 00:10:44.926 "data_offset": 0, 00:10:44.926 "data_size": 63488 00:10:44.926 }, 00:10:44.926 { 00:10:44.926 "name": "BaseBdev4", 00:10:44.926 "uuid": "110bc70b-7a7e-4d0b-9f43-2044486e7e8b", 00:10:44.926 "is_configured": true, 00:10:44.926 "data_offset": 2048, 00:10:44.926 "data_size": 63488 00:10:44.926 } 00:10:44.926 ] 00:10:44.926 }' 00:10:44.926 17:32:15 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:44.926 17:32:15 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:45.186 17:32:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:10:45.186 17:32:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:45.186 17:32:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:45.186 17:32:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:45.186 17:32:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:45.186 17:32:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:10:45.186 17:32:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:10:45.186 17:32:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:45.186 17:32:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:45.186 [2024-11-27 17:32:16.331313] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:45.186 17:32:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:45.186 17:32:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:45.186 17:32:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:45.186 17:32:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:45.186 17:32:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:45.186 17:32:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:45.186 17:32:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:45.186 17:32:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:45.186 17:32:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:45.186 17:32:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:45.186 17:32:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:45.186 17:32:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:45.186 17:32:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:45.186 17:32:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:45.186 17:32:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:45.186 17:32:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:45.446 17:32:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:45.446 "name": "Existed_Raid", 00:10:45.446 "uuid": "4253209d-2d69-4c9d-98cf-3ba7ccd58600", 00:10:45.446 "strip_size_kb": 64, 00:10:45.446 "state": "configuring", 00:10:45.446 "raid_level": "concat", 00:10:45.446 "superblock": true, 00:10:45.446 "num_base_bdevs": 4, 00:10:45.446 "num_base_bdevs_discovered": 3, 00:10:45.446 "num_base_bdevs_operational": 4, 00:10:45.446 "base_bdevs_list": [ 00:10:45.446 { 00:10:45.446 "name": "BaseBdev1", 00:10:45.446 "uuid": "b91a97ef-5142-4af3-997a-d2c83bbdf055", 00:10:45.446 "is_configured": true, 00:10:45.446 "data_offset": 2048, 00:10:45.446 "data_size": 63488 00:10:45.446 }, 00:10:45.446 { 00:10:45.446 "name": null, 00:10:45.446 "uuid": "6d03b07a-708d-48f9-9d9b-3c6aa9450d1f", 00:10:45.446 "is_configured": false, 00:10:45.446 "data_offset": 0, 00:10:45.446 "data_size": 63488 00:10:45.446 }, 00:10:45.446 { 00:10:45.446 "name": "BaseBdev3", 00:10:45.446 "uuid": "987a8497-3566-482e-9484-33509955d1d2", 00:10:45.446 "is_configured": true, 00:10:45.446 "data_offset": 2048, 00:10:45.446 "data_size": 63488 00:10:45.446 }, 00:10:45.446 { 00:10:45.446 "name": "BaseBdev4", 00:10:45.446 "uuid": "110bc70b-7a7e-4d0b-9f43-2044486e7e8b", 00:10:45.446 "is_configured": true, 00:10:45.446 "data_offset": 2048, 00:10:45.446 "data_size": 63488 00:10:45.446 } 00:10:45.446 ] 00:10:45.446 }' 00:10:45.446 17:32:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:45.446 17:32:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:45.706 17:32:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:45.706 17:32:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:45.706 17:32:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:10:45.706 17:32:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:45.706 17:32:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:45.706 17:32:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:10:45.706 17:32:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:10:45.706 17:32:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:45.706 17:32:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:45.706 [2024-11-27 17:32:16.774561] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:10:45.706 17:32:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:45.706 17:32:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:45.706 17:32:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:45.706 17:32:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:45.706 17:32:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:45.706 17:32:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:45.706 17:32:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:45.706 17:32:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:45.706 17:32:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:45.706 17:32:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:45.706 17:32:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:45.706 17:32:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:45.706 17:32:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:45.706 17:32:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:45.706 17:32:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:45.706 17:32:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:45.706 17:32:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:45.706 "name": "Existed_Raid", 00:10:45.706 "uuid": "4253209d-2d69-4c9d-98cf-3ba7ccd58600", 00:10:45.706 "strip_size_kb": 64, 00:10:45.706 "state": "configuring", 00:10:45.706 "raid_level": "concat", 00:10:45.706 "superblock": true, 00:10:45.706 "num_base_bdevs": 4, 00:10:45.706 "num_base_bdevs_discovered": 2, 00:10:45.706 "num_base_bdevs_operational": 4, 00:10:45.706 "base_bdevs_list": [ 00:10:45.706 { 00:10:45.706 "name": null, 00:10:45.706 "uuid": "b91a97ef-5142-4af3-997a-d2c83bbdf055", 00:10:45.706 "is_configured": false, 00:10:45.706 "data_offset": 0, 00:10:45.706 "data_size": 63488 00:10:45.706 }, 00:10:45.706 { 00:10:45.706 "name": null, 00:10:45.706 "uuid": "6d03b07a-708d-48f9-9d9b-3c6aa9450d1f", 00:10:45.706 "is_configured": false, 00:10:45.706 "data_offset": 0, 00:10:45.706 "data_size": 63488 00:10:45.706 }, 00:10:45.706 { 00:10:45.706 "name": "BaseBdev3", 00:10:45.707 "uuid": "987a8497-3566-482e-9484-33509955d1d2", 00:10:45.707 "is_configured": true, 00:10:45.707 "data_offset": 2048, 00:10:45.707 "data_size": 63488 00:10:45.707 }, 00:10:45.707 { 00:10:45.707 "name": "BaseBdev4", 00:10:45.707 "uuid": "110bc70b-7a7e-4d0b-9f43-2044486e7e8b", 00:10:45.707 "is_configured": true, 00:10:45.707 "data_offset": 2048, 00:10:45.707 "data_size": 63488 00:10:45.707 } 00:10:45.707 ] 00:10:45.707 }' 00:10:45.707 17:32:16 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:45.707 17:32:16 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:46.276 17:32:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:46.276 17:32:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:46.276 17:32:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:46.276 17:32:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:10:46.276 17:32:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:46.276 17:32:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:10:46.276 17:32:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:10:46.276 17:32:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:46.276 17:32:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:46.276 [2024-11-27 17:32:17.289397] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:46.276 17:32:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:46.276 17:32:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring concat 64 4 00:10:46.276 17:32:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:46.276 17:32:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:46.276 17:32:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:46.276 17:32:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:46.276 17:32:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:46.276 17:32:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:46.276 17:32:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:46.276 17:32:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:46.276 17:32:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:46.276 17:32:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:46.276 17:32:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:46.276 17:32:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:46.276 17:32:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:46.276 17:32:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:46.276 17:32:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:46.276 "name": "Existed_Raid", 00:10:46.276 "uuid": "4253209d-2d69-4c9d-98cf-3ba7ccd58600", 00:10:46.276 "strip_size_kb": 64, 00:10:46.276 "state": "configuring", 00:10:46.276 "raid_level": "concat", 00:10:46.276 "superblock": true, 00:10:46.276 "num_base_bdevs": 4, 00:10:46.276 "num_base_bdevs_discovered": 3, 00:10:46.276 "num_base_bdevs_operational": 4, 00:10:46.276 "base_bdevs_list": [ 00:10:46.276 { 00:10:46.276 "name": null, 00:10:46.276 "uuid": "b91a97ef-5142-4af3-997a-d2c83bbdf055", 00:10:46.276 "is_configured": false, 00:10:46.276 "data_offset": 0, 00:10:46.276 "data_size": 63488 00:10:46.276 }, 00:10:46.276 { 00:10:46.276 "name": "BaseBdev2", 00:10:46.276 "uuid": "6d03b07a-708d-48f9-9d9b-3c6aa9450d1f", 00:10:46.276 "is_configured": true, 00:10:46.276 "data_offset": 2048, 00:10:46.276 "data_size": 63488 00:10:46.276 }, 00:10:46.276 { 00:10:46.276 "name": "BaseBdev3", 00:10:46.276 "uuid": "987a8497-3566-482e-9484-33509955d1d2", 00:10:46.276 "is_configured": true, 00:10:46.276 "data_offset": 2048, 00:10:46.276 "data_size": 63488 00:10:46.276 }, 00:10:46.276 { 00:10:46.276 "name": "BaseBdev4", 00:10:46.276 "uuid": "110bc70b-7a7e-4d0b-9f43-2044486e7e8b", 00:10:46.276 "is_configured": true, 00:10:46.276 "data_offset": 2048, 00:10:46.276 "data_size": 63488 00:10:46.276 } 00:10:46.276 ] 00:10:46.276 }' 00:10:46.276 17:32:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:46.276 17:32:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:46.846 17:32:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:46.846 17:32:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:46.846 17:32:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:10:46.846 17:32:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:46.846 17:32:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:46.846 17:32:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:10:46.846 17:32:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:46.846 17:32:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:46.846 17:32:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:46.846 17:32:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:10:46.846 17:32:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:46.846 17:32:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u b91a97ef-5142-4af3-997a-d2c83bbdf055 00:10:46.846 17:32:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:46.846 17:32:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:46.846 [2024-11-27 17:32:17.844988] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:10:46.846 [2024-11-27 17:32:17.845290] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:10:46.846 [2024-11-27 17:32:17.845307] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:10:46.846 [2024-11-27 17:32:17.845600] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002a10 00:10:46.846 [2024-11-27 17:32:17.845720] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:10:46.846 [2024-11-27 17:32:17.845732] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001c80 00:10:46.846 NewBaseBdev 00:10:46.846 [2024-11-27 17:32:17.845832] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:46.846 17:32:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:46.846 17:32:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:10:46.846 17:32:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=NewBaseBdev 00:10:46.846 17:32:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:10:46.846 17:32:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:10:46.846 17:32:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:10:46.846 17:32:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:10:46.846 17:32:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:10:46.846 17:32:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:46.846 17:32:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:46.846 17:32:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:46.846 17:32:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:10:46.846 17:32:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:46.846 17:32:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:46.846 [ 00:10:46.846 { 00:10:46.846 "name": "NewBaseBdev", 00:10:46.846 "aliases": [ 00:10:46.846 "b91a97ef-5142-4af3-997a-d2c83bbdf055" 00:10:46.846 ], 00:10:46.846 "product_name": "Malloc disk", 00:10:46.846 "block_size": 512, 00:10:46.846 "num_blocks": 65536, 00:10:46.846 "uuid": "b91a97ef-5142-4af3-997a-d2c83bbdf055", 00:10:46.846 "assigned_rate_limits": { 00:10:46.846 "rw_ios_per_sec": 0, 00:10:46.846 "rw_mbytes_per_sec": 0, 00:10:46.846 "r_mbytes_per_sec": 0, 00:10:46.846 "w_mbytes_per_sec": 0 00:10:46.846 }, 00:10:46.846 "claimed": true, 00:10:46.846 "claim_type": "exclusive_write", 00:10:46.846 "zoned": false, 00:10:46.846 "supported_io_types": { 00:10:46.846 "read": true, 00:10:46.846 "write": true, 00:10:46.846 "unmap": true, 00:10:46.846 "flush": true, 00:10:46.846 "reset": true, 00:10:46.846 "nvme_admin": false, 00:10:46.846 "nvme_io": false, 00:10:46.846 "nvme_io_md": false, 00:10:46.846 "write_zeroes": true, 00:10:46.846 "zcopy": true, 00:10:46.846 "get_zone_info": false, 00:10:46.846 "zone_management": false, 00:10:46.846 "zone_append": false, 00:10:46.846 "compare": false, 00:10:46.846 "compare_and_write": false, 00:10:46.846 "abort": true, 00:10:46.846 "seek_hole": false, 00:10:46.846 "seek_data": false, 00:10:46.846 "copy": true, 00:10:46.846 "nvme_iov_md": false 00:10:46.846 }, 00:10:46.846 "memory_domains": [ 00:10:46.846 { 00:10:46.846 "dma_device_id": "system", 00:10:46.846 "dma_device_type": 1 00:10:46.846 }, 00:10:46.846 { 00:10:46.846 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:46.846 "dma_device_type": 2 00:10:46.846 } 00:10:46.846 ], 00:10:46.846 "driver_specific": {} 00:10:46.846 } 00:10:46.846 ] 00:10:46.846 17:32:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:46.846 17:32:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:10:46.846 17:32:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online concat 64 4 00:10:46.846 17:32:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:10:46.846 17:32:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:46.846 17:32:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:46.846 17:32:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:46.846 17:32:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:46.846 17:32:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:46.846 17:32:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:46.846 17:32:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:46.847 17:32:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:46.847 17:32:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:46.847 17:32:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:10:46.847 17:32:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:46.847 17:32:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:46.847 17:32:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:46.847 17:32:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:46.847 "name": "Existed_Raid", 00:10:46.847 "uuid": "4253209d-2d69-4c9d-98cf-3ba7ccd58600", 00:10:46.847 "strip_size_kb": 64, 00:10:46.847 "state": "online", 00:10:46.847 "raid_level": "concat", 00:10:46.847 "superblock": true, 00:10:46.847 "num_base_bdevs": 4, 00:10:46.847 "num_base_bdevs_discovered": 4, 00:10:46.847 "num_base_bdevs_operational": 4, 00:10:46.847 "base_bdevs_list": [ 00:10:46.847 { 00:10:46.847 "name": "NewBaseBdev", 00:10:46.847 "uuid": "b91a97ef-5142-4af3-997a-d2c83bbdf055", 00:10:46.847 "is_configured": true, 00:10:46.847 "data_offset": 2048, 00:10:46.847 "data_size": 63488 00:10:46.847 }, 00:10:46.847 { 00:10:46.847 "name": "BaseBdev2", 00:10:46.847 "uuid": "6d03b07a-708d-48f9-9d9b-3c6aa9450d1f", 00:10:46.847 "is_configured": true, 00:10:46.847 "data_offset": 2048, 00:10:46.847 "data_size": 63488 00:10:46.847 }, 00:10:46.847 { 00:10:46.847 "name": "BaseBdev3", 00:10:46.847 "uuid": "987a8497-3566-482e-9484-33509955d1d2", 00:10:46.847 "is_configured": true, 00:10:46.847 "data_offset": 2048, 00:10:46.847 "data_size": 63488 00:10:46.847 }, 00:10:46.847 { 00:10:46.847 "name": "BaseBdev4", 00:10:46.847 "uuid": "110bc70b-7a7e-4d0b-9f43-2044486e7e8b", 00:10:46.847 "is_configured": true, 00:10:46.847 "data_offset": 2048, 00:10:46.847 "data_size": 63488 00:10:46.847 } 00:10:46.847 ] 00:10:46.847 }' 00:10:46.847 17:32:17 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:46.847 17:32:17 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:47.107 17:32:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:10:47.107 17:32:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:10:47.107 17:32:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:10:47.107 17:32:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:10:47.107 17:32:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:10:47.107 17:32:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:10:47.107 17:32:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:10:47.107 17:32:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:10:47.107 17:32:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:47.107 17:32:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:47.107 [2024-11-27 17:32:18.292547] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:47.367 17:32:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:47.367 17:32:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:47.367 "name": "Existed_Raid", 00:10:47.367 "aliases": [ 00:10:47.367 "4253209d-2d69-4c9d-98cf-3ba7ccd58600" 00:10:47.367 ], 00:10:47.367 "product_name": "Raid Volume", 00:10:47.367 "block_size": 512, 00:10:47.367 "num_blocks": 253952, 00:10:47.367 "uuid": "4253209d-2d69-4c9d-98cf-3ba7ccd58600", 00:10:47.367 "assigned_rate_limits": { 00:10:47.367 "rw_ios_per_sec": 0, 00:10:47.367 "rw_mbytes_per_sec": 0, 00:10:47.367 "r_mbytes_per_sec": 0, 00:10:47.367 "w_mbytes_per_sec": 0 00:10:47.367 }, 00:10:47.367 "claimed": false, 00:10:47.367 "zoned": false, 00:10:47.367 "supported_io_types": { 00:10:47.367 "read": true, 00:10:47.367 "write": true, 00:10:47.367 "unmap": true, 00:10:47.367 "flush": true, 00:10:47.367 "reset": true, 00:10:47.367 "nvme_admin": false, 00:10:47.367 "nvme_io": false, 00:10:47.367 "nvme_io_md": false, 00:10:47.367 "write_zeroes": true, 00:10:47.367 "zcopy": false, 00:10:47.367 "get_zone_info": false, 00:10:47.367 "zone_management": false, 00:10:47.367 "zone_append": false, 00:10:47.367 "compare": false, 00:10:47.367 "compare_and_write": false, 00:10:47.367 "abort": false, 00:10:47.367 "seek_hole": false, 00:10:47.367 "seek_data": false, 00:10:47.367 "copy": false, 00:10:47.367 "nvme_iov_md": false 00:10:47.367 }, 00:10:47.367 "memory_domains": [ 00:10:47.367 { 00:10:47.367 "dma_device_id": "system", 00:10:47.367 "dma_device_type": 1 00:10:47.367 }, 00:10:47.367 { 00:10:47.367 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:47.367 "dma_device_type": 2 00:10:47.367 }, 00:10:47.367 { 00:10:47.367 "dma_device_id": "system", 00:10:47.367 "dma_device_type": 1 00:10:47.367 }, 00:10:47.367 { 00:10:47.367 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:47.367 "dma_device_type": 2 00:10:47.367 }, 00:10:47.367 { 00:10:47.367 "dma_device_id": "system", 00:10:47.367 "dma_device_type": 1 00:10:47.367 }, 00:10:47.367 { 00:10:47.367 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:47.367 "dma_device_type": 2 00:10:47.367 }, 00:10:47.367 { 00:10:47.367 "dma_device_id": "system", 00:10:47.367 "dma_device_type": 1 00:10:47.367 }, 00:10:47.367 { 00:10:47.367 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:47.367 "dma_device_type": 2 00:10:47.367 } 00:10:47.367 ], 00:10:47.367 "driver_specific": { 00:10:47.367 "raid": { 00:10:47.367 "uuid": "4253209d-2d69-4c9d-98cf-3ba7ccd58600", 00:10:47.367 "strip_size_kb": 64, 00:10:47.367 "state": "online", 00:10:47.367 "raid_level": "concat", 00:10:47.367 "superblock": true, 00:10:47.367 "num_base_bdevs": 4, 00:10:47.367 "num_base_bdevs_discovered": 4, 00:10:47.367 "num_base_bdevs_operational": 4, 00:10:47.367 "base_bdevs_list": [ 00:10:47.367 { 00:10:47.367 "name": "NewBaseBdev", 00:10:47.367 "uuid": "b91a97ef-5142-4af3-997a-d2c83bbdf055", 00:10:47.367 "is_configured": true, 00:10:47.367 "data_offset": 2048, 00:10:47.367 "data_size": 63488 00:10:47.367 }, 00:10:47.367 { 00:10:47.367 "name": "BaseBdev2", 00:10:47.367 "uuid": "6d03b07a-708d-48f9-9d9b-3c6aa9450d1f", 00:10:47.367 "is_configured": true, 00:10:47.367 "data_offset": 2048, 00:10:47.367 "data_size": 63488 00:10:47.367 }, 00:10:47.367 { 00:10:47.367 "name": "BaseBdev3", 00:10:47.367 "uuid": "987a8497-3566-482e-9484-33509955d1d2", 00:10:47.367 "is_configured": true, 00:10:47.367 "data_offset": 2048, 00:10:47.367 "data_size": 63488 00:10:47.367 }, 00:10:47.367 { 00:10:47.367 "name": "BaseBdev4", 00:10:47.367 "uuid": "110bc70b-7a7e-4d0b-9f43-2044486e7e8b", 00:10:47.367 "is_configured": true, 00:10:47.367 "data_offset": 2048, 00:10:47.367 "data_size": 63488 00:10:47.367 } 00:10:47.367 ] 00:10:47.367 } 00:10:47.367 } 00:10:47.367 }' 00:10:47.367 17:32:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:47.367 17:32:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:10:47.367 BaseBdev2 00:10:47.367 BaseBdev3 00:10:47.367 BaseBdev4' 00:10:47.367 17:32:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:47.367 17:32:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:47.367 17:32:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:47.367 17:32:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:10:47.367 17:32:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:47.367 17:32:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:47.367 17:32:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:47.367 17:32:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:47.367 17:32:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:47.367 17:32:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:47.367 17:32:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:47.367 17:32:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:47.367 17:32:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:10:47.367 17:32:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:47.367 17:32:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:47.367 17:32:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:47.367 17:32:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:47.367 17:32:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:47.367 17:32:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:47.367 17:32:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:10:47.368 17:32:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:47.368 17:32:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:47.368 17:32:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:47.368 17:32:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:47.628 17:32:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:47.628 17:32:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:47.628 17:32:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:47.628 17:32:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:47.628 17:32:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:10:47.628 17:32:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:47.628 17:32:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:47.628 17:32:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:47.628 17:32:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:47.628 17:32:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:47.628 17:32:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:10:47.628 17:32:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:47.628 17:32:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:47.628 [2024-11-27 17:32:18.623655] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:10:47.628 [2024-11-27 17:32:18.623727] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:47.628 [2024-11-27 17:32:18.623824] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:47.628 [2024-11-27 17:32:18.623921] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:47.628 [2024-11-27 17:32:18.623962] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name Existed_Raid, state offline 00:10:47.628 17:32:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:47.628 17:32:18 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 82675 00:10:47.628 17:32:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@950 -- # '[' -z 82675 ']' 00:10:47.628 17:32:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # kill -0 82675 00:10:47.628 17:32:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@955 -- # uname 00:10:47.628 17:32:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:10:47.628 17:32:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 82675 00:10:47.628 killing process with pid 82675 00:10:47.628 17:32:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:10:47.628 17:32:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:10:47.628 17:32:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@968 -- # echo 'killing process with pid 82675' 00:10:47.628 17:32:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@969 -- # kill 82675 00:10:47.628 [2024-11-27 17:32:18.672282] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:10:47.628 17:32:18 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@974 -- # wait 82675 00:10:47.628 [2024-11-27 17:32:18.747344] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:10:48.206 ************************************ 00:10:48.206 END TEST raid_state_function_test_sb 00:10:48.206 ************************************ 00:10:48.206 17:32:19 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:10:48.206 00:10:48.206 real 0m9.785s 00:10:48.206 user 0m16.385s 00:10:48.206 sys 0m2.120s 00:10:48.206 17:32:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1126 -- # xtrace_disable 00:10:48.206 17:32:19 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:10:48.206 17:32:19 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test concat 4 00:10:48.206 17:32:19 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:10:48.206 17:32:19 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:10:48.206 17:32:19 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:10:48.206 ************************************ 00:10:48.206 START TEST raid_superblock_test 00:10:48.206 ************************************ 00:10:48.206 17:32:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1125 -- # raid_superblock_test concat 4 00:10:48.206 17:32:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=concat 00:10:48.206 17:32:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=4 00:10:48.206 17:32:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:10:48.206 17:32:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:10:48.206 17:32:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:10:48.206 17:32:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:10:48.206 17:32:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:10:48.206 17:32:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:10:48.206 17:32:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:10:48.206 17:32:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:10:48.206 17:32:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:10:48.206 17:32:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:10:48.206 17:32:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:10:48.207 17:32:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' concat '!=' raid1 ']' 00:10:48.207 17:32:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@405 -- # strip_size=64 00:10:48.207 17:32:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@406 -- # strip_size_create_arg='-z 64' 00:10:48.207 17:32:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=83327 00:10:48.207 17:32:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:10:48.207 17:32:19 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 83327 00:10:48.207 17:32:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@831 -- # '[' -z 83327 ']' 00:10:48.207 17:32:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:48.207 17:32:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:10:48.207 17:32:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:48.207 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:48.207 17:32:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:10:48.207 17:32:19 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:48.207 [2024-11-27 17:32:19.273853] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:10:48.207 [2024-11-27 17:32:19.274064] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid83327 ] 00:10:48.472 [2024-11-27 17:32:19.399717] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:48.472 [2024-11-27 17:32:19.466883] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:10:48.472 [2024-11-27 17:32:19.542494] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:48.472 [2024-11-27 17:32:19.542633] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:49.041 17:32:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:10:49.041 17:32:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@864 -- # return 0 00:10:49.041 17:32:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:10:49.041 17:32:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:10:49.041 17:32:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:10:49.041 17:32:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:10:49.041 17:32:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:10:49.041 17:32:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:10:49.041 17:32:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:10:49.041 17:32:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:10:49.041 17:32:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:10:49.041 17:32:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:49.041 17:32:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:49.041 malloc1 00:10:49.041 17:32:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:49.041 17:32:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:10:49.041 17:32:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:49.041 17:32:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:49.041 [2024-11-27 17:32:20.128498] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:10:49.042 [2024-11-27 17:32:20.128603] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:49.042 [2024-11-27 17:32:20.128660] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:10:49.042 [2024-11-27 17:32:20.128709] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:49.042 [2024-11-27 17:32:20.131207] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:49.042 [2024-11-27 17:32:20.131275] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:10:49.042 pt1 00:10:49.042 17:32:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:49.042 17:32:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:10:49.042 17:32:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:10:49.042 17:32:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:10:49.042 17:32:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:10:49.042 17:32:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:10:49.042 17:32:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:10:49.042 17:32:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:10:49.042 17:32:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:10:49.042 17:32:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:10:49.042 17:32:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:49.042 17:32:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:49.042 malloc2 00:10:49.042 17:32:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:49.042 17:32:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:10:49.042 17:32:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:49.042 17:32:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:49.042 [2024-11-27 17:32:20.183206] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:10:49.042 [2024-11-27 17:32:20.183386] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:49.042 [2024-11-27 17:32:20.183466] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:10:49.042 [2024-11-27 17:32:20.183546] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:49.042 [2024-11-27 17:32:20.188325] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:49.042 [2024-11-27 17:32:20.188468] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:10:49.042 pt2 00:10:49.042 17:32:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:49.042 17:32:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:10:49.042 17:32:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:10:49.042 17:32:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc3 00:10:49.042 17:32:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt3 00:10:49.042 17:32:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000003 00:10:49.042 17:32:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:10:49.042 17:32:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:10:49.042 17:32:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:10:49.042 17:32:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc3 00:10:49.042 17:32:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:49.042 17:32:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:49.042 malloc3 00:10:49.042 17:32:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:49.042 17:32:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:10:49.042 17:32:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:49.042 17:32:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:49.042 [2024-11-27 17:32:20.223325] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:10:49.042 [2024-11-27 17:32:20.223379] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:49.042 [2024-11-27 17:32:20.223417] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:10:49.042 [2024-11-27 17:32:20.223428] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:49.042 [2024-11-27 17:32:20.225811] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:49.042 [2024-11-27 17:32:20.225847] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:10:49.302 pt3 00:10:49.302 17:32:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:49.302 17:32:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:10:49.302 17:32:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:10:49.302 17:32:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc4 00:10:49.302 17:32:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt4 00:10:49.302 17:32:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000004 00:10:49.302 17:32:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:10:49.302 17:32:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:10:49.302 17:32:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:10:49.302 17:32:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc4 00:10:49.302 17:32:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:49.302 17:32:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:49.302 malloc4 00:10:49.302 17:32:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:49.302 17:32:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:10:49.302 17:32:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:49.302 17:32:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:49.302 [2024-11-27 17:32:20.258062] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:10:49.302 [2024-11-27 17:32:20.258176] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:49.302 [2024-11-27 17:32:20.258222] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:10:49.302 [2024-11-27 17:32:20.258257] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:49.302 [2024-11-27 17:32:20.260721] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:49.302 [2024-11-27 17:32:20.260789] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:10:49.302 pt4 00:10:49.302 17:32:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:49.302 17:32:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:10:49.302 17:32:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:10:49.302 17:32:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''pt1 pt2 pt3 pt4'\''' -n raid_bdev1 -s 00:10:49.302 17:32:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:49.302 17:32:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:49.302 [2024-11-27 17:32:20.270093] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:10:49.302 [2024-11-27 17:32:20.272352] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:10:49.302 [2024-11-27 17:32:20.272457] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:10:49.302 [2024-11-27 17:32:20.272533] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:10:49.302 [2024-11-27 17:32:20.272719] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:10:49.302 [2024-11-27 17:32:20.272767] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:10:49.302 [2024-11-27 17:32:20.273043] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:10:49.302 [2024-11-27 17:32:20.273244] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:10:49.302 [2024-11-27 17:32:20.273286] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:10:49.302 [2024-11-27 17:32:20.273450] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:49.302 17:32:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:49.302 17:32:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online concat 64 4 00:10:49.302 17:32:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:49.302 17:32:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:49.302 17:32:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:49.302 17:32:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:49.302 17:32:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:49.302 17:32:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:49.302 17:32:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:49.302 17:32:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:49.302 17:32:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:49.302 17:32:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:49.302 17:32:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:49.302 17:32:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:49.302 17:32:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:49.302 17:32:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:49.302 17:32:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:49.302 "name": "raid_bdev1", 00:10:49.302 "uuid": "ed8af488-9e31-42ea-b9b7-9904ce04ad33", 00:10:49.302 "strip_size_kb": 64, 00:10:49.302 "state": "online", 00:10:49.302 "raid_level": "concat", 00:10:49.302 "superblock": true, 00:10:49.302 "num_base_bdevs": 4, 00:10:49.302 "num_base_bdevs_discovered": 4, 00:10:49.302 "num_base_bdevs_operational": 4, 00:10:49.302 "base_bdevs_list": [ 00:10:49.302 { 00:10:49.302 "name": "pt1", 00:10:49.302 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:49.302 "is_configured": true, 00:10:49.302 "data_offset": 2048, 00:10:49.302 "data_size": 63488 00:10:49.302 }, 00:10:49.302 { 00:10:49.302 "name": "pt2", 00:10:49.302 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:49.302 "is_configured": true, 00:10:49.302 "data_offset": 2048, 00:10:49.302 "data_size": 63488 00:10:49.302 }, 00:10:49.302 { 00:10:49.302 "name": "pt3", 00:10:49.302 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:49.302 "is_configured": true, 00:10:49.302 "data_offset": 2048, 00:10:49.302 "data_size": 63488 00:10:49.302 }, 00:10:49.302 { 00:10:49.302 "name": "pt4", 00:10:49.302 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:49.302 "is_configured": true, 00:10:49.302 "data_offset": 2048, 00:10:49.302 "data_size": 63488 00:10:49.302 } 00:10:49.302 ] 00:10:49.302 }' 00:10:49.303 17:32:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:49.303 17:32:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:49.561 17:32:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:10:49.561 17:32:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:10:49.561 17:32:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:10:49.562 17:32:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:10:49.562 17:32:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:10:49.562 17:32:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:10:49.562 17:32:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:10:49.562 17:32:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:10:49.562 17:32:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:49.562 17:32:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:49.562 [2024-11-27 17:32:20.721603] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:49.562 17:32:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:49.822 17:32:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:49.822 "name": "raid_bdev1", 00:10:49.822 "aliases": [ 00:10:49.822 "ed8af488-9e31-42ea-b9b7-9904ce04ad33" 00:10:49.822 ], 00:10:49.822 "product_name": "Raid Volume", 00:10:49.822 "block_size": 512, 00:10:49.822 "num_blocks": 253952, 00:10:49.822 "uuid": "ed8af488-9e31-42ea-b9b7-9904ce04ad33", 00:10:49.822 "assigned_rate_limits": { 00:10:49.822 "rw_ios_per_sec": 0, 00:10:49.822 "rw_mbytes_per_sec": 0, 00:10:49.822 "r_mbytes_per_sec": 0, 00:10:49.822 "w_mbytes_per_sec": 0 00:10:49.822 }, 00:10:49.822 "claimed": false, 00:10:49.822 "zoned": false, 00:10:49.822 "supported_io_types": { 00:10:49.822 "read": true, 00:10:49.822 "write": true, 00:10:49.822 "unmap": true, 00:10:49.822 "flush": true, 00:10:49.822 "reset": true, 00:10:49.822 "nvme_admin": false, 00:10:49.822 "nvme_io": false, 00:10:49.822 "nvme_io_md": false, 00:10:49.822 "write_zeroes": true, 00:10:49.822 "zcopy": false, 00:10:49.822 "get_zone_info": false, 00:10:49.822 "zone_management": false, 00:10:49.822 "zone_append": false, 00:10:49.822 "compare": false, 00:10:49.822 "compare_and_write": false, 00:10:49.822 "abort": false, 00:10:49.822 "seek_hole": false, 00:10:49.822 "seek_data": false, 00:10:49.822 "copy": false, 00:10:49.822 "nvme_iov_md": false 00:10:49.822 }, 00:10:49.822 "memory_domains": [ 00:10:49.822 { 00:10:49.822 "dma_device_id": "system", 00:10:49.822 "dma_device_type": 1 00:10:49.822 }, 00:10:49.822 { 00:10:49.822 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:49.822 "dma_device_type": 2 00:10:49.822 }, 00:10:49.822 { 00:10:49.822 "dma_device_id": "system", 00:10:49.822 "dma_device_type": 1 00:10:49.822 }, 00:10:49.822 { 00:10:49.822 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:49.822 "dma_device_type": 2 00:10:49.822 }, 00:10:49.822 { 00:10:49.822 "dma_device_id": "system", 00:10:49.822 "dma_device_type": 1 00:10:49.822 }, 00:10:49.822 { 00:10:49.822 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:49.822 "dma_device_type": 2 00:10:49.822 }, 00:10:49.822 { 00:10:49.822 "dma_device_id": "system", 00:10:49.822 "dma_device_type": 1 00:10:49.822 }, 00:10:49.822 { 00:10:49.822 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:49.822 "dma_device_type": 2 00:10:49.822 } 00:10:49.822 ], 00:10:49.822 "driver_specific": { 00:10:49.822 "raid": { 00:10:49.822 "uuid": "ed8af488-9e31-42ea-b9b7-9904ce04ad33", 00:10:49.822 "strip_size_kb": 64, 00:10:49.822 "state": "online", 00:10:49.822 "raid_level": "concat", 00:10:49.822 "superblock": true, 00:10:49.822 "num_base_bdevs": 4, 00:10:49.822 "num_base_bdevs_discovered": 4, 00:10:49.822 "num_base_bdevs_operational": 4, 00:10:49.822 "base_bdevs_list": [ 00:10:49.822 { 00:10:49.822 "name": "pt1", 00:10:49.822 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:49.822 "is_configured": true, 00:10:49.822 "data_offset": 2048, 00:10:49.822 "data_size": 63488 00:10:49.822 }, 00:10:49.822 { 00:10:49.822 "name": "pt2", 00:10:49.822 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:49.822 "is_configured": true, 00:10:49.822 "data_offset": 2048, 00:10:49.822 "data_size": 63488 00:10:49.822 }, 00:10:49.822 { 00:10:49.822 "name": "pt3", 00:10:49.822 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:49.822 "is_configured": true, 00:10:49.822 "data_offset": 2048, 00:10:49.822 "data_size": 63488 00:10:49.822 }, 00:10:49.822 { 00:10:49.822 "name": "pt4", 00:10:49.822 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:49.822 "is_configured": true, 00:10:49.822 "data_offset": 2048, 00:10:49.822 "data_size": 63488 00:10:49.822 } 00:10:49.822 ] 00:10:49.822 } 00:10:49.822 } 00:10:49.822 }' 00:10:49.822 17:32:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:49.822 17:32:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:10:49.822 pt2 00:10:49.822 pt3 00:10:49.822 pt4' 00:10:49.822 17:32:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:49.822 17:32:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:49.822 17:32:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:49.822 17:32:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:10:49.822 17:32:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:49.822 17:32:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:49.822 17:32:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:49.822 17:32:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:49.822 17:32:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:49.822 17:32:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:49.822 17:32:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:49.822 17:32:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:10:49.822 17:32:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:49.822 17:32:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:49.822 17:32:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:49.822 17:32:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:49.822 17:32:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:49.822 17:32:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:49.822 17:32:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:49.822 17:32:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:10:49.822 17:32:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:49.822 17:32:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:49.823 17:32:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:49.823 17:32:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:49.823 17:32:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:49.823 17:32:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:49.823 17:32:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:49.823 17:32:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt4 00:10:49.823 17:32:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:49.823 17:32:20 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:49.823 17:32:20 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:50.083 17:32:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:50.083 17:32:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:50.083 17:32:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:50.083 17:32:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:10:50.083 17:32:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:10:50.083 17:32:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:50.083 17:32:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:50.083 [2024-11-27 17:32:21.052961] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:50.083 17:32:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:50.083 17:32:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=ed8af488-9e31-42ea-b9b7-9904ce04ad33 00:10:50.083 17:32:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z ed8af488-9e31-42ea-b9b7-9904ce04ad33 ']' 00:10:50.083 17:32:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:10:50.083 17:32:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:50.083 17:32:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:50.083 [2024-11-27 17:32:21.100625] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:10:50.083 [2024-11-27 17:32:21.100657] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:50.083 [2024-11-27 17:32:21.100744] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:50.083 [2024-11-27 17:32:21.100824] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:50.083 [2024-11-27 17:32:21.100846] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:10:50.083 17:32:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:50.083 17:32:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:50.083 17:32:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:50.083 17:32:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:50.083 17:32:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:10:50.083 17:32:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:50.083 17:32:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:10:50.083 17:32:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:10:50.083 17:32:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:10:50.084 17:32:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:10:50.084 17:32:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:50.084 17:32:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:50.084 17:32:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:50.084 17:32:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:10:50.084 17:32:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:10:50.084 17:32:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:50.084 17:32:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:50.084 17:32:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:50.084 17:32:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:10:50.084 17:32:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt3 00:10:50.084 17:32:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:50.084 17:32:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:50.084 17:32:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:50.084 17:32:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:10:50.084 17:32:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt4 00:10:50.084 17:32:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:50.084 17:32:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:50.084 17:32:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:50.084 17:32:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:10:50.084 17:32:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:50.084 17:32:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:50.084 17:32:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:10:50.084 17:32:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:50.084 17:32:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:10:50.084 17:32:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:10:50.084 17:32:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@650 -- # local es=0 00:10:50.084 17:32:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:10:50.084 17:32:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:10:50.084 17:32:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:10:50.084 17:32:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:10:50.084 17:32:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:10:50.084 17:32:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:10:50.084 17:32:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:50.084 17:32:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:50.084 [2024-11-27 17:32:21.256381] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:10:50.084 [2024-11-27 17:32:21.258523] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:10:50.084 [2024-11-27 17:32:21.258601] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc3 is claimed 00:10:50.084 [2024-11-27 17:32:21.258676] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc4 is claimed 00:10:50.084 [2024-11-27 17:32:21.258780] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:10:50.084 [2024-11-27 17:32:21.258866] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:10:50.084 [2024-11-27 17:32:21.258927] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc3 00:10:50.084 [2024-11-27 17:32:21.258980] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc4 00:10:50.084 [2024-11-27 17:32:21.259059] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:10:50.084 [2024-11-27 17:32:21.259102] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state configuring 00:10:50.084 request: 00:10:50.084 { 00:10:50.084 "name": "raid_bdev1", 00:10:50.084 "raid_level": "concat", 00:10:50.084 "base_bdevs": [ 00:10:50.084 "malloc1", 00:10:50.084 "malloc2", 00:10:50.084 "malloc3", 00:10:50.084 "malloc4" 00:10:50.084 ], 00:10:50.084 "strip_size_kb": 64, 00:10:50.084 "superblock": false, 00:10:50.084 "method": "bdev_raid_create", 00:10:50.084 "req_id": 1 00:10:50.084 } 00:10:50.084 Got JSON-RPC error response 00:10:50.084 response: 00:10:50.084 { 00:10:50.084 "code": -17, 00:10:50.084 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:10:50.084 } 00:10:50.084 17:32:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:10:50.084 17:32:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@653 -- # es=1 00:10:50.084 17:32:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:10:50.084 17:32:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:10:50.084 17:32:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:10:50.084 17:32:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:10:50.084 17:32:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:50.084 17:32:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:50.084 17:32:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:50.345 17:32:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:50.345 17:32:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:10:50.345 17:32:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:10:50.345 17:32:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:10:50.345 17:32:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:50.345 17:32:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:50.345 [2024-11-27 17:32:21.320237] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:10:50.345 [2024-11-27 17:32:21.320314] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:50.345 [2024-11-27 17:32:21.320371] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:10:50.345 [2024-11-27 17:32:21.320402] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:50.345 [2024-11-27 17:32:21.322885] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:50.345 [2024-11-27 17:32:21.322948] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:10:50.345 [2024-11-27 17:32:21.323041] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:10:50.345 [2024-11-27 17:32:21.323101] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:10:50.345 pt1 00:10:50.345 17:32:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:50.345 17:32:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring concat 64 4 00:10:50.345 17:32:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:50.345 17:32:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:50.345 17:32:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:50.345 17:32:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:50.345 17:32:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:50.345 17:32:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:50.345 17:32:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:50.345 17:32:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:50.345 17:32:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:50.345 17:32:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:50.345 17:32:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:50.345 17:32:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:50.345 17:32:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:50.345 17:32:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:50.345 17:32:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:50.345 "name": "raid_bdev1", 00:10:50.345 "uuid": "ed8af488-9e31-42ea-b9b7-9904ce04ad33", 00:10:50.345 "strip_size_kb": 64, 00:10:50.345 "state": "configuring", 00:10:50.345 "raid_level": "concat", 00:10:50.345 "superblock": true, 00:10:50.345 "num_base_bdevs": 4, 00:10:50.345 "num_base_bdevs_discovered": 1, 00:10:50.345 "num_base_bdevs_operational": 4, 00:10:50.345 "base_bdevs_list": [ 00:10:50.345 { 00:10:50.345 "name": "pt1", 00:10:50.345 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:50.345 "is_configured": true, 00:10:50.345 "data_offset": 2048, 00:10:50.345 "data_size": 63488 00:10:50.345 }, 00:10:50.345 { 00:10:50.345 "name": null, 00:10:50.345 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:50.345 "is_configured": false, 00:10:50.345 "data_offset": 2048, 00:10:50.345 "data_size": 63488 00:10:50.345 }, 00:10:50.345 { 00:10:50.345 "name": null, 00:10:50.345 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:50.345 "is_configured": false, 00:10:50.345 "data_offset": 2048, 00:10:50.345 "data_size": 63488 00:10:50.345 }, 00:10:50.345 { 00:10:50.345 "name": null, 00:10:50.345 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:50.345 "is_configured": false, 00:10:50.345 "data_offset": 2048, 00:10:50.345 "data_size": 63488 00:10:50.345 } 00:10:50.345 ] 00:10:50.345 }' 00:10:50.345 17:32:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:50.345 17:32:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:50.606 17:32:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 4 -gt 2 ']' 00:10:50.606 17:32:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@472 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:10:50.606 17:32:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:50.606 17:32:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:50.606 [2024-11-27 17:32:21.767496] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:10:50.606 [2024-11-27 17:32:21.767551] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:50.606 [2024-11-27 17:32:21.767573] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009c80 00:10:50.606 [2024-11-27 17:32:21.767584] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:50.606 [2024-11-27 17:32:21.768012] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:50.606 [2024-11-27 17:32:21.768029] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:10:50.606 [2024-11-27 17:32:21.768111] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:10:50.606 [2024-11-27 17:32:21.768144] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:10:50.606 pt2 00:10:50.606 17:32:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:50.606 17:32:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@473 -- # rpc_cmd bdev_passthru_delete pt2 00:10:50.606 17:32:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:50.606 17:32:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:50.606 [2024-11-27 17:32:21.779506] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: pt2 00:10:50.606 17:32:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:50.606 17:32:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@474 -- # verify_raid_bdev_state raid_bdev1 configuring concat 64 4 00:10:50.606 17:32:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:50.606 17:32:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:10:50.606 17:32:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:50.606 17:32:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:50.606 17:32:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:50.606 17:32:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:50.606 17:32:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:50.606 17:32:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:50.606 17:32:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:50.606 17:32:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:50.606 17:32:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:50.606 17:32:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:50.606 17:32:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:50.866 17:32:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:50.866 17:32:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:50.866 "name": "raid_bdev1", 00:10:50.866 "uuid": "ed8af488-9e31-42ea-b9b7-9904ce04ad33", 00:10:50.866 "strip_size_kb": 64, 00:10:50.866 "state": "configuring", 00:10:50.866 "raid_level": "concat", 00:10:50.866 "superblock": true, 00:10:50.866 "num_base_bdevs": 4, 00:10:50.866 "num_base_bdevs_discovered": 1, 00:10:50.866 "num_base_bdevs_operational": 4, 00:10:50.866 "base_bdevs_list": [ 00:10:50.866 { 00:10:50.866 "name": "pt1", 00:10:50.866 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:50.866 "is_configured": true, 00:10:50.866 "data_offset": 2048, 00:10:50.866 "data_size": 63488 00:10:50.866 }, 00:10:50.866 { 00:10:50.866 "name": null, 00:10:50.866 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:50.866 "is_configured": false, 00:10:50.866 "data_offset": 0, 00:10:50.866 "data_size": 63488 00:10:50.866 }, 00:10:50.866 { 00:10:50.866 "name": null, 00:10:50.866 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:50.866 "is_configured": false, 00:10:50.866 "data_offset": 2048, 00:10:50.866 "data_size": 63488 00:10:50.866 }, 00:10:50.866 { 00:10:50.866 "name": null, 00:10:50.866 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:50.866 "is_configured": false, 00:10:50.866 "data_offset": 2048, 00:10:50.866 "data_size": 63488 00:10:50.866 } 00:10:50.866 ] 00:10:50.866 }' 00:10:50.866 17:32:21 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:50.866 17:32:21 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:51.127 17:32:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:10:51.127 17:32:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:10:51.127 17:32:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:10:51.127 17:32:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:51.127 17:32:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:51.127 [2024-11-27 17:32:22.202775] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:10:51.127 [2024-11-27 17:32:22.202833] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:51.127 [2024-11-27 17:32:22.202851] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009f80 00:10:51.127 [2024-11-27 17:32:22.202862] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:51.127 [2024-11-27 17:32:22.203265] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:51.127 [2024-11-27 17:32:22.203296] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:10:51.127 [2024-11-27 17:32:22.203364] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:10:51.127 [2024-11-27 17:32:22.203390] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:10:51.127 pt2 00:10:51.127 17:32:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:51.127 17:32:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:10:51.127 17:32:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:10:51.127 17:32:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:10:51.127 17:32:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:51.127 17:32:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:51.127 [2024-11-27 17:32:22.214720] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:10:51.127 [2024-11-27 17:32:22.214784] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:51.127 [2024-11-27 17:32:22.214801] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:10:51.127 [2024-11-27 17:32:22.214821] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:51.127 [2024-11-27 17:32:22.215148] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:51.127 [2024-11-27 17:32:22.215195] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:10:51.127 [2024-11-27 17:32:22.215253] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:10:51.127 [2024-11-27 17:32:22.215276] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:10:51.127 pt3 00:10:51.127 17:32:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:51.127 17:32:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:10:51.127 17:32:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:10:51.127 17:32:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:10:51.127 17:32:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:51.127 17:32:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:51.127 [2024-11-27 17:32:22.226722] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:10:51.127 [2024-11-27 17:32:22.226766] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:51.127 [2024-11-27 17:32:22.226778] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a580 00:10:51.127 [2024-11-27 17:32:22.226788] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:51.127 [2024-11-27 17:32:22.227073] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:51.127 [2024-11-27 17:32:22.227091] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:10:51.127 [2024-11-27 17:32:22.227135] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt4 00:10:51.127 [2024-11-27 17:32:22.227190] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:10:51.127 [2024-11-27 17:32:22.227289] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:10:51.127 [2024-11-27 17:32:22.227300] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:10:51.127 [2024-11-27 17:32:22.227551] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:10:51.127 [2024-11-27 17:32:22.227672] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:10:51.127 [2024-11-27 17:32:22.227680] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:10:51.127 [2024-11-27 17:32:22.227774] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:51.127 pt4 00:10:51.127 17:32:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:51.127 17:32:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:10:51.127 17:32:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:10:51.127 17:32:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online concat 64 4 00:10:51.127 17:32:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:51.127 17:32:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:51.127 17:32:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:51.127 17:32:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:51.127 17:32:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:51.127 17:32:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:51.127 17:32:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:51.127 17:32:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:51.127 17:32:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:51.127 17:32:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:51.127 17:32:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:51.127 17:32:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:51.127 17:32:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:51.127 17:32:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:51.127 17:32:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:51.127 "name": "raid_bdev1", 00:10:51.127 "uuid": "ed8af488-9e31-42ea-b9b7-9904ce04ad33", 00:10:51.127 "strip_size_kb": 64, 00:10:51.127 "state": "online", 00:10:51.127 "raid_level": "concat", 00:10:51.127 "superblock": true, 00:10:51.127 "num_base_bdevs": 4, 00:10:51.127 "num_base_bdevs_discovered": 4, 00:10:51.127 "num_base_bdevs_operational": 4, 00:10:51.127 "base_bdevs_list": [ 00:10:51.127 { 00:10:51.127 "name": "pt1", 00:10:51.127 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:51.127 "is_configured": true, 00:10:51.127 "data_offset": 2048, 00:10:51.127 "data_size": 63488 00:10:51.127 }, 00:10:51.127 { 00:10:51.127 "name": "pt2", 00:10:51.127 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:51.127 "is_configured": true, 00:10:51.127 "data_offset": 2048, 00:10:51.127 "data_size": 63488 00:10:51.127 }, 00:10:51.127 { 00:10:51.127 "name": "pt3", 00:10:51.127 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:51.127 "is_configured": true, 00:10:51.127 "data_offset": 2048, 00:10:51.127 "data_size": 63488 00:10:51.127 }, 00:10:51.127 { 00:10:51.127 "name": "pt4", 00:10:51.127 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:51.127 "is_configured": true, 00:10:51.127 "data_offset": 2048, 00:10:51.127 "data_size": 63488 00:10:51.127 } 00:10:51.127 ] 00:10:51.127 }' 00:10:51.127 17:32:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:51.127 17:32:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:51.698 17:32:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:10:51.698 17:32:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:10:51.698 17:32:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:10:51.698 17:32:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:10:51.698 17:32:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:10:51.698 17:32:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:10:51.698 17:32:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:10:51.698 17:32:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:51.698 17:32:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:51.698 17:32:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:10:51.698 [2024-11-27 17:32:22.662314] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:51.698 17:32:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:51.698 17:32:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:10:51.698 "name": "raid_bdev1", 00:10:51.698 "aliases": [ 00:10:51.698 "ed8af488-9e31-42ea-b9b7-9904ce04ad33" 00:10:51.698 ], 00:10:51.698 "product_name": "Raid Volume", 00:10:51.698 "block_size": 512, 00:10:51.698 "num_blocks": 253952, 00:10:51.698 "uuid": "ed8af488-9e31-42ea-b9b7-9904ce04ad33", 00:10:51.698 "assigned_rate_limits": { 00:10:51.698 "rw_ios_per_sec": 0, 00:10:51.698 "rw_mbytes_per_sec": 0, 00:10:51.698 "r_mbytes_per_sec": 0, 00:10:51.698 "w_mbytes_per_sec": 0 00:10:51.698 }, 00:10:51.698 "claimed": false, 00:10:51.698 "zoned": false, 00:10:51.698 "supported_io_types": { 00:10:51.698 "read": true, 00:10:51.698 "write": true, 00:10:51.698 "unmap": true, 00:10:51.698 "flush": true, 00:10:51.698 "reset": true, 00:10:51.698 "nvme_admin": false, 00:10:51.698 "nvme_io": false, 00:10:51.698 "nvme_io_md": false, 00:10:51.698 "write_zeroes": true, 00:10:51.698 "zcopy": false, 00:10:51.698 "get_zone_info": false, 00:10:51.698 "zone_management": false, 00:10:51.698 "zone_append": false, 00:10:51.698 "compare": false, 00:10:51.698 "compare_and_write": false, 00:10:51.698 "abort": false, 00:10:51.698 "seek_hole": false, 00:10:51.698 "seek_data": false, 00:10:51.698 "copy": false, 00:10:51.698 "nvme_iov_md": false 00:10:51.698 }, 00:10:51.698 "memory_domains": [ 00:10:51.698 { 00:10:51.698 "dma_device_id": "system", 00:10:51.698 "dma_device_type": 1 00:10:51.698 }, 00:10:51.698 { 00:10:51.698 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:51.698 "dma_device_type": 2 00:10:51.698 }, 00:10:51.698 { 00:10:51.698 "dma_device_id": "system", 00:10:51.698 "dma_device_type": 1 00:10:51.698 }, 00:10:51.698 { 00:10:51.698 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:51.698 "dma_device_type": 2 00:10:51.698 }, 00:10:51.698 { 00:10:51.698 "dma_device_id": "system", 00:10:51.698 "dma_device_type": 1 00:10:51.698 }, 00:10:51.698 { 00:10:51.698 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:51.698 "dma_device_type": 2 00:10:51.698 }, 00:10:51.698 { 00:10:51.698 "dma_device_id": "system", 00:10:51.698 "dma_device_type": 1 00:10:51.698 }, 00:10:51.698 { 00:10:51.698 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:10:51.698 "dma_device_type": 2 00:10:51.698 } 00:10:51.698 ], 00:10:51.698 "driver_specific": { 00:10:51.698 "raid": { 00:10:51.698 "uuid": "ed8af488-9e31-42ea-b9b7-9904ce04ad33", 00:10:51.698 "strip_size_kb": 64, 00:10:51.698 "state": "online", 00:10:51.698 "raid_level": "concat", 00:10:51.698 "superblock": true, 00:10:51.698 "num_base_bdevs": 4, 00:10:51.698 "num_base_bdevs_discovered": 4, 00:10:51.698 "num_base_bdevs_operational": 4, 00:10:51.698 "base_bdevs_list": [ 00:10:51.698 { 00:10:51.698 "name": "pt1", 00:10:51.698 "uuid": "00000000-0000-0000-0000-000000000001", 00:10:51.698 "is_configured": true, 00:10:51.698 "data_offset": 2048, 00:10:51.698 "data_size": 63488 00:10:51.698 }, 00:10:51.698 { 00:10:51.698 "name": "pt2", 00:10:51.698 "uuid": "00000000-0000-0000-0000-000000000002", 00:10:51.698 "is_configured": true, 00:10:51.698 "data_offset": 2048, 00:10:51.698 "data_size": 63488 00:10:51.698 }, 00:10:51.698 { 00:10:51.698 "name": "pt3", 00:10:51.698 "uuid": "00000000-0000-0000-0000-000000000003", 00:10:51.698 "is_configured": true, 00:10:51.698 "data_offset": 2048, 00:10:51.698 "data_size": 63488 00:10:51.698 }, 00:10:51.698 { 00:10:51.698 "name": "pt4", 00:10:51.698 "uuid": "00000000-0000-0000-0000-000000000004", 00:10:51.698 "is_configured": true, 00:10:51.698 "data_offset": 2048, 00:10:51.698 "data_size": 63488 00:10:51.698 } 00:10:51.698 ] 00:10:51.698 } 00:10:51.698 } 00:10:51.698 }' 00:10:51.698 17:32:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:10:51.698 17:32:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:10:51.698 pt2 00:10:51.698 pt3 00:10:51.698 pt4' 00:10:51.698 17:32:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:51.698 17:32:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:10:51.698 17:32:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:51.698 17:32:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:10:51.698 17:32:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:51.698 17:32:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:51.698 17:32:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:51.698 17:32:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:51.698 17:32:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:51.698 17:32:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:51.698 17:32:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:51.698 17:32:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:51.698 17:32:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:10:51.698 17:32:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:51.698 17:32:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:51.698 17:32:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:51.959 17:32:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:51.959 17:32:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:51.959 17:32:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:51.959 17:32:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:10:51.959 17:32:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:51.959 17:32:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:51.959 17:32:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:51.959 17:32:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:51.959 17:32:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:51.959 17:32:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:51.959 17:32:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:10:51.959 17:32:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:10:51.959 17:32:22 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt4 00:10:51.959 17:32:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:51.959 17:32:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:51.959 17:32:22 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:51.959 17:32:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:10:51.959 17:32:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:10:51.959 17:32:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:10:51.959 17:32:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:10:51.959 17:32:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:51.959 17:32:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:51.959 [2024-11-27 17:32:23.013649] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:10:51.959 17:32:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:51.959 17:32:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' ed8af488-9e31-42ea-b9b7-9904ce04ad33 '!=' ed8af488-9e31-42ea-b9b7-9904ce04ad33 ']' 00:10:51.959 17:32:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy concat 00:10:51.959 17:32:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:10:51.959 17:32:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@200 -- # return 1 00:10:51.959 17:32:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 83327 00:10:51.959 17:32:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@950 -- # '[' -z 83327 ']' 00:10:51.959 17:32:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # kill -0 83327 00:10:51.959 17:32:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@955 -- # uname 00:10:51.959 17:32:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:10:51.959 17:32:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 83327 00:10:51.959 17:32:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:10:51.959 17:32:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:10:51.959 17:32:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 83327' 00:10:51.959 killing process with pid 83327 00:10:51.959 17:32:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@969 -- # kill 83327 00:10:51.959 [2024-11-27 17:32:23.084936] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:10:51.959 [2024-11-27 17:32:23.085081] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:51.959 [2024-11-27 17:32:23.085188] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:51.959 [2024-11-27 17:32:23.085236] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:10:51.959 17:32:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@974 -- # wait 83327 00:10:52.220 [2024-11-27 17:32:23.163329] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:10:52.480 17:32:23 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:10:52.480 00:10:52.480 real 0m4.341s 00:10:52.480 user 0m6.653s 00:10:52.480 sys 0m0.973s 00:10:52.480 17:32:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:10:52.480 ************************************ 00:10:52.480 END TEST raid_superblock_test 00:10:52.480 ************************************ 00:10:52.480 17:32:23 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:10:52.480 17:32:23 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test concat 4 read 00:10:52.480 17:32:23 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:10:52.480 17:32:23 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:10:52.480 17:32:23 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:10:52.480 ************************************ 00:10:52.480 START TEST raid_read_error_test 00:10:52.480 ************************************ 00:10:52.480 17:32:23 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1125 -- # raid_io_error_test concat 4 read 00:10:52.480 17:32:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=concat 00:10:52.480 17:32:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=4 00:10:52.480 17:32:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:10:52.480 17:32:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:10:52.480 17:32:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:52.480 17:32:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:10:52.480 17:32:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:52.480 17:32:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:52.480 17:32:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:10:52.480 17:32:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:52.480 17:32:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:52.480 17:32:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:10:52.480 17:32:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:52.480 17:32:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:52.480 17:32:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev4 00:10:52.480 17:32:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:52.480 17:32:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:52.480 17:32:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:10:52.480 17:32:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:10:52.480 17:32:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:10:52.480 17:32:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:10:52.480 17:32:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:10:52.480 17:32:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:10:52.480 17:32:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:10:52.480 17:32:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' concat '!=' raid1 ']' 00:10:52.480 17:32:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:10:52.480 17:32:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:10:52.480 17:32:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:10:52.480 17:32:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.FSdr7MWvOx 00:10:52.480 17:32:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=83582 00:10:52.480 17:32:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:10:52.480 17:32:23 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 83582 00:10:52.480 17:32:23 bdev_raid.raid_read_error_test -- common/autotest_common.sh@831 -- # '[' -z 83582 ']' 00:10:52.480 17:32:23 bdev_raid.raid_read_error_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:52.480 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:52.480 17:32:23 bdev_raid.raid_read_error_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:10:52.480 17:32:23 bdev_raid.raid_read_error_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:52.480 17:32:23 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:10:52.480 17:32:23 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:52.741 [2024-11-27 17:32:23.712975] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:10:52.741 [2024-11-27 17:32:23.713104] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid83582 ] 00:10:52.741 [2024-11-27 17:32:23.858946] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:52.741 [2024-11-27 17:32:23.926427] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:10:53.001 [2024-11-27 17:32:24.002175] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:53.001 [2024-11-27 17:32:24.002212] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:53.572 17:32:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:10:53.572 17:32:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@864 -- # return 0 00:10:53.572 17:32:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:53.572 17:32:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:10:53.572 17:32:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:53.572 17:32:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:53.572 BaseBdev1_malloc 00:10:53.572 17:32:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:53.572 17:32:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:10:53.572 17:32:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:53.572 17:32:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:53.572 true 00:10:53.572 17:32:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:53.572 17:32:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:10:53.572 17:32:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:53.572 17:32:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:53.572 [2024-11-27 17:32:24.568112] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:10:53.572 [2024-11-27 17:32:24.568266] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:53.572 [2024-11-27 17:32:24.568295] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:10:53.572 [2024-11-27 17:32:24.568305] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:53.572 [2024-11-27 17:32:24.570861] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:53.572 [2024-11-27 17:32:24.570907] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:10:53.572 BaseBdev1 00:10:53.572 17:32:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:53.572 17:32:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:53.572 17:32:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:10:53.572 17:32:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:53.572 17:32:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:53.572 BaseBdev2_malloc 00:10:53.572 17:32:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:53.572 17:32:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:10:53.572 17:32:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:53.572 17:32:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:53.572 true 00:10:53.572 17:32:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:53.572 17:32:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:10:53.572 17:32:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:53.572 17:32:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:53.572 [2024-11-27 17:32:24.629375] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:10:53.572 [2024-11-27 17:32:24.629513] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:53.572 [2024-11-27 17:32:24.629552] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:10:53.572 [2024-11-27 17:32:24.629566] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:53.572 [2024-11-27 17:32:24.632599] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:53.572 [2024-11-27 17:32:24.632670] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:10:53.572 BaseBdev2 00:10:53.572 17:32:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:53.573 17:32:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:53.573 17:32:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:10:53.573 17:32:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:53.573 17:32:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:53.573 BaseBdev3_malloc 00:10:53.573 17:32:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:53.573 17:32:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:10:53.573 17:32:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:53.573 17:32:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:53.573 true 00:10:53.573 17:32:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:53.573 17:32:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:10:53.573 17:32:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:53.573 17:32:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:53.573 [2024-11-27 17:32:24.675951] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:10:53.573 [2024-11-27 17:32:24.675999] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:53.573 [2024-11-27 17:32:24.676019] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008780 00:10:53.573 [2024-11-27 17:32:24.676028] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:53.573 [2024-11-27 17:32:24.678428] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:53.573 [2024-11-27 17:32:24.678496] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:10:53.573 BaseBdev3 00:10:53.573 17:32:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:53.573 17:32:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:53.573 17:32:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:10:53.573 17:32:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:53.573 17:32:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:53.573 BaseBdev4_malloc 00:10:53.573 17:32:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:53.573 17:32:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev4_malloc 00:10:53.573 17:32:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:53.573 17:32:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:53.573 true 00:10:53.573 17:32:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:53.573 17:32:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev4_malloc -p BaseBdev4 00:10:53.573 17:32:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:53.573 17:32:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:53.573 [2024-11-27 17:32:24.722316] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev4_malloc 00:10:53.573 [2024-11-27 17:32:24.722359] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:53.573 [2024-11-27 17:32:24.722382] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:10:53.573 [2024-11-27 17:32:24.722390] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:53.573 [2024-11-27 17:32:24.724690] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:53.573 [2024-11-27 17:32:24.724725] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:10:53.573 BaseBdev4 00:10:53.573 17:32:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:53.573 17:32:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 -s 00:10:53.573 17:32:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:53.573 17:32:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:53.573 [2024-11-27 17:32:24.734376] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:53.573 [2024-11-27 17:32:24.736443] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:53.573 [2024-11-27 17:32:24.736565] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:53.573 [2024-11-27 17:32:24.736648] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:10:53.573 [2024-11-27 17:32:24.736847] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000002000 00:10:53.573 [2024-11-27 17:32:24.736868] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:10:53.573 [2024-11-27 17:32:24.737128] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:10:53.573 [2024-11-27 17:32:24.737283] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000002000 00:10:53.573 [2024-11-27 17:32:24.737302] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000002000 00:10:53.573 [2024-11-27 17:32:24.737436] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:53.573 17:32:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:53.573 17:32:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online concat 64 4 00:10:53.573 17:32:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:53.573 17:32:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:53.573 17:32:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:53.573 17:32:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:53.573 17:32:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:53.573 17:32:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:53.573 17:32:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:53.573 17:32:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:53.573 17:32:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:53.573 17:32:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:53.573 17:32:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:53.573 17:32:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:53.573 17:32:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:53.833 17:32:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:53.833 17:32:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:53.833 "name": "raid_bdev1", 00:10:53.833 "uuid": "cb7902ba-6806-4ada-948c-a8ec6ed23fab", 00:10:53.833 "strip_size_kb": 64, 00:10:53.833 "state": "online", 00:10:53.833 "raid_level": "concat", 00:10:53.833 "superblock": true, 00:10:53.833 "num_base_bdevs": 4, 00:10:53.833 "num_base_bdevs_discovered": 4, 00:10:53.833 "num_base_bdevs_operational": 4, 00:10:53.833 "base_bdevs_list": [ 00:10:53.833 { 00:10:53.833 "name": "BaseBdev1", 00:10:53.833 "uuid": "ec7d70c2-5e82-5c36-a21d-8e988e4b4c89", 00:10:53.833 "is_configured": true, 00:10:53.833 "data_offset": 2048, 00:10:53.833 "data_size": 63488 00:10:53.833 }, 00:10:53.833 { 00:10:53.833 "name": "BaseBdev2", 00:10:53.833 "uuid": "a2c1505b-8b27-5f86-a1ae-2a76381bca93", 00:10:53.833 "is_configured": true, 00:10:53.833 "data_offset": 2048, 00:10:53.833 "data_size": 63488 00:10:53.833 }, 00:10:53.833 { 00:10:53.833 "name": "BaseBdev3", 00:10:53.833 "uuid": "14caf647-2145-5c92-9c6b-ac9140cfb803", 00:10:53.833 "is_configured": true, 00:10:53.833 "data_offset": 2048, 00:10:53.833 "data_size": 63488 00:10:53.833 }, 00:10:53.833 { 00:10:53.833 "name": "BaseBdev4", 00:10:53.833 "uuid": "a7ba05e7-ffa1-5b3a-a786-919d0f1d77e7", 00:10:53.833 "is_configured": true, 00:10:53.833 "data_offset": 2048, 00:10:53.833 "data_size": 63488 00:10:53.833 } 00:10:53.833 ] 00:10:53.833 }' 00:10:53.833 17:32:24 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:53.833 17:32:24 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:54.094 17:32:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:10:54.094 17:32:25 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:10:54.094 [2024-11-27 17:32:25.205982] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000026d0 00:10:55.052 17:32:26 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:10:55.052 17:32:26 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:55.052 17:32:26 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:55.052 17:32:26 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:55.052 17:32:26 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:10:55.052 17:32:26 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ concat = \r\a\i\d\1 ]] 00:10:55.052 17:32:26 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=4 00:10:55.052 17:32:26 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online concat 64 4 00:10:55.052 17:32:26 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:55.052 17:32:26 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:55.052 17:32:26 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:55.052 17:32:26 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:55.052 17:32:26 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:55.052 17:32:26 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:55.052 17:32:26 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:55.052 17:32:26 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:55.053 17:32:26 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:55.053 17:32:26 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:55.053 17:32:26 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:55.053 17:32:26 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:55.053 17:32:26 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:55.053 17:32:26 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:55.053 17:32:26 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:55.053 "name": "raid_bdev1", 00:10:55.053 "uuid": "cb7902ba-6806-4ada-948c-a8ec6ed23fab", 00:10:55.053 "strip_size_kb": 64, 00:10:55.053 "state": "online", 00:10:55.053 "raid_level": "concat", 00:10:55.053 "superblock": true, 00:10:55.053 "num_base_bdevs": 4, 00:10:55.053 "num_base_bdevs_discovered": 4, 00:10:55.053 "num_base_bdevs_operational": 4, 00:10:55.053 "base_bdevs_list": [ 00:10:55.053 { 00:10:55.053 "name": "BaseBdev1", 00:10:55.053 "uuid": "ec7d70c2-5e82-5c36-a21d-8e988e4b4c89", 00:10:55.053 "is_configured": true, 00:10:55.053 "data_offset": 2048, 00:10:55.053 "data_size": 63488 00:10:55.053 }, 00:10:55.053 { 00:10:55.053 "name": "BaseBdev2", 00:10:55.053 "uuid": "a2c1505b-8b27-5f86-a1ae-2a76381bca93", 00:10:55.053 "is_configured": true, 00:10:55.053 "data_offset": 2048, 00:10:55.053 "data_size": 63488 00:10:55.053 }, 00:10:55.053 { 00:10:55.053 "name": "BaseBdev3", 00:10:55.053 "uuid": "14caf647-2145-5c92-9c6b-ac9140cfb803", 00:10:55.053 "is_configured": true, 00:10:55.053 "data_offset": 2048, 00:10:55.053 "data_size": 63488 00:10:55.053 }, 00:10:55.053 { 00:10:55.053 "name": "BaseBdev4", 00:10:55.053 "uuid": "a7ba05e7-ffa1-5b3a-a786-919d0f1d77e7", 00:10:55.053 "is_configured": true, 00:10:55.053 "data_offset": 2048, 00:10:55.053 "data_size": 63488 00:10:55.053 } 00:10:55.053 ] 00:10:55.053 }' 00:10:55.053 17:32:26 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:55.053 17:32:26 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:55.623 17:32:26 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:10:55.623 17:32:26 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:55.623 17:32:26 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:55.623 [2024-11-27 17:32:26.598583] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:10:55.623 [2024-11-27 17:32:26.598685] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:55.623 [2024-11-27 17:32:26.601364] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:55.623 [2024-11-27 17:32:26.601424] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:55.623 [2024-11-27 17:32:26.601479] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:55.623 [2024-11-27 17:32:26.601489] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002000 name raid_bdev1, state offline 00:10:55.623 { 00:10:55.623 "results": [ 00:10:55.623 { 00:10:55.623 "job": "raid_bdev1", 00:10:55.623 "core_mask": "0x1", 00:10:55.623 "workload": "randrw", 00:10:55.623 "percentage": 50, 00:10:55.623 "status": "finished", 00:10:55.623 "queue_depth": 1, 00:10:55.623 "io_size": 131072, 00:10:55.623 "runtime": 1.393263, 00:10:55.623 "iops": 14695.717893893687, 00:10:55.623 "mibps": 1836.9647367367108, 00:10:55.623 "io_failed": 1, 00:10:55.623 "io_timeout": 0, 00:10:55.623 "avg_latency_us": 95.5739069533743, 00:10:55.623 "min_latency_us": 24.817467248908297, 00:10:55.623 "max_latency_us": 1423.7624454148472 00:10:55.623 } 00:10:55.623 ], 00:10:55.623 "core_count": 1 00:10:55.623 } 00:10:55.623 17:32:26 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:55.623 17:32:26 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 83582 00:10:55.623 17:32:26 bdev_raid.raid_read_error_test -- common/autotest_common.sh@950 -- # '[' -z 83582 ']' 00:10:55.623 17:32:26 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # kill -0 83582 00:10:55.623 17:32:26 bdev_raid.raid_read_error_test -- common/autotest_common.sh@955 -- # uname 00:10:55.623 17:32:26 bdev_raid.raid_read_error_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:10:55.623 17:32:26 bdev_raid.raid_read_error_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 83582 00:10:55.623 17:32:26 bdev_raid.raid_read_error_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:10:55.623 17:32:26 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:10:55.623 17:32:26 bdev_raid.raid_read_error_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 83582' 00:10:55.623 killing process with pid 83582 00:10:55.623 17:32:26 bdev_raid.raid_read_error_test -- common/autotest_common.sh@969 -- # kill 83582 00:10:55.623 [2024-11-27 17:32:26.648902] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:10:55.623 17:32:26 bdev_raid.raid_read_error_test -- common/autotest_common.sh@974 -- # wait 83582 00:10:55.623 [2024-11-27 17:32:26.713148] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:10:56.194 17:32:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.FSdr7MWvOx 00:10:56.194 17:32:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:10:56.194 17:32:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:10:56.194 17:32:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.72 00:10:56.194 17:32:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy concat 00:10:56.194 17:32:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:10:56.194 17:32:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:10:56.194 17:32:27 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.72 != \0\.\0\0 ]] 00:10:56.194 00:10:56.194 real 0m3.491s 00:10:56.194 user 0m4.209s 00:10:56.194 sys 0m0.651s 00:10:56.194 17:32:27 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:10:56.194 ************************************ 00:10:56.194 END TEST raid_read_error_test 00:10:56.194 ************************************ 00:10:56.194 17:32:27 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:56.194 17:32:27 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test concat 4 write 00:10:56.194 17:32:27 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:10:56.194 17:32:27 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:10:56.194 17:32:27 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:10:56.194 ************************************ 00:10:56.194 START TEST raid_write_error_test 00:10:56.194 ************************************ 00:10:56.194 17:32:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1125 -- # raid_io_error_test concat 4 write 00:10:56.194 17:32:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=concat 00:10:56.194 17:32:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=4 00:10:56.194 17:32:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:10:56.194 17:32:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:10:56.194 17:32:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:56.194 17:32:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:10:56.194 17:32:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:56.194 17:32:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:56.194 17:32:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:10:56.194 17:32:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:56.194 17:32:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:56.194 17:32:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:10:56.194 17:32:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:56.194 17:32:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:56.194 17:32:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev4 00:10:56.194 17:32:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:10:56.194 17:32:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:10:56.194 17:32:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:10:56.194 17:32:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:10:56.194 17:32:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:10:56.194 17:32:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:10:56.194 17:32:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:10:56.194 17:32:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:10:56.194 17:32:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:10:56.194 17:32:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' concat '!=' raid1 ']' 00:10:56.195 17:32:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@801 -- # strip_size=64 00:10:56.195 17:32:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@802 -- # create_arg+=' -z 64' 00:10:56.195 17:32:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:10:56.195 17:32:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.fmKy5PWRJS 00:10:56.195 17:32:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=83711 00:10:56.195 17:32:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:10:56.195 17:32:27 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 83711 00:10:56.195 17:32:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@831 -- # '[' -z 83711 ']' 00:10:56.195 17:32:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:56.195 17:32:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:10:56.195 17:32:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:56.195 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:56.195 17:32:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:10:56.195 17:32:27 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:56.195 [2024-11-27 17:32:27.277795] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:10:56.195 [2024-11-27 17:32:27.277939] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid83711 ] 00:10:56.455 [2024-11-27 17:32:27.424976] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:56.455 [2024-11-27 17:32:27.492766] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:10:56.455 [2024-11-27 17:32:27.568718] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:56.455 [2024-11-27 17:32:27.568761] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:10:57.025 17:32:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:10:57.025 17:32:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@864 -- # return 0 00:10:57.025 17:32:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:57.025 17:32:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:10:57.025 17:32:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:57.025 17:32:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:57.025 BaseBdev1_malloc 00:10:57.025 17:32:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:57.025 17:32:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:10:57.025 17:32:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:57.025 17:32:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:57.025 true 00:10:57.025 17:32:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:57.025 17:32:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:10:57.025 17:32:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:57.025 17:32:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:57.025 [2024-11-27 17:32:28.135243] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:10:57.025 [2024-11-27 17:32:28.135307] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:57.025 [2024-11-27 17:32:28.135354] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:10:57.025 [2024-11-27 17:32:28.135370] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:57.025 [2024-11-27 17:32:28.137841] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:57.025 [2024-11-27 17:32:28.137876] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:10:57.025 BaseBdev1 00:10:57.025 17:32:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:57.025 17:32:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:57.025 17:32:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:10:57.025 17:32:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:57.025 17:32:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:57.025 BaseBdev2_malloc 00:10:57.025 17:32:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:57.025 17:32:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:10:57.026 17:32:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:57.026 17:32:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:57.026 true 00:10:57.026 17:32:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:57.026 17:32:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:10:57.026 17:32:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:57.026 17:32:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:57.026 [2024-11-27 17:32:28.191380] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:10:57.026 [2024-11-27 17:32:28.191437] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:57.026 [2024-11-27 17:32:28.191475] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:10:57.026 [2024-11-27 17:32:28.191484] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:57.026 [2024-11-27 17:32:28.193877] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:57.026 [2024-11-27 17:32:28.193912] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:10:57.026 BaseBdev2 00:10:57.026 17:32:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:57.026 17:32:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:57.026 17:32:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:10:57.026 17:32:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:57.026 17:32:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:57.286 BaseBdev3_malloc 00:10:57.286 17:32:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:57.286 17:32:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:10:57.286 17:32:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:57.286 17:32:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:57.286 true 00:10:57.286 17:32:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:57.286 17:32:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:10:57.286 17:32:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:57.286 17:32:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:57.286 [2024-11-27 17:32:28.237990] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:10:57.286 [2024-11-27 17:32:28.238039] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:57.286 [2024-11-27 17:32:28.238058] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008780 00:10:57.286 [2024-11-27 17:32:28.238067] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:57.286 [2024-11-27 17:32:28.240463] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:57.286 [2024-11-27 17:32:28.240495] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:10:57.286 BaseBdev3 00:10:57.286 17:32:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:57.286 17:32:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:10:57.286 17:32:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:10:57.286 17:32:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:57.286 17:32:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:57.286 BaseBdev4_malloc 00:10:57.286 17:32:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:57.286 17:32:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev4_malloc 00:10:57.286 17:32:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:57.286 17:32:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:57.286 true 00:10:57.286 17:32:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:57.286 17:32:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev4_malloc -p BaseBdev4 00:10:57.286 17:32:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:57.286 17:32:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:57.286 [2024-11-27 17:32:28.284573] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev4_malloc 00:10:57.286 [2024-11-27 17:32:28.284617] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:10:57.286 [2024-11-27 17:32:28.284654] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:10:57.286 [2024-11-27 17:32:28.284663] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:10:57.286 [2024-11-27 17:32:28.287011] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:10:57.286 [2024-11-27 17:32:28.287086] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:10:57.286 BaseBdev4 00:10:57.286 17:32:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:57.286 17:32:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -z 64 -r concat -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 -s 00:10:57.286 17:32:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:57.286 17:32:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:57.286 [2024-11-27 17:32:28.296624] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:10:57.286 [2024-11-27 17:32:28.298740] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:10:57.286 [2024-11-27 17:32:28.298815] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:10:57.286 [2024-11-27 17:32:28.298876] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:10:57.286 [2024-11-27 17:32:28.299070] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000002000 00:10:57.286 [2024-11-27 17:32:28.299082] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 253952, blocklen 512 00:10:57.286 [2024-11-27 17:32:28.299333] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:10:57.286 [2024-11-27 17:32:28.299466] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000002000 00:10:57.286 [2024-11-27 17:32:28.299481] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000002000 00:10:57.286 [2024-11-27 17:32:28.299610] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:57.286 17:32:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:57.286 17:32:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online concat 64 4 00:10:57.286 17:32:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:57.286 17:32:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:57.286 17:32:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:57.286 17:32:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:57.287 17:32:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:57.287 17:32:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:57.287 17:32:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:57.287 17:32:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:57.287 17:32:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:57.287 17:32:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:57.287 17:32:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:57.287 17:32:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:57.287 17:32:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:57.287 17:32:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:57.287 17:32:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:57.287 "name": "raid_bdev1", 00:10:57.287 "uuid": "8acec929-555b-44fc-b076-b3c85c7a68d2", 00:10:57.287 "strip_size_kb": 64, 00:10:57.287 "state": "online", 00:10:57.287 "raid_level": "concat", 00:10:57.287 "superblock": true, 00:10:57.287 "num_base_bdevs": 4, 00:10:57.287 "num_base_bdevs_discovered": 4, 00:10:57.287 "num_base_bdevs_operational": 4, 00:10:57.287 "base_bdevs_list": [ 00:10:57.287 { 00:10:57.287 "name": "BaseBdev1", 00:10:57.287 "uuid": "361643bd-6b22-5c61-ada6-fa03db9c0e00", 00:10:57.287 "is_configured": true, 00:10:57.287 "data_offset": 2048, 00:10:57.287 "data_size": 63488 00:10:57.287 }, 00:10:57.287 { 00:10:57.287 "name": "BaseBdev2", 00:10:57.287 "uuid": "783ec13b-3824-5659-bc70-e1714266f7d8", 00:10:57.287 "is_configured": true, 00:10:57.287 "data_offset": 2048, 00:10:57.287 "data_size": 63488 00:10:57.287 }, 00:10:57.287 { 00:10:57.287 "name": "BaseBdev3", 00:10:57.287 "uuid": "edcb631d-ed8d-59c1-9309-3071624da41c", 00:10:57.287 "is_configured": true, 00:10:57.287 "data_offset": 2048, 00:10:57.287 "data_size": 63488 00:10:57.287 }, 00:10:57.287 { 00:10:57.287 "name": "BaseBdev4", 00:10:57.287 "uuid": "62c7a146-49e4-57cf-b6cf-f485eabd1b5a", 00:10:57.287 "is_configured": true, 00:10:57.287 "data_offset": 2048, 00:10:57.287 "data_size": 63488 00:10:57.287 } 00:10:57.287 ] 00:10:57.287 }' 00:10:57.287 17:32:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:57.287 17:32:28 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:57.547 17:32:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:10:57.547 17:32:28 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:10:57.807 [2024-11-27 17:32:28.824209] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000026d0 00:10:58.748 17:32:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:10:58.748 17:32:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:58.748 17:32:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:58.748 17:32:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:58.748 17:32:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:10:58.748 17:32:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ concat = \r\a\i\d\1 ]] 00:10:58.748 17:32:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=4 00:10:58.748 17:32:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online concat 64 4 00:10:58.748 17:32:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:10:58.748 17:32:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:10:58.748 17:32:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=concat 00:10:58.748 17:32:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:10:58.748 17:32:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:10:58.748 17:32:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:10:58.748 17:32:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:10:58.748 17:32:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:10:58.748 17:32:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:10:58.748 17:32:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:10:58.748 17:32:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:10:58.748 17:32:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:58.748 17:32:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:58.748 17:32:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:58.748 17:32:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:10:58.748 "name": "raid_bdev1", 00:10:58.748 "uuid": "8acec929-555b-44fc-b076-b3c85c7a68d2", 00:10:58.748 "strip_size_kb": 64, 00:10:58.748 "state": "online", 00:10:58.748 "raid_level": "concat", 00:10:58.748 "superblock": true, 00:10:58.748 "num_base_bdevs": 4, 00:10:58.748 "num_base_bdevs_discovered": 4, 00:10:58.748 "num_base_bdevs_operational": 4, 00:10:58.748 "base_bdevs_list": [ 00:10:58.748 { 00:10:58.748 "name": "BaseBdev1", 00:10:58.748 "uuid": "361643bd-6b22-5c61-ada6-fa03db9c0e00", 00:10:58.748 "is_configured": true, 00:10:58.748 "data_offset": 2048, 00:10:58.748 "data_size": 63488 00:10:58.748 }, 00:10:58.748 { 00:10:58.748 "name": "BaseBdev2", 00:10:58.748 "uuid": "783ec13b-3824-5659-bc70-e1714266f7d8", 00:10:58.748 "is_configured": true, 00:10:58.748 "data_offset": 2048, 00:10:58.748 "data_size": 63488 00:10:58.748 }, 00:10:58.748 { 00:10:58.748 "name": "BaseBdev3", 00:10:58.748 "uuid": "edcb631d-ed8d-59c1-9309-3071624da41c", 00:10:58.748 "is_configured": true, 00:10:58.748 "data_offset": 2048, 00:10:58.748 "data_size": 63488 00:10:58.748 }, 00:10:58.748 { 00:10:58.748 "name": "BaseBdev4", 00:10:58.748 "uuid": "62c7a146-49e4-57cf-b6cf-f485eabd1b5a", 00:10:58.748 "is_configured": true, 00:10:58.748 "data_offset": 2048, 00:10:58.748 "data_size": 63488 00:10:58.748 } 00:10:58.748 ] 00:10:58.748 }' 00:10:58.748 17:32:29 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:10:58.748 17:32:29 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:59.319 17:32:30 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:10:59.319 17:32:30 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:59.319 17:32:30 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:59.319 [2024-11-27 17:32:30.244832] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:10:59.319 [2024-11-27 17:32:30.244926] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:10:59.319 [2024-11-27 17:32:30.247476] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:10:59.319 [2024-11-27 17:32:30.247583] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:10:59.319 [2024-11-27 17:32:30.247667] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:10:59.319 [2024-11-27 17:32:30.247709] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002000 name raid_bdev1, state offline 00:10:59.319 17:32:30 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:59.319 { 00:10:59.319 "results": [ 00:10:59.319 { 00:10:59.319 "job": "raid_bdev1", 00:10:59.319 "core_mask": "0x1", 00:10:59.319 "workload": "randrw", 00:10:59.319 "percentage": 50, 00:10:59.319 "status": "finished", 00:10:59.319 "queue_depth": 1, 00:10:59.319 "io_size": 131072, 00:10:59.319 "runtime": 1.421225, 00:10:59.319 "iops": 14776.68912381924, 00:10:59.319 "mibps": 1847.086140477405, 00:10:59.319 "io_failed": 1, 00:10:59.319 "io_timeout": 0, 00:10:59.319 "avg_latency_us": 95.11341627268602, 00:10:59.319 "min_latency_us": 24.705676855895195, 00:10:59.319 "max_latency_us": 1387.989519650655 00:10:59.319 } 00:10:59.319 ], 00:10:59.319 "core_count": 1 00:10:59.319 } 00:10:59.319 17:32:30 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 83711 00:10:59.319 17:32:30 bdev_raid.raid_write_error_test -- common/autotest_common.sh@950 -- # '[' -z 83711 ']' 00:10:59.319 17:32:30 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # kill -0 83711 00:10:59.319 17:32:30 bdev_raid.raid_write_error_test -- common/autotest_common.sh@955 -- # uname 00:10:59.319 17:32:30 bdev_raid.raid_write_error_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:10:59.319 17:32:30 bdev_raid.raid_write_error_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 83711 00:10:59.319 17:32:30 bdev_raid.raid_write_error_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:10:59.319 killing process with pid 83711 00:10:59.319 17:32:30 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:10:59.319 17:32:30 bdev_raid.raid_write_error_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 83711' 00:10:59.319 17:32:30 bdev_raid.raid_write_error_test -- common/autotest_common.sh@969 -- # kill 83711 00:10:59.319 [2024-11-27 17:32:30.285928] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:10:59.319 17:32:30 bdev_raid.raid_write_error_test -- common/autotest_common.sh@974 -- # wait 83711 00:10:59.319 [2024-11-27 17:32:30.350323] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:10:59.579 17:32:30 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.fmKy5PWRJS 00:10:59.579 17:32:30 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:10:59.579 17:32:30 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:10:59.579 17:32:30 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.70 00:10:59.579 17:32:30 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy concat 00:10:59.579 ************************************ 00:10:59.579 END TEST raid_write_error_test 00:10:59.579 ************************************ 00:10:59.579 17:32:30 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:10:59.579 17:32:30 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@200 -- # return 1 00:10:59.579 17:32:30 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@849 -- # [[ 0.70 != \0\.\0\0 ]] 00:10:59.579 00:10:59.579 real 0m3.553s 00:10:59.579 user 0m4.327s 00:10:59.579 sys 0m0.660s 00:10:59.579 17:32:30 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:10:59.579 17:32:30 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:10:59.840 17:32:30 bdev_raid -- bdev/bdev_raid.sh@967 -- # for level in raid0 concat raid1 00:10:59.840 17:32:30 bdev_raid -- bdev/bdev_raid.sh@968 -- # run_test raid_state_function_test raid_state_function_test raid1 4 false 00:10:59.840 17:32:30 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:10:59.840 17:32:30 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:10:59.840 17:32:30 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:10:59.840 ************************************ 00:10:59.840 START TEST raid_state_function_test 00:10:59.840 ************************************ 00:10:59.840 17:32:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1125 -- # raid_state_function_test raid1 4 false 00:10:59.840 17:32:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:10:59.840 17:32:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=4 00:10:59.840 17:32:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:10:59.840 17:32:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:10:59.840 17:32:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:10:59.840 17:32:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:59.840 17:32:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:10:59.840 17:32:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:59.840 17:32:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:59.840 17:32:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:10:59.840 17:32:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:59.840 17:32:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:59.840 17:32:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:10:59.840 17:32:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:59.840 17:32:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:59.840 17:32:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev4 00:10:59.840 17:32:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:10:59.840 17:32:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:10:59.840 17:32:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:10:59.840 17:32:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:10:59.840 17:32:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:10:59.840 17:32:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:10:59.840 17:32:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:10:59.840 17:32:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:10:59.840 17:32:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:10:59.840 17:32:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:10:59.840 17:32:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:10:59.841 17:32:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:10:59.841 17:32:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=83844 00:10:59.841 17:32:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:10:59.841 17:32:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 83844' 00:10:59.841 Process raid pid: 83844 00:10:59.841 17:32:30 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 83844 00:10:59.841 17:32:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@831 -- # '[' -z 83844 ']' 00:10:59.841 17:32:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:59.841 17:32:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:10:59.841 17:32:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:59.841 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:59.841 17:32:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:10:59.841 17:32:30 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:10:59.841 [2024-11-27 17:32:30.897697] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:10:59.841 [2024-11-27 17:32:30.897914] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:11:00.101 [2024-11-27 17:32:31.044237] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:11:00.101 [2024-11-27 17:32:31.113846] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:11:00.101 [2024-11-27 17:32:31.190945] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:00.101 [2024-11-27 17:32:31.191062] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:00.671 17:32:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:11:00.671 17:32:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@864 -- # return 0 00:11:00.671 17:32:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:11:00.671 17:32:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:00.671 17:32:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:00.671 [2024-11-27 17:32:31.726471] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:11:00.671 [2024-11-27 17:32:31.726590] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:11:00.671 [2024-11-27 17:32:31.726609] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:11:00.671 [2024-11-27 17:32:31.726619] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:11:00.671 [2024-11-27 17:32:31.726641] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:11:00.671 [2024-11-27 17:32:31.726655] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:11:00.671 [2024-11-27 17:32:31.726661] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:11:00.671 [2024-11-27 17:32:31.726676] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:11:00.671 17:32:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:00.671 17:32:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:11:00.671 17:32:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:00.671 17:32:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:00.671 17:32:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:00.671 17:32:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:00.671 17:32:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:00.671 17:32:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:00.671 17:32:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:00.671 17:32:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:00.671 17:32:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:00.671 17:32:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:00.671 17:32:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:00.671 17:32:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:00.671 17:32:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:00.671 17:32:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:00.671 17:32:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:00.671 "name": "Existed_Raid", 00:11:00.671 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:00.671 "strip_size_kb": 0, 00:11:00.671 "state": "configuring", 00:11:00.671 "raid_level": "raid1", 00:11:00.671 "superblock": false, 00:11:00.671 "num_base_bdevs": 4, 00:11:00.671 "num_base_bdevs_discovered": 0, 00:11:00.671 "num_base_bdevs_operational": 4, 00:11:00.671 "base_bdevs_list": [ 00:11:00.671 { 00:11:00.671 "name": "BaseBdev1", 00:11:00.671 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:00.671 "is_configured": false, 00:11:00.672 "data_offset": 0, 00:11:00.672 "data_size": 0 00:11:00.672 }, 00:11:00.672 { 00:11:00.672 "name": "BaseBdev2", 00:11:00.672 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:00.672 "is_configured": false, 00:11:00.672 "data_offset": 0, 00:11:00.672 "data_size": 0 00:11:00.672 }, 00:11:00.672 { 00:11:00.672 "name": "BaseBdev3", 00:11:00.672 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:00.672 "is_configured": false, 00:11:00.672 "data_offset": 0, 00:11:00.672 "data_size": 0 00:11:00.672 }, 00:11:00.672 { 00:11:00.672 "name": "BaseBdev4", 00:11:00.672 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:00.672 "is_configured": false, 00:11:00.672 "data_offset": 0, 00:11:00.672 "data_size": 0 00:11:00.672 } 00:11:00.672 ] 00:11:00.672 }' 00:11:00.672 17:32:31 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:00.672 17:32:31 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:00.931 17:32:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:11:00.931 17:32:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:00.931 17:32:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:00.931 [2024-11-27 17:32:32.097716] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:11:00.931 [2024-11-27 17:32:32.097765] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:11:00.931 17:32:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:00.931 17:32:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:11:00.931 17:32:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:00.931 17:32:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:00.931 [2024-11-27 17:32:32.109729] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:11:00.931 [2024-11-27 17:32:32.109817] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:11:00.931 [2024-11-27 17:32:32.109844] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:11:00.931 [2024-11-27 17:32:32.109866] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:11:00.931 [2024-11-27 17:32:32.109883] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:11:00.931 [2024-11-27 17:32:32.109904] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:11:00.931 [2024-11-27 17:32:32.109936] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:11:00.931 [2024-11-27 17:32:32.109957] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:11:00.931 17:32:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:00.931 17:32:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:11:00.931 17:32:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:00.931 17:32:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:01.191 [2024-11-27 17:32:32.136950] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:11:01.191 BaseBdev1 00:11:01.191 17:32:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:01.191 17:32:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:11:01.191 17:32:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:11:01.191 17:32:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:11:01.191 17:32:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:11:01.191 17:32:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:11:01.191 17:32:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:11:01.191 17:32:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:11:01.191 17:32:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:01.191 17:32:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:01.191 17:32:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:01.191 17:32:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:11:01.191 17:32:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:01.191 17:32:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:01.191 [ 00:11:01.191 { 00:11:01.191 "name": "BaseBdev1", 00:11:01.191 "aliases": [ 00:11:01.191 "07b7172f-1053-4a57-8d5b-80ef5f27b256" 00:11:01.191 ], 00:11:01.191 "product_name": "Malloc disk", 00:11:01.191 "block_size": 512, 00:11:01.191 "num_blocks": 65536, 00:11:01.191 "uuid": "07b7172f-1053-4a57-8d5b-80ef5f27b256", 00:11:01.191 "assigned_rate_limits": { 00:11:01.191 "rw_ios_per_sec": 0, 00:11:01.191 "rw_mbytes_per_sec": 0, 00:11:01.191 "r_mbytes_per_sec": 0, 00:11:01.191 "w_mbytes_per_sec": 0 00:11:01.191 }, 00:11:01.191 "claimed": true, 00:11:01.191 "claim_type": "exclusive_write", 00:11:01.191 "zoned": false, 00:11:01.191 "supported_io_types": { 00:11:01.191 "read": true, 00:11:01.191 "write": true, 00:11:01.191 "unmap": true, 00:11:01.191 "flush": true, 00:11:01.191 "reset": true, 00:11:01.191 "nvme_admin": false, 00:11:01.191 "nvme_io": false, 00:11:01.191 "nvme_io_md": false, 00:11:01.191 "write_zeroes": true, 00:11:01.191 "zcopy": true, 00:11:01.191 "get_zone_info": false, 00:11:01.192 "zone_management": false, 00:11:01.192 "zone_append": false, 00:11:01.192 "compare": false, 00:11:01.192 "compare_and_write": false, 00:11:01.192 "abort": true, 00:11:01.192 "seek_hole": false, 00:11:01.192 "seek_data": false, 00:11:01.192 "copy": true, 00:11:01.192 "nvme_iov_md": false 00:11:01.192 }, 00:11:01.192 "memory_domains": [ 00:11:01.192 { 00:11:01.192 "dma_device_id": "system", 00:11:01.192 "dma_device_type": 1 00:11:01.192 }, 00:11:01.192 { 00:11:01.192 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:01.192 "dma_device_type": 2 00:11:01.192 } 00:11:01.192 ], 00:11:01.192 "driver_specific": {} 00:11:01.192 } 00:11:01.192 ] 00:11:01.192 17:32:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:01.192 17:32:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:11:01.192 17:32:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:11:01.192 17:32:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:01.192 17:32:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:01.192 17:32:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:01.192 17:32:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:01.192 17:32:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:01.192 17:32:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:01.192 17:32:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:01.192 17:32:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:01.192 17:32:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:01.192 17:32:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:01.192 17:32:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:01.192 17:32:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:01.192 17:32:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:01.192 17:32:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:01.192 17:32:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:01.192 "name": "Existed_Raid", 00:11:01.192 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:01.192 "strip_size_kb": 0, 00:11:01.192 "state": "configuring", 00:11:01.192 "raid_level": "raid1", 00:11:01.192 "superblock": false, 00:11:01.192 "num_base_bdevs": 4, 00:11:01.192 "num_base_bdevs_discovered": 1, 00:11:01.192 "num_base_bdevs_operational": 4, 00:11:01.192 "base_bdevs_list": [ 00:11:01.192 { 00:11:01.192 "name": "BaseBdev1", 00:11:01.192 "uuid": "07b7172f-1053-4a57-8d5b-80ef5f27b256", 00:11:01.192 "is_configured": true, 00:11:01.192 "data_offset": 0, 00:11:01.192 "data_size": 65536 00:11:01.192 }, 00:11:01.192 { 00:11:01.192 "name": "BaseBdev2", 00:11:01.192 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:01.192 "is_configured": false, 00:11:01.192 "data_offset": 0, 00:11:01.192 "data_size": 0 00:11:01.192 }, 00:11:01.192 { 00:11:01.192 "name": "BaseBdev3", 00:11:01.192 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:01.192 "is_configured": false, 00:11:01.192 "data_offset": 0, 00:11:01.192 "data_size": 0 00:11:01.192 }, 00:11:01.192 { 00:11:01.192 "name": "BaseBdev4", 00:11:01.192 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:01.192 "is_configured": false, 00:11:01.192 "data_offset": 0, 00:11:01.192 "data_size": 0 00:11:01.192 } 00:11:01.192 ] 00:11:01.192 }' 00:11:01.192 17:32:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:01.192 17:32:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:01.451 17:32:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:11:01.451 17:32:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:01.451 17:32:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:01.451 [2024-11-27 17:32:32.628141] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:11:01.451 [2024-11-27 17:32:32.628208] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:11:01.451 17:32:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:01.451 17:32:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:11:01.451 17:32:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:01.451 17:32:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:01.451 [2024-11-27 17:32:32.636190] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:11:01.451 [2024-11-27 17:32:32.638364] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:11:01.451 [2024-11-27 17:32:32.638452] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:11:01.451 [2024-11-27 17:32:32.638467] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:11:01.451 [2024-11-27 17:32:32.638476] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:11:01.451 [2024-11-27 17:32:32.638483] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:11:01.451 [2024-11-27 17:32:32.638491] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:11:01.711 17:32:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:01.711 17:32:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:11:01.711 17:32:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:11:01.711 17:32:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:11:01.711 17:32:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:01.711 17:32:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:01.711 17:32:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:01.711 17:32:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:01.711 17:32:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:01.711 17:32:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:01.711 17:32:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:01.711 17:32:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:01.711 17:32:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:01.711 17:32:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:01.711 17:32:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:01.711 17:32:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:01.711 17:32:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:01.711 17:32:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:01.711 17:32:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:01.711 "name": "Existed_Raid", 00:11:01.711 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:01.711 "strip_size_kb": 0, 00:11:01.711 "state": "configuring", 00:11:01.711 "raid_level": "raid1", 00:11:01.711 "superblock": false, 00:11:01.711 "num_base_bdevs": 4, 00:11:01.711 "num_base_bdevs_discovered": 1, 00:11:01.711 "num_base_bdevs_operational": 4, 00:11:01.711 "base_bdevs_list": [ 00:11:01.711 { 00:11:01.711 "name": "BaseBdev1", 00:11:01.711 "uuid": "07b7172f-1053-4a57-8d5b-80ef5f27b256", 00:11:01.711 "is_configured": true, 00:11:01.711 "data_offset": 0, 00:11:01.711 "data_size": 65536 00:11:01.711 }, 00:11:01.711 { 00:11:01.711 "name": "BaseBdev2", 00:11:01.711 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:01.711 "is_configured": false, 00:11:01.711 "data_offset": 0, 00:11:01.711 "data_size": 0 00:11:01.711 }, 00:11:01.711 { 00:11:01.711 "name": "BaseBdev3", 00:11:01.711 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:01.711 "is_configured": false, 00:11:01.711 "data_offset": 0, 00:11:01.711 "data_size": 0 00:11:01.711 }, 00:11:01.711 { 00:11:01.711 "name": "BaseBdev4", 00:11:01.711 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:01.711 "is_configured": false, 00:11:01.711 "data_offset": 0, 00:11:01.711 "data_size": 0 00:11:01.711 } 00:11:01.711 ] 00:11:01.711 }' 00:11:01.711 17:32:32 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:01.711 17:32:32 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:01.971 17:32:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:11:01.971 17:32:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:01.971 17:32:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:01.971 [2024-11-27 17:32:33.098602] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:11:01.971 BaseBdev2 00:11:01.971 17:32:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:01.971 17:32:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:11:01.971 17:32:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:11:01.971 17:32:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:11:01.971 17:32:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:11:01.971 17:32:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:11:01.971 17:32:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:11:01.971 17:32:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:11:01.972 17:32:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:01.972 17:32:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:01.972 17:32:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:01.972 17:32:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:11:01.972 17:32:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:01.972 17:32:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:01.972 [ 00:11:01.972 { 00:11:01.972 "name": "BaseBdev2", 00:11:01.972 "aliases": [ 00:11:01.972 "ccdf9d0c-1f43-4652-bbf5-5db02db5503a" 00:11:01.972 ], 00:11:01.972 "product_name": "Malloc disk", 00:11:01.972 "block_size": 512, 00:11:01.972 "num_blocks": 65536, 00:11:01.972 "uuid": "ccdf9d0c-1f43-4652-bbf5-5db02db5503a", 00:11:01.972 "assigned_rate_limits": { 00:11:01.972 "rw_ios_per_sec": 0, 00:11:01.972 "rw_mbytes_per_sec": 0, 00:11:01.972 "r_mbytes_per_sec": 0, 00:11:01.972 "w_mbytes_per_sec": 0 00:11:01.972 }, 00:11:01.972 "claimed": true, 00:11:01.972 "claim_type": "exclusive_write", 00:11:01.972 "zoned": false, 00:11:01.972 "supported_io_types": { 00:11:01.972 "read": true, 00:11:01.972 "write": true, 00:11:01.972 "unmap": true, 00:11:01.972 "flush": true, 00:11:01.972 "reset": true, 00:11:01.972 "nvme_admin": false, 00:11:01.972 "nvme_io": false, 00:11:01.972 "nvme_io_md": false, 00:11:01.972 "write_zeroes": true, 00:11:01.972 "zcopy": true, 00:11:01.972 "get_zone_info": false, 00:11:01.972 "zone_management": false, 00:11:01.972 "zone_append": false, 00:11:01.972 "compare": false, 00:11:01.972 "compare_and_write": false, 00:11:01.972 "abort": true, 00:11:01.972 "seek_hole": false, 00:11:01.972 "seek_data": false, 00:11:01.972 "copy": true, 00:11:01.972 "nvme_iov_md": false 00:11:01.972 }, 00:11:01.972 "memory_domains": [ 00:11:01.972 { 00:11:01.972 "dma_device_id": "system", 00:11:01.972 "dma_device_type": 1 00:11:01.972 }, 00:11:01.972 { 00:11:01.972 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:01.972 "dma_device_type": 2 00:11:01.972 } 00:11:01.972 ], 00:11:01.972 "driver_specific": {} 00:11:01.972 } 00:11:01.972 ] 00:11:01.972 17:32:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:01.972 17:32:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:11:01.972 17:32:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:11:01.972 17:32:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:11:01.972 17:32:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:11:01.972 17:32:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:01.972 17:32:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:01.972 17:32:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:01.972 17:32:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:01.972 17:32:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:01.972 17:32:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:01.972 17:32:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:01.972 17:32:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:01.972 17:32:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:01.972 17:32:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:01.972 17:32:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:01.972 17:32:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:01.972 17:32:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:02.231 17:32:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:02.231 17:32:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:02.231 "name": "Existed_Raid", 00:11:02.231 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:02.231 "strip_size_kb": 0, 00:11:02.231 "state": "configuring", 00:11:02.231 "raid_level": "raid1", 00:11:02.231 "superblock": false, 00:11:02.231 "num_base_bdevs": 4, 00:11:02.231 "num_base_bdevs_discovered": 2, 00:11:02.231 "num_base_bdevs_operational": 4, 00:11:02.231 "base_bdevs_list": [ 00:11:02.231 { 00:11:02.231 "name": "BaseBdev1", 00:11:02.231 "uuid": "07b7172f-1053-4a57-8d5b-80ef5f27b256", 00:11:02.231 "is_configured": true, 00:11:02.231 "data_offset": 0, 00:11:02.231 "data_size": 65536 00:11:02.231 }, 00:11:02.231 { 00:11:02.231 "name": "BaseBdev2", 00:11:02.231 "uuid": "ccdf9d0c-1f43-4652-bbf5-5db02db5503a", 00:11:02.231 "is_configured": true, 00:11:02.231 "data_offset": 0, 00:11:02.231 "data_size": 65536 00:11:02.231 }, 00:11:02.231 { 00:11:02.231 "name": "BaseBdev3", 00:11:02.231 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:02.231 "is_configured": false, 00:11:02.231 "data_offset": 0, 00:11:02.231 "data_size": 0 00:11:02.231 }, 00:11:02.231 { 00:11:02.231 "name": "BaseBdev4", 00:11:02.231 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:02.231 "is_configured": false, 00:11:02.231 "data_offset": 0, 00:11:02.231 "data_size": 0 00:11:02.231 } 00:11:02.231 ] 00:11:02.231 }' 00:11:02.231 17:32:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:02.231 17:32:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:02.491 17:32:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:11:02.491 17:32:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:02.491 17:32:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:02.491 [2024-11-27 17:32:33.587138] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:11:02.491 BaseBdev3 00:11:02.491 17:32:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:02.491 17:32:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:11:02.491 17:32:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:11:02.491 17:32:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:11:02.491 17:32:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:11:02.491 17:32:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:11:02.491 17:32:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:11:02.491 17:32:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:11:02.491 17:32:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:02.491 17:32:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:02.491 17:32:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:02.491 17:32:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:11:02.491 17:32:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:02.491 17:32:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:02.491 [ 00:11:02.491 { 00:11:02.491 "name": "BaseBdev3", 00:11:02.491 "aliases": [ 00:11:02.491 "f585671b-2b22-4875-9938-9de149c1eda6" 00:11:02.491 ], 00:11:02.491 "product_name": "Malloc disk", 00:11:02.491 "block_size": 512, 00:11:02.491 "num_blocks": 65536, 00:11:02.491 "uuid": "f585671b-2b22-4875-9938-9de149c1eda6", 00:11:02.491 "assigned_rate_limits": { 00:11:02.491 "rw_ios_per_sec": 0, 00:11:02.491 "rw_mbytes_per_sec": 0, 00:11:02.491 "r_mbytes_per_sec": 0, 00:11:02.491 "w_mbytes_per_sec": 0 00:11:02.491 }, 00:11:02.491 "claimed": true, 00:11:02.491 "claim_type": "exclusive_write", 00:11:02.491 "zoned": false, 00:11:02.491 "supported_io_types": { 00:11:02.491 "read": true, 00:11:02.491 "write": true, 00:11:02.491 "unmap": true, 00:11:02.492 "flush": true, 00:11:02.492 "reset": true, 00:11:02.492 "nvme_admin": false, 00:11:02.492 "nvme_io": false, 00:11:02.492 "nvme_io_md": false, 00:11:02.492 "write_zeroes": true, 00:11:02.492 "zcopy": true, 00:11:02.492 "get_zone_info": false, 00:11:02.492 "zone_management": false, 00:11:02.492 "zone_append": false, 00:11:02.492 "compare": false, 00:11:02.492 "compare_and_write": false, 00:11:02.492 "abort": true, 00:11:02.492 "seek_hole": false, 00:11:02.492 "seek_data": false, 00:11:02.492 "copy": true, 00:11:02.492 "nvme_iov_md": false 00:11:02.492 }, 00:11:02.492 "memory_domains": [ 00:11:02.492 { 00:11:02.492 "dma_device_id": "system", 00:11:02.492 "dma_device_type": 1 00:11:02.492 }, 00:11:02.492 { 00:11:02.492 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:02.492 "dma_device_type": 2 00:11:02.492 } 00:11:02.492 ], 00:11:02.492 "driver_specific": {} 00:11:02.492 } 00:11:02.492 ] 00:11:02.492 17:32:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:02.492 17:32:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:11:02.492 17:32:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:11:02.492 17:32:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:11:02.492 17:32:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:11:02.492 17:32:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:02.492 17:32:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:02.492 17:32:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:02.492 17:32:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:02.492 17:32:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:02.492 17:32:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:02.492 17:32:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:02.492 17:32:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:02.492 17:32:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:02.492 17:32:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:02.492 17:32:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:02.492 17:32:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:02.492 17:32:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:02.492 17:32:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:02.751 17:32:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:02.751 "name": "Existed_Raid", 00:11:02.751 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:02.751 "strip_size_kb": 0, 00:11:02.751 "state": "configuring", 00:11:02.751 "raid_level": "raid1", 00:11:02.751 "superblock": false, 00:11:02.751 "num_base_bdevs": 4, 00:11:02.751 "num_base_bdevs_discovered": 3, 00:11:02.751 "num_base_bdevs_operational": 4, 00:11:02.751 "base_bdevs_list": [ 00:11:02.751 { 00:11:02.751 "name": "BaseBdev1", 00:11:02.751 "uuid": "07b7172f-1053-4a57-8d5b-80ef5f27b256", 00:11:02.751 "is_configured": true, 00:11:02.751 "data_offset": 0, 00:11:02.751 "data_size": 65536 00:11:02.751 }, 00:11:02.751 { 00:11:02.751 "name": "BaseBdev2", 00:11:02.751 "uuid": "ccdf9d0c-1f43-4652-bbf5-5db02db5503a", 00:11:02.751 "is_configured": true, 00:11:02.751 "data_offset": 0, 00:11:02.751 "data_size": 65536 00:11:02.751 }, 00:11:02.751 { 00:11:02.751 "name": "BaseBdev3", 00:11:02.751 "uuid": "f585671b-2b22-4875-9938-9de149c1eda6", 00:11:02.751 "is_configured": true, 00:11:02.751 "data_offset": 0, 00:11:02.751 "data_size": 65536 00:11:02.751 }, 00:11:02.751 { 00:11:02.751 "name": "BaseBdev4", 00:11:02.751 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:02.751 "is_configured": false, 00:11:02.751 "data_offset": 0, 00:11:02.751 "data_size": 0 00:11:02.751 } 00:11:02.751 ] 00:11:02.751 }' 00:11:02.751 17:32:33 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:02.752 17:32:33 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:03.012 17:32:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:11:03.012 17:32:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:03.012 17:32:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:03.012 [2024-11-27 17:32:34.087278] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:11:03.012 [2024-11-27 17:32:34.087401] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:11:03.012 [2024-11-27 17:32:34.087415] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:11:03.012 [2024-11-27 17:32:34.087740] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:11:03.012 [2024-11-27 17:32:34.087916] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:11:03.012 [2024-11-27 17:32:34.087930] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:11:03.012 [2024-11-27 17:32:34.088201] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:03.012 BaseBdev4 00:11:03.012 17:32:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:03.012 17:32:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev4 00:11:03.012 17:32:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev4 00:11:03.012 17:32:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:11:03.012 17:32:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:11:03.012 17:32:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:11:03.012 17:32:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:11:03.012 17:32:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:11:03.012 17:32:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:03.012 17:32:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:03.012 17:32:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:03.012 17:32:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:11:03.012 17:32:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:03.012 17:32:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:03.012 [ 00:11:03.012 { 00:11:03.012 "name": "BaseBdev4", 00:11:03.012 "aliases": [ 00:11:03.012 "f0f872d1-aff8-429d-b9a5-3b4fbb1e28de" 00:11:03.012 ], 00:11:03.012 "product_name": "Malloc disk", 00:11:03.012 "block_size": 512, 00:11:03.012 "num_blocks": 65536, 00:11:03.012 "uuid": "f0f872d1-aff8-429d-b9a5-3b4fbb1e28de", 00:11:03.012 "assigned_rate_limits": { 00:11:03.012 "rw_ios_per_sec": 0, 00:11:03.012 "rw_mbytes_per_sec": 0, 00:11:03.012 "r_mbytes_per_sec": 0, 00:11:03.012 "w_mbytes_per_sec": 0 00:11:03.012 }, 00:11:03.012 "claimed": true, 00:11:03.012 "claim_type": "exclusive_write", 00:11:03.012 "zoned": false, 00:11:03.012 "supported_io_types": { 00:11:03.012 "read": true, 00:11:03.012 "write": true, 00:11:03.012 "unmap": true, 00:11:03.012 "flush": true, 00:11:03.012 "reset": true, 00:11:03.012 "nvme_admin": false, 00:11:03.012 "nvme_io": false, 00:11:03.012 "nvme_io_md": false, 00:11:03.012 "write_zeroes": true, 00:11:03.012 "zcopy": true, 00:11:03.012 "get_zone_info": false, 00:11:03.012 "zone_management": false, 00:11:03.012 "zone_append": false, 00:11:03.012 "compare": false, 00:11:03.012 "compare_and_write": false, 00:11:03.012 "abort": true, 00:11:03.012 "seek_hole": false, 00:11:03.012 "seek_data": false, 00:11:03.012 "copy": true, 00:11:03.012 "nvme_iov_md": false 00:11:03.012 }, 00:11:03.012 "memory_domains": [ 00:11:03.012 { 00:11:03.012 "dma_device_id": "system", 00:11:03.012 "dma_device_type": 1 00:11:03.012 }, 00:11:03.012 { 00:11:03.012 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:03.012 "dma_device_type": 2 00:11:03.012 } 00:11:03.012 ], 00:11:03.012 "driver_specific": {} 00:11:03.012 } 00:11:03.012 ] 00:11:03.012 17:32:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:03.012 17:32:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:11:03.012 17:32:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:11:03.012 17:32:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:11:03.012 17:32:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 4 00:11:03.012 17:32:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:03.012 17:32:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:03.012 17:32:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:03.012 17:32:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:03.012 17:32:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:03.012 17:32:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:03.012 17:32:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:03.012 17:32:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:03.012 17:32:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:03.012 17:32:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:03.012 17:32:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:03.012 17:32:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:03.012 17:32:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:03.012 17:32:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:03.012 17:32:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:03.012 "name": "Existed_Raid", 00:11:03.012 "uuid": "a5c0a359-5322-4a9d-b9af-e4793dea0cd4", 00:11:03.012 "strip_size_kb": 0, 00:11:03.012 "state": "online", 00:11:03.012 "raid_level": "raid1", 00:11:03.012 "superblock": false, 00:11:03.012 "num_base_bdevs": 4, 00:11:03.012 "num_base_bdevs_discovered": 4, 00:11:03.012 "num_base_bdevs_operational": 4, 00:11:03.012 "base_bdevs_list": [ 00:11:03.012 { 00:11:03.012 "name": "BaseBdev1", 00:11:03.012 "uuid": "07b7172f-1053-4a57-8d5b-80ef5f27b256", 00:11:03.012 "is_configured": true, 00:11:03.012 "data_offset": 0, 00:11:03.012 "data_size": 65536 00:11:03.012 }, 00:11:03.012 { 00:11:03.012 "name": "BaseBdev2", 00:11:03.012 "uuid": "ccdf9d0c-1f43-4652-bbf5-5db02db5503a", 00:11:03.012 "is_configured": true, 00:11:03.012 "data_offset": 0, 00:11:03.012 "data_size": 65536 00:11:03.012 }, 00:11:03.012 { 00:11:03.012 "name": "BaseBdev3", 00:11:03.012 "uuid": "f585671b-2b22-4875-9938-9de149c1eda6", 00:11:03.012 "is_configured": true, 00:11:03.012 "data_offset": 0, 00:11:03.012 "data_size": 65536 00:11:03.012 }, 00:11:03.012 { 00:11:03.012 "name": "BaseBdev4", 00:11:03.012 "uuid": "f0f872d1-aff8-429d-b9a5-3b4fbb1e28de", 00:11:03.012 "is_configured": true, 00:11:03.012 "data_offset": 0, 00:11:03.012 "data_size": 65536 00:11:03.012 } 00:11:03.012 ] 00:11:03.012 }' 00:11:03.012 17:32:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:03.012 17:32:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:03.582 17:32:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:11:03.582 17:32:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:11:03.582 17:32:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:11:03.582 17:32:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:11:03.582 17:32:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:11:03.582 17:32:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:11:03.582 17:32:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:11:03.582 17:32:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:11:03.582 17:32:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:03.582 17:32:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:03.582 [2024-11-27 17:32:34.594913] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:11:03.582 17:32:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:03.582 17:32:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:11:03.582 "name": "Existed_Raid", 00:11:03.582 "aliases": [ 00:11:03.582 "a5c0a359-5322-4a9d-b9af-e4793dea0cd4" 00:11:03.582 ], 00:11:03.582 "product_name": "Raid Volume", 00:11:03.582 "block_size": 512, 00:11:03.582 "num_blocks": 65536, 00:11:03.582 "uuid": "a5c0a359-5322-4a9d-b9af-e4793dea0cd4", 00:11:03.582 "assigned_rate_limits": { 00:11:03.582 "rw_ios_per_sec": 0, 00:11:03.582 "rw_mbytes_per_sec": 0, 00:11:03.582 "r_mbytes_per_sec": 0, 00:11:03.582 "w_mbytes_per_sec": 0 00:11:03.582 }, 00:11:03.582 "claimed": false, 00:11:03.582 "zoned": false, 00:11:03.582 "supported_io_types": { 00:11:03.582 "read": true, 00:11:03.582 "write": true, 00:11:03.582 "unmap": false, 00:11:03.582 "flush": false, 00:11:03.582 "reset": true, 00:11:03.582 "nvme_admin": false, 00:11:03.582 "nvme_io": false, 00:11:03.582 "nvme_io_md": false, 00:11:03.582 "write_zeroes": true, 00:11:03.582 "zcopy": false, 00:11:03.582 "get_zone_info": false, 00:11:03.582 "zone_management": false, 00:11:03.582 "zone_append": false, 00:11:03.582 "compare": false, 00:11:03.582 "compare_and_write": false, 00:11:03.582 "abort": false, 00:11:03.582 "seek_hole": false, 00:11:03.582 "seek_data": false, 00:11:03.582 "copy": false, 00:11:03.582 "nvme_iov_md": false 00:11:03.582 }, 00:11:03.582 "memory_domains": [ 00:11:03.582 { 00:11:03.582 "dma_device_id": "system", 00:11:03.583 "dma_device_type": 1 00:11:03.583 }, 00:11:03.583 { 00:11:03.583 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:03.583 "dma_device_type": 2 00:11:03.583 }, 00:11:03.583 { 00:11:03.583 "dma_device_id": "system", 00:11:03.583 "dma_device_type": 1 00:11:03.583 }, 00:11:03.583 { 00:11:03.583 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:03.583 "dma_device_type": 2 00:11:03.583 }, 00:11:03.583 { 00:11:03.583 "dma_device_id": "system", 00:11:03.583 "dma_device_type": 1 00:11:03.583 }, 00:11:03.583 { 00:11:03.583 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:03.583 "dma_device_type": 2 00:11:03.583 }, 00:11:03.583 { 00:11:03.583 "dma_device_id": "system", 00:11:03.583 "dma_device_type": 1 00:11:03.583 }, 00:11:03.583 { 00:11:03.583 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:03.583 "dma_device_type": 2 00:11:03.583 } 00:11:03.583 ], 00:11:03.583 "driver_specific": { 00:11:03.583 "raid": { 00:11:03.583 "uuid": "a5c0a359-5322-4a9d-b9af-e4793dea0cd4", 00:11:03.583 "strip_size_kb": 0, 00:11:03.583 "state": "online", 00:11:03.583 "raid_level": "raid1", 00:11:03.583 "superblock": false, 00:11:03.583 "num_base_bdevs": 4, 00:11:03.583 "num_base_bdevs_discovered": 4, 00:11:03.583 "num_base_bdevs_operational": 4, 00:11:03.583 "base_bdevs_list": [ 00:11:03.583 { 00:11:03.583 "name": "BaseBdev1", 00:11:03.583 "uuid": "07b7172f-1053-4a57-8d5b-80ef5f27b256", 00:11:03.583 "is_configured": true, 00:11:03.583 "data_offset": 0, 00:11:03.583 "data_size": 65536 00:11:03.583 }, 00:11:03.583 { 00:11:03.583 "name": "BaseBdev2", 00:11:03.583 "uuid": "ccdf9d0c-1f43-4652-bbf5-5db02db5503a", 00:11:03.583 "is_configured": true, 00:11:03.583 "data_offset": 0, 00:11:03.583 "data_size": 65536 00:11:03.583 }, 00:11:03.583 { 00:11:03.583 "name": "BaseBdev3", 00:11:03.583 "uuid": "f585671b-2b22-4875-9938-9de149c1eda6", 00:11:03.583 "is_configured": true, 00:11:03.583 "data_offset": 0, 00:11:03.583 "data_size": 65536 00:11:03.583 }, 00:11:03.583 { 00:11:03.583 "name": "BaseBdev4", 00:11:03.583 "uuid": "f0f872d1-aff8-429d-b9a5-3b4fbb1e28de", 00:11:03.583 "is_configured": true, 00:11:03.583 "data_offset": 0, 00:11:03.583 "data_size": 65536 00:11:03.583 } 00:11:03.583 ] 00:11:03.583 } 00:11:03.583 } 00:11:03.583 }' 00:11:03.583 17:32:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:11:03.583 17:32:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:11:03.583 BaseBdev2 00:11:03.583 BaseBdev3 00:11:03.583 BaseBdev4' 00:11:03.583 17:32:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:03.583 17:32:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:11:03.583 17:32:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:03.583 17:32:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:03.583 17:32:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:11:03.583 17:32:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:03.583 17:32:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:03.583 17:32:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:03.583 17:32:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:03.583 17:32:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:03.583 17:32:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:03.583 17:32:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:11:03.583 17:32:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:03.583 17:32:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:03.583 17:32:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:03.844 17:32:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:03.844 17:32:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:03.844 17:32:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:03.844 17:32:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:03.844 17:32:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:11:03.844 17:32:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:03.844 17:32:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:03.844 17:32:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:03.844 17:32:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:03.844 17:32:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:03.844 17:32:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:03.844 17:32:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:03.844 17:32:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:03.844 17:32:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:11:03.844 17:32:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:03.844 17:32:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:03.844 17:32:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:03.844 17:32:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:03.844 17:32:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:03.844 17:32:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:11:03.844 17:32:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:03.844 17:32:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:03.844 [2024-11-27 17:32:34.918020] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:11:03.844 17:32:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:03.844 17:32:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:11:03.844 17:32:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:11:03.844 17:32:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:11:03.844 17:32:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@199 -- # return 0 00:11:03.844 17:32:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:11:03.844 17:32:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 3 00:11:03.844 17:32:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:03.844 17:32:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:03.844 17:32:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:03.844 17:32:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:03.844 17:32:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:11:03.844 17:32:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:03.844 17:32:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:03.844 17:32:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:03.844 17:32:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:03.844 17:32:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:03.844 17:32:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:03.844 17:32:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:03.844 17:32:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:03.844 17:32:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:03.844 17:32:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:03.844 "name": "Existed_Raid", 00:11:03.844 "uuid": "a5c0a359-5322-4a9d-b9af-e4793dea0cd4", 00:11:03.844 "strip_size_kb": 0, 00:11:03.844 "state": "online", 00:11:03.844 "raid_level": "raid1", 00:11:03.844 "superblock": false, 00:11:03.845 "num_base_bdevs": 4, 00:11:03.845 "num_base_bdevs_discovered": 3, 00:11:03.845 "num_base_bdevs_operational": 3, 00:11:03.845 "base_bdevs_list": [ 00:11:03.845 { 00:11:03.845 "name": null, 00:11:03.845 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:03.845 "is_configured": false, 00:11:03.845 "data_offset": 0, 00:11:03.845 "data_size": 65536 00:11:03.845 }, 00:11:03.845 { 00:11:03.845 "name": "BaseBdev2", 00:11:03.845 "uuid": "ccdf9d0c-1f43-4652-bbf5-5db02db5503a", 00:11:03.845 "is_configured": true, 00:11:03.845 "data_offset": 0, 00:11:03.845 "data_size": 65536 00:11:03.845 }, 00:11:03.845 { 00:11:03.845 "name": "BaseBdev3", 00:11:03.845 "uuid": "f585671b-2b22-4875-9938-9de149c1eda6", 00:11:03.845 "is_configured": true, 00:11:03.845 "data_offset": 0, 00:11:03.845 "data_size": 65536 00:11:03.845 }, 00:11:03.845 { 00:11:03.845 "name": "BaseBdev4", 00:11:03.845 "uuid": "f0f872d1-aff8-429d-b9a5-3b4fbb1e28de", 00:11:03.845 "is_configured": true, 00:11:03.845 "data_offset": 0, 00:11:03.845 "data_size": 65536 00:11:03.845 } 00:11:03.845 ] 00:11:03.845 }' 00:11:03.845 17:32:34 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:03.845 17:32:34 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:04.415 17:32:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:11:04.415 17:32:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:11:04.415 17:32:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:04.415 17:32:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:04.415 17:32:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:11:04.415 17:32:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:04.415 17:32:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:04.415 17:32:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:11:04.415 17:32:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:11:04.415 17:32:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:11:04.415 17:32:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:04.415 17:32:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:04.415 [2024-11-27 17:32:35.378250] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:11:04.415 17:32:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:04.415 17:32:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:11:04.415 17:32:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:11:04.415 17:32:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:04.415 17:32:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:04.415 17:32:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:04.415 17:32:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:11:04.415 17:32:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:04.415 17:32:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:11:04.415 17:32:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:11:04.415 17:32:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:11:04.415 17:32:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:04.415 17:32:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:04.415 [2024-11-27 17:32:35.454275] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:11:04.415 17:32:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:04.415 17:32:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:11:04.415 17:32:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:11:04.415 17:32:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:04.415 17:32:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:04.415 17:32:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:04.415 17:32:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:11:04.415 17:32:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:04.415 17:32:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:11:04.415 17:32:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:11:04.415 17:32:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev4 00:11:04.415 17:32:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:04.415 17:32:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:04.415 [2024-11-27 17:32:35.534940] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev4 00:11:04.415 [2024-11-27 17:32:35.535035] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:11:04.415 [2024-11-27 17:32:35.555519] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:11:04.415 [2024-11-27 17:32:35.555573] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:11:04.415 [2024-11-27 17:32:35.555600] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:11:04.415 17:32:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:04.415 17:32:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:11:04.415 17:32:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:11:04.415 17:32:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:04.415 17:32:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:04.415 17:32:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:04.415 17:32:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:11:04.415 17:32:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:04.676 17:32:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:11:04.676 17:32:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:11:04.676 17:32:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 4 -gt 2 ']' 00:11:04.676 17:32:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:11:04.676 17:32:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:11:04.676 17:32:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:11:04.676 17:32:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:04.676 17:32:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:04.676 BaseBdev2 00:11:04.676 17:32:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:04.676 17:32:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:11:04.676 17:32:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:11:04.676 17:32:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:11:04.676 17:32:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:11:04.676 17:32:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:11:04.676 17:32:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:11:04.676 17:32:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:11:04.676 17:32:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:04.676 17:32:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:04.676 17:32:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:04.676 17:32:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:11:04.676 17:32:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:04.676 17:32:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:04.676 [ 00:11:04.676 { 00:11:04.676 "name": "BaseBdev2", 00:11:04.676 "aliases": [ 00:11:04.676 "b98327fe-5d91-4798-a1f2-f12a740e5156" 00:11:04.676 ], 00:11:04.676 "product_name": "Malloc disk", 00:11:04.676 "block_size": 512, 00:11:04.676 "num_blocks": 65536, 00:11:04.676 "uuid": "b98327fe-5d91-4798-a1f2-f12a740e5156", 00:11:04.676 "assigned_rate_limits": { 00:11:04.676 "rw_ios_per_sec": 0, 00:11:04.676 "rw_mbytes_per_sec": 0, 00:11:04.676 "r_mbytes_per_sec": 0, 00:11:04.676 "w_mbytes_per_sec": 0 00:11:04.676 }, 00:11:04.676 "claimed": false, 00:11:04.677 "zoned": false, 00:11:04.677 "supported_io_types": { 00:11:04.677 "read": true, 00:11:04.677 "write": true, 00:11:04.677 "unmap": true, 00:11:04.677 "flush": true, 00:11:04.677 "reset": true, 00:11:04.677 "nvme_admin": false, 00:11:04.677 "nvme_io": false, 00:11:04.677 "nvme_io_md": false, 00:11:04.677 "write_zeroes": true, 00:11:04.677 "zcopy": true, 00:11:04.677 "get_zone_info": false, 00:11:04.677 "zone_management": false, 00:11:04.677 "zone_append": false, 00:11:04.677 "compare": false, 00:11:04.677 "compare_and_write": false, 00:11:04.677 "abort": true, 00:11:04.677 "seek_hole": false, 00:11:04.677 "seek_data": false, 00:11:04.677 "copy": true, 00:11:04.677 "nvme_iov_md": false 00:11:04.677 }, 00:11:04.677 "memory_domains": [ 00:11:04.677 { 00:11:04.677 "dma_device_id": "system", 00:11:04.677 "dma_device_type": 1 00:11:04.677 }, 00:11:04.677 { 00:11:04.677 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:04.677 "dma_device_type": 2 00:11:04.677 } 00:11:04.677 ], 00:11:04.677 "driver_specific": {} 00:11:04.677 } 00:11:04.677 ] 00:11:04.677 17:32:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:04.677 17:32:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:11:04.677 17:32:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:11:04.677 17:32:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:11:04.677 17:32:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:11:04.677 17:32:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:04.677 17:32:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:04.677 BaseBdev3 00:11:04.677 17:32:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:04.677 17:32:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:11:04.677 17:32:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:11:04.677 17:32:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:11:04.677 17:32:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:11:04.677 17:32:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:11:04.677 17:32:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:11:04.677 17:32:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:11:04.677 17:32:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:04.677 17:32:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:04.677 17:32:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:04.677 17:32:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:11:04.677 17:32:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:04.677 17:32:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:04.677 [ 00:11:04.677 { 00:11:04.677 "name": "BaseBdev3", 00:11:04.677 "aliases": [ 00:11:04.677 "1bc274d6-5e32-4c98-8904-57eecd247e0b" 00:11:04.677 ], 00:11:04.677 "product_name": "Malloc disk", 00:11:04.677 "block_size": 512, 00:11:04.677 "num_blocks": 65536, 00:11:04.677 "uuid": "1bc274d6-5e32-4c98-8904-57eecd247e0b", 00:11:04.677 "assigned_rate_limits": { 00:11:04.677 "rw_ios_per_sec": 0, 00:11:04.677 "rw_mbytes_per_sec": 0, 00:11:04.677 "r_mbytes_per_sec": 0, 00:11:04.677 "w_mbytes_per_sec": 0 00:11:04.677 }, 00:11:04.677 "claimed": false, 00:11:04.677 "zoned": false, 00:11:04.677 "supported_io_types": { 00:11:04.677 "read": true, 00:11:04.677 "write": true, 00:11:04.677 "unmap": true, 00:11:04.677 "flush": true, 00:11:04.677 "reset": true, 00:11:04.677 "nvme_admin": false, 00:11:04.677 "nvme_io": false, 00:11:04.677 "nvme_io_md": false, 00:11:04.677 "write_zeroes": true, 00:11:04.677 "zcopy": true, 00:11:04.677 "get_zone_info": false, 00:11:04.677 "zone_management": false, 00:11:04.677 "zone_append": false, 00:11:04.677 "compare": false, 00:11:04.677 "compare_and_write": false, 00:11:04.677 "abort": true, 00:11:04.677 "seek_hole": false, 00:11:04.677 "seek_data": false, 00:11:04.677 "copy": true, 00:11:04.677 "nvme_iov_md": false 00:11:04.677 }, 00:11:04.677 "memory_domains": [ 00:11:04.677 { 00:11:04.677 "dma_device_id": "system", 00:11:04.677 "dma_device_type": 1 00:11:04.677 }, 00:11:04.677 { 00:11:04.677 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:04.677 "dma_device_type": 2 00:11:04.677 } 00:11:04.677 ], 00:11:04.677 "driver_specific": {} 00:11:04.677 } 00:11:04.677 ] 00:11:04.677 17:32:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:04.677 17:32:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:11:04.677 17:32:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:11:04.677 17:32:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:11:04.677 17:32:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:11:04.677 17:32:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:04.677 17:32:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:04.677 BaseBdev4 00:11:04.677 17:32:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:04.677 17:32:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev4 00:11:04.677 17:32:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev4 00:11:04.677 17:32:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:11:04.677 17:32:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:11:04.677 17:32:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:11:04.677 17:32:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:11:04.677 17:32:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:11:04.677 17:32:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:04.677 17:32:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:04.677 17:32:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:04.677 17:32:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:11:04.677 17:32:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:04.677 17:32:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:04.677 [ 00:11:04.677 { 00:11:04.677 "name": "BaseBdev4", 00:11:04.677 "aliases": [ 00:11:04.677 "c0a4ab88-e694-401c-b920-800a8cef9bf2" 00:11:04.677 ], 00:11:04.677 "product_name": "Malloc disk", 00:11:04.677 "block_size": 512, 00:11:04.677 "num_blocks": 65536, 00:11:04.677 "uuid": "c0a4ab88-e694-401c-b920-800a8cef9bf2", 00:11:04.677 "assigned_rate_limits": { 00:11:04.677 "rw_ios_per_sec": 0, 00:11:04.677 "rw_mbytes_per_sec": 0, 00:11:04.677 "r_mbytes_per_sec": 0, 00:11:04.677 "w_mbytes_per_sec": 0 00:11:04.677 }, 00:11:04.677 "claimed": false, 00:11:04.677 "zoned": false, 00:11:04.677 "supported_io_types": { 00:11:04.677 "read": true, 00:11:04.677 "write": true, 00:11:04.677 "unmap": true, 00:11:04.677 "flush": true, 00:11:04.677 "reset": true, 00:11:04.677 "nvme_admin": false, 00:11:04.677 "nvme_io": false, 00:11:04.677 "nvme_io_md": false, 00:11:04.677 "write_zeroes": true, 00:11:04.677 "zcopy": true, 00:11:04.677 "get_zone_info": false, 00:11:04.677 "zone_management": false, 00:11:04.677 "zone_append": false, 00:11:04.677 "compare": false, 00:11:04.677 "compare_and_write": false, 00:11:04.677 "abort": true, 00:11:04.677 "seek_hole": false, 00:11:04.677 "seek_data": false, 00:11:04.677 "copy": true, 00:11:04.677 "nvme_iov_md": false 00:11:04.677 }, 00:11:04.677 "memory_domains": [ 00:11:04.677 { 00:11:04.677 "dma_device_id": "system", 00:11:04.677 "dma_device_type": 1 00:11:04.677 }, 00:11:04.677 { 00:11:04.677 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:04.677 "dma_device_type": 2 00:11:04.677 } 00:11:04.677 ], 00:11:04.677 "driver_specific": {} 00:11:04.677 } 00:11:04.677 ] 00:11:04.677 17:32:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:04.677 17:32:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:11:04.677 17:32:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:11:04.677 17:32:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:11:04.677 17:32:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:11:04.677 17:32:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:04.677 17:32:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:04.677 [2024-11-27 17:32:35.799970] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:11:04.677 [2024-11-27 17:32:35.800063] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:11:04.677 [2024-11-27 17:32:35.800102] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:11:04.677 [2024-11-27 17:32:35.802221] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:11:04.677 [2024-11-27 17:32:35.802314] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:11:04.677 17:32:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:04.677 17:32:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:11:04.677 17:32:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:04.678 17:32:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:04.678 17:32:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:04.678 17:32:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:04.678 17:32:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:04.678 17:32:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:04.678 17:32:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:04.678 17:32:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:04.678 17:32:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:04.678 17:32:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:04.678 17:32:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:04.678 17:32:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:04.678 17:32:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:04.678 17:32:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:04.678 17:32:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:04.678 "name": "Existed_Raid", 00:11:04.678 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:04.678 "strip_size_kb": 0, 00:11:04.678 "state": "configuring", 00:11:04.678 "raid_level": "raid1", 00:11:04.678 "superblock": false, 00:11:04.678 "num_base_bdevs": 4, 00:11:04.678 "num_base_bdevs_discovered": 3, 00:11:04.678 "num_base_bdevs_operational": 4, 00:11:04.678 "base_bdevs_list": [ 00:11:04.678 { 00:11:04.678 "name": "BaseBdev1", 00:11:04.678 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:04.678 "is_configured": false, 00:11:04.678 "data_offset": 0, 00:11:04.678 "data_size": 0 00:11:04.678 }, 00:11:04.678 { 00:11:04.678 "name": "BaseBdev2", 00:11:04.678 "uuid": "b98327fe-5d91-4798-a1f2-f12a740e5156", 00:11:04.678 "is_configured": true, 00:11:04.678 "data_offset": 0, 00:11:04.678 "data_size": 65536 00:11:04.678 }, 00:11:04.678 { 00:11:04.678 "name": "BaseBdev3", 00:11:04.678 "uuid": "1bc274d6-5e32-4c98-8904-57eecd247e0b", 00:11:04.678 "is_configured": true, 00:11:04.678 "data_offset": 0, 00:11:04.678 "data_size": 65536 00:11:04.678 }, 00:11:04.678 { 00:11:04.678 "name": "BaseBdev4", 00:11:04.678 "uuid": "c0a4ab88-e694-401c-b920-800a8cef9bf2", 00:11:04.678 "is_configured": true, 00:11:04.678 "data_offset": 0, 00:11:04.678 "data_size": 65536 00:11:04.678 } 00:11:04.678 ] 00:11:04.678 }' 00:11:04.678 17:32:35 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:04.678 17:32:35 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:05.247 17:32:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:11:05.247 17:32:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:05.247 17:32:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:05.247 [2024-11-27 17:32:36.223252] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:11:05.247 17:32:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:05.247 17:32:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:11:05.247 17:32:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:05.247 17:32:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:05.247 17:32:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:05.247 17:32:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:05.247 17:32:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:05.247 17:32:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:05.247 17:32:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:05.247 17:32:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:05.247 17:32:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:05.247 17:32:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:05.247 17:32:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:05.247 17:32:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:05.247 17:32:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:05.247 17:32:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:05.247 17:32:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:05.247 "name": "Existed_Raid", 00:11:05.247 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:05.247 "strip_size_kb": 0, 00:11:05.247 "state": "configuring", 00:11:05.247 "raid_level": "raid1", 00:11:05.247 "superblock": false, 00:11:05.247 "num_base_bdevs": 4, 00:11:05.247 "num_base_bdevs_discovered": 2, 00:11:05.247 "num_base_bdevs_operational": 4, 00:11:05.247 "base_bdevs_list": [ 00:11:05.247 { 00:11:05.247 "name": "BaseBdev1", 00:11:05.247 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:05.247 "is_configured": false, 00:11:05.247 "data_offset": 0, 00:11:05.247 "data_size": 0 00:11:05.247 }, 00:11:05.247 { 00:11:05.247 "name": null, 00:11:05.247 "uuid": "b98327fe-5d91-4798-a1f2-f12a740e5156", 00:11:05.247 "is_configured": false, 00:11:05.247 "data_offset": 0, 00:11:05.247 "data_size": 65536 00:11:05.247 }, 00:11:05.247 { 00:11:05.247 "name": "BaseBdev3", 00:11:05.247 "uuid": "1bc274d6-5e32-4c98-8904-57eecd247e0b", 00:11:05.247 "is_configured": true, 00:11:05.247 "data_offset": 0, 00:11:05.247 "data_size": 65536 00:11:05.247 }, 00:11:05.247 { 00:11:05.247 "name": "BaseBdev4", 00:11:05.247 "uuid": "c0a4ab88-e694-401c-b920-800a8cef9bf2", 00:11:05.247 "is_configured": true, 00:11:05.247 "data_offset": 0, 00:11:05.247 "data_size": 65536 00:11:05.247 } 00:11:05.247 ] 00:11:05.247 }' 00:11:05.247 17:32:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:05.247 17:32:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:05.507 17:32:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:05.507 17:32:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:05.508 17:32:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:11:05.508 17:32:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:05.508 17:32:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:05.508 17:32:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:11:05.508 17:32:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:11:05.508 17:32:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:05.508 17:32:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:05.769 [2024-11-27 17:32:36.707553] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:11:05.769 BaseBdev1 00:11:05.769 17:32:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:05.769 17:32:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:11:05.769 17:32:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:11:05.769 17:32:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:11:05.769 17:32:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:11:05.769 17:32:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:11:05.769 17:32:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:11:05.769 17:32:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:11:05.769 17:32:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:05.769 17:32:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:05.769 17:32:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:05.769 17:32:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:11:05.769 17:32:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:05.769 17:32:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:05.769 [ 00:11:05.770 { 00:11:05.770 "name": "BaseBdev1", 00:11:05.770 "aliases": [ 00:11:05.770 "f6c97c6b-cf05-4a2f-8744-6665284eafb8" 00:11:05.770 ], 00:11:05.770 "product_name": "Malloc disk", 00:11:05.770 "block_size": 512, 00:11:05.770 "num_blocks": 65536, 00:11:05.770 "uuid": "f6c97c6b-cf05-4a2f-8744-6665284eafb8", 00:11:05.770 "assigned_rate_limits": { 00:11:05.770 "rw_ios_per_sec": 0, 00:11:05.770 "rw_mbytes_per_sec": 0, 00:11:05.770 "r_mbytes_per_sec": 0, 00:11:05.770 "w_mbytes_per_sec": 0 00:11:05.770 }, 00:11:05.770 "claimed": true, 00:11:05.770 "claim_type": "exclusive_write", 00:11:05.770 "zoned": false, 00:11:05.770 "supported_io_types": { 00:11:05.770 "read": true, 00:11:05.770 "write": true, 00:11:05.770 "unmap": true, 00:11:05.770 "flush": true, 00:11:05.770 "reset": true, 00:11:05.770 "nvme_admin": false, 00:11:05.770 "nvme_io": false, 00:11:05.770 "nvme_io_md": false, 00:11:05.770 "write_zeroes": true, 00:11:05.770 "zcopy": true, 00:11:05.770 "get_zone_info": false, 00:11:05.770 "zone_management": false, 00:11:05.770 "zone_append": false, 00:11:05.770 "compare": false, 00:11:05.770 "compare_and_write": false, 00:11:05.770 "abort": true, 00:11:05.770 "seek_hole": false, 00:11:05.770 "seek_data": false, 00:11:05.770 "copy": true, 00:11:05.770 "nvme_iov_md": false 00:11:05.770 }, 00:11:05.770 "memory_domains": [ 00:11:05.770 { 00:11:05.770 "dma_device_id": "system", 00:11:05.770 "dma_device_type": 1 00:11:05.770 }, 00:11:05.770 { 00:11:05.770 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:05.770 "dma_device_type": 2 00:11:05.770 } 00:11:05.770 ], 00:11:05.770 "driver_specific": {} 00:11:05.770 } 00:11:05.770 ] 00:11:05.770 17:32:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:05.770 17:32:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:11:05.770 17:32:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:11:05.770 17:32:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:05.770 17:32:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:05.770 17:32:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:05.770 17:32:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:05.770 17:32:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:05.770 17:32:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:05.770 17:32:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:05.770 17:32:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:05.770 17:32:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:05.770 17:32:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:05.770 17:32:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:05.770 17:32:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:05.770 17:32:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:05.770 17:32:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:05.770 17:32:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:05.770 "name": "Existed_Raid", 00:11:05.770 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:05.770 "strip_size_kb": 0, 00:11:05.770 "state": "configuring", 00:11:05.770 "raid_level": "raid1", 00:11:05.770 "superblock": false, 00:11:05.770 "num_base_bdevs": 4, 00:11:05.770 "num_base_bdevs_discovered": 3, 00:11:05.770 "num_base_bdevs_operational": 4, 00:11:05.770 "base_bdevs_list": [ 00:11:05.770 { 00:11:05.770 "name": "BaseBdev1", 00:11:05.770 "uuid": "f6c97c6b-cf05-4a2f-8744-6665284eafb8", 00:11:05.770 "is_configured": true, 00:11:05.770 "data_offset": 0, 00:11:05.770 "data_size": 65536 00:11:05.770 }, 00:11:05.770 { 00:11:05.770 "name": null, 00:11:05.770 "uuid": "b98327fe-5d91-4798-a1f2-f12a740e5156", 00:11:05.770 "is_configured": false, 00:11:05.770 "data_offset": 0, 00:11:05.770 "data_size": 65536 00:11:05.770 }, 00:11:05.770 { 00:11:05.770 "name": "BaseBdev3", 00:11:05.770 "uuid": "1bc274d6-5e32-4c98-8904-57eecd247e0b", 00:11:05.770 "is_configured": true, 00:11:05.770 "data_offset": 0, 00:11:05.770 "data_size": 65536 00:11:05.770 }, 00:11:05.770 { 00:11:05.770 "name": "BaseBdev4", 00:11:05.770 "uuid": "c0a4ab88-e694-401c-b920-800a8cef9bf2", 00:11:05.770 "is_configured": true, 00:11:05.770 "data_offset": 0, 00:11:05.770 "data_size": 65536 00:11:05.770 } 00:11:05.770 ] 00:11:05.770 }' 00:11:05.770 17:32:36 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:05.770 17:32:36 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:06.030 17:32:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:06.030 17:32:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:11:06.030 17:32:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:06.030 17:32:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:06.030 17:32:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:06.290 17:32:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:11:06.290 17:32:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:11:06.290 17:32:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:06.290 17:32:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:06.290 [2024-11-27 17:32:37.250695] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:11:06.290 17:32:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:06.290 17:32:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:11:06.290 17:32:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:06.290 17:32:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:06.290 17:32:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:06.290 17:32:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:06.290 17:32:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:06.290 17:32:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:06.290 17:32:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:06.290 17:32:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:06.290 17:32:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:06.290 17:32:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:06.290 17:32:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:06.290 17:32:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:06.290 17:32:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:06.290 17:32:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:06.290 17:32:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:06.290 "name": "Existed_Raid", 00:11:06.290 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:06.290 "strip_size_kb": 0, 00:11:06.290 "state": "configuring", 00:11:06.290 "raid_level": "raid1", 00:11:06.290 "superblock": false, 00:11:06.290 "num_base_bdevs": 4, 00:11:06.290 "num_base_bdevs_discovered": 2, 00:11:06.290 "num_base_bdevs_operational": 4, 00:11:06.290 "base_bdevs_list": [ 00:11:06.290 { 00:11:06.290 "name": "BaseBdev1", 00:11:06.290 "uuid": "f6c97c6b-cf05-4a2f-8744-6665284eafb8", 00:11:06.290 "is_configured": true, 00:11:06.290 "data_offset": 0, 00:11:06.290 "data_size": 65536 00:11:06.290 }, 00:11:06.290 { 00:11:06.290 "name": null, 00:11:06.290 "uuid": "b98327fe-5d91-4798-a1f2-f12a740e5156", 00:11:06.290 "is_configured": false, 00:11:06.290 "data_offset": 0, 00:11:06.290 "data_size": 65536 00:11:06.290 }, 00:11:06.290 { 00:11:06.290 "name": null, 00:11:06.290 "uuid": "1bc274d6-5e32-4c98-8904-57eecd247e0b", 00:11:06.290 "is_configured": false, 00:11:06.290 "data_offset": 0, 00:11:06.290 "data_size": 65536 00:11:06.290 }, 00:11:06.290 { 00:11:06.290 "name": "BaseBdev4", 00:11:06.290 "uuid": "c0a4ab88-e694-401c-b920-800a8cef9bf2", 00:11:06.290 "is_configured": true, 00:11:06.290 "data_offset": 0, 00:11:06.290 "data_size": 65536 00:11:06.290 } 00:11:06.290 ] 00:11:06.290 }' 00:11:06.290 17:32:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:06.290 17:32:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:06.549 17:32:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:06.549 17:32:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:06.549 17:32:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:06.550 17:32:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:11:06.550 17:32:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:06.550 17:32:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:11:06.550 17:32:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:11:06.550 17:32:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:06.550 17:32:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:06.550 [2024-11-27 17:32:37.729943] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:11:06.550 17:32:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:06.550 17:32:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:11:06.550 17:32:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:06.550 17:32:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:06.550 17:32:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:06.550 17:32:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:06.550 17:32:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:06.550 17:32:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:06.550 17:32:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:06.550 17:32:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:06.550 17:32:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:06.812 17:32:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:06.812 17:32:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:06.812 17:32:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:06.812 17:32:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:06.812 17:32:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:06.812 17:32:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:06.812 "name": "Existed_Raid", 00:11:06.812 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:06.812 "strip_size_kb": 0, 00:11:06.812 "state": "configuring", 00:11:06.812 "raid_level": "raid1", 00:11:06.812 "superblock": false, 00:11:06.812 "num_base_bdevs": 4, 00:11:06.812 "num_base_bdevs_discovered": 3, 00:11:06.812 "num_base_bdevs_operational": 4, 00:11:06.812 "base_bdevs_list": [ 00:11:06.812 { 00:11:06.812 "name": "BaseBdev1", 00:11:06.812 "uuid": "f6c97c6b-cf05-4a2f-8744-6665284eafb8", 00:11:06.812 "is_configured": true, 00:11:06.812 "data_offset": 0, 00:11:06.812 "data_size": 65536 00:11:06.812 }, 00:11:06.812 { 00:11:06.812 "name": null, 00:11:06.812 "uuid": "b98327fe-5d91-4798-a1f2-f12a740e5156", 00:11:06.812 "is_configured": false, 00:11:06.812 "data_offset": 0, 00:11:06.812 "data_size": 65536 00:11:06.812 }, 00:11:06.812 { 00:11:06.812 "name": "BaseBdev3", 00:11:06.813 "uuid": "1bc274d6-5e32-4c98-8904-57eecd247e0b", 00:11:06.813 "is_configured": true, 00:11:06.813 "data_offset": 0, 00:11:06.813 "data_size": 65536 00:11:06.813 }, 00:11:06.813 { 00:11:06.813 "name": "BaseBdev4", 00:11:06.813 "uuid": "c0a4ab88-e694-401c-b920-800a8cef9bf2", 00:11:06.813 "is_configured": true, 00:11:06.813 "data_offset": 0, 00:11:06.813 "data_size": 65536 00:11:06.813 } 00:11:06.813 ] 00:11:06.813 }' 00:11:06.813 17:32:37 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:06.813 17:32:37 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:07.076 17:32:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:11:07.076 17:32:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:07.076 17:32:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:07.076 17:32:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:07.076 17:32:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:07.076 17:32:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:11:07.076 17:32:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:11:07.076 17:32:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:07.076 17:32:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:07.076 [2024-11-27 17:32:38.213189] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:11:07.076 17:32:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:07.076 17:32:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:11:07.076 17:32:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:07.076 17:32:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:07.076 17:32:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:07.076 17:32:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:07.076 17:32:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:07.076 17:32:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:07.076 17:32:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:07.076 17:32:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:07.076 17:32:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:07.076 17:32:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:07.076 17:32:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:07.076 17:32:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:07.076 17:32:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:07.076 17:32:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:07.337 17:32:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:07.337 "name": "Existed_Raid", 00:11:07.337 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:07.337 "strip_size_kb": 0, 00:11:07.337 "state": "configuring", 00:11:07.337 "raid_level": "raid1", 00:11:07.337 "superblock": false, 00:11:07.337 "num_base_bdevs": 4, 00:11:07.337 "num_base_bdevs_discovered": 2, 00:11:07.337 "num_base_bdevs_operational": 4, 00:11:07.337 "base_bdevs_list": [ 00:11:07.337 { 00:11:07.337 "name": null, 00:11:07.337 "uuid": "f6c97c6b-cf05-4a2f-8744-6665284eafb8", 00:11:07.337 "is_configured": false, 00:11:07.337 "data_offset": 0, 00:11:07.337 "data_size": 65536 00:11:07.337 }, 00:11:07.337 { 00:11:07.337 "name": null, 00:11:07.337 "uuid": "b98327fe-5d91-4798-a1f2-f12a740e5156", 00:11:07.337 "is_configured": false, 00:11:07.337 "data_offset": 0, 00:11:07.337 "data_size": 65536 00:11:07.337 }, 00:11:07.337 { 00:11:07.337 "name": "BaseBdev3", 00:11:07.337 "uuid": "1bc274d6-5e32-4c98-8904-57eecd247e0b", 00:11:07.337 "is_configured": true, 00:11:07.337 "data_offset": 0, 00:11:07.337 "data_size": 65536 00:11:07.337 }, 00:11:07.337 { 00:11:07.337 "name": "BaseBdev4", 00:11:07.337 "uuid": "c0a4ab88-e694-401c-b920-800a8cef9bf2", 00:11:07.337 "is_configured": true, 00:11:07.337 "data_offset": 0, 00:11:07.337 "data_size": 65536 00:11:07.337 } 00:11:07.337 ] 00:11:07.337 }' 00:11:07.337 17:32:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:07.337 17:32:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:07.597 17:32:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:07.597 17:32:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:11:07.597 17:32:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:07.597 17:32:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:07.597 17:32:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:07.597 17:32:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:11:07.597 17:32:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:11:07.597 17:32:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:07.597 17:32:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:07.597 [2024-11-27 17:32:38.728175] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:11:07.597 17:32:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:07.597 17:32:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:11:07.597 17:32:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:07.597 17:32:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:07.597 17:32:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:07.597 17:32:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:07.597 17:32:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:07.597 17:32:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:07.597 17:32:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:07.597 17:32:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:07.597 17:32:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:07.597 17:32:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:07.597 17:32:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:07.597 17:32:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:07.597 17:32:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:07.597 17:32:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:07.597 17:32:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:07.597 "name": "Existed_Raid", 00:11:07.597 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:07.597 "strip_size_kb": 0, 00:11:07.597 "state": "configuring", 00:11:07.597 "raid_level": "raid1", 00:11:07.597 "superblock": false, 00:11:07.597 "num_base_bdevs": 4, 00:11:07.597 "num_base_bdevs_discovered": 3, 00:11:07.597 "num_base_bdevs_operational": 4, 00:11:07.597 "base_bdevs_list": [ 00:11:07.597 { 00:11:07.597 "name": null, 00:11:07.597 "uuid": "f6c97c6b-cf05-4a2f-8744-6665284eafb8", 00:11:07.597 "is_configured": false, 00:11:07.597 "data_offset": 0, 00:11:07.597 "data_size": 65536 00:11:07.597 }, 00:11:07.597 { 00:11:07.597 "name": "BaseBdev2", 00:11:07.597 "uuid": "b98327fe-5d91-4798-a1f2-f12a740e5156", 00:11:07.597 "is_configured": true, 00:11:07.597 "data_offset": 0, 00:11:07.597 "data_size": 65536 00:11:07.597 }, 00:11:07.597 { 00:11:07.597 "name": "BaseBdev3", 00:11:07.597 "uuid": "1bc274d6-5e32-4c98-8904-57eecd247e0b", 00:11:07.597 "is_configured": true, 00:11:07.597 "data_offset": 0, 00:11:07.597 "data_size": 65536 00:11:07.597 }, 00:11:07.597 { 00:11:07.597 "name": "BaseBdev4", 00:11:07.597 "uuid": "c0a4ab88-e694-401c-b920-800a8cef9bf2", 00:11:07.597 "is_configured": true, 00:11:07.597 "data_offset": 0, 00:11:07.597 "data_size": 65536 00:11:07.597 } 00:11:07.597 ] 00:11:07.597 }' 00:11:07.597 17:32:38 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:07.597 17:32:38 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:08.166 17:32:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:11:08.166 17:32:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:08.166 17:32:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:08.166 17:32:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:08.166 17:32:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:08.166 17:32:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:11:08.166 17:32:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:08.166 17:32:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:08.166 17:32:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:08.166 17:32:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:11:08.166 17:32:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:08.167 17:32:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u f6c97c6b-cf05-4a2f-8744-6665284eafb8 00:11:08.167 17:32:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:08.167 17:32:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:08.167 [2024-11-27 17:32:39.244088] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:11:08.167 [2024-11-27 17:32:39.244205] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:11:08.167 [2024-11-27 17:32:39.244237] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:11:08.167 [2024-11-27 17:32:39.244583] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002a10 00:11:08.167 [2024-11-27 17:32:39.244772] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:11:08.167 [2024-11-27 17:32:39.244813] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001c80 00:11:08.167 [2024-11-27 17:32:39.245037] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:08.167 NewBaseBdev 00:11:08.167 17:32:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:08.167 17:32:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:11:08.167 17:32:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=NewBaseBdev 00:11:08.167 17:32:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:11:08.167 17:32:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@901 -- # local i 00:11:08.167 17:32:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:11:08.167 17:32:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:11:08.167 17:32:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:11:08.167 17:32:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:08.167 17:32:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:08.167 17:32:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:08.167 17:32:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:11:08.167 17:32:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:08.167 17:32:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:08.167 [ 00:11:08.167 { 00:11:08.167 "name": "NewBaseBdev", 00:11:08.167 "aliases": [ 00:11:08.167 "f6c97c6b-cf05-4a2f-8744-6665284eafb8" 00:11:08.167 ], 00:11:08.167 "product_name": "Malloc disk", 00:11:08.167 "block_size": 512, 00:11:08.167 "num_blocks": 65536, 00:11:08.167 "uuid": "f6c97c6b-cf05-4a2f-8744-6665284eafb8", 00:11:08.167 "assigned_rate_limits": { 00:11:08.167 "rw_ios_per_sec": 0, 00:11:08.167 "rw_mbytes_per_sec": 0, 00:11:08.167 "r_mbytes_per_sec": 0, 00:11:08.167 "w_mbytes_per_sec": 0 00:11:08.167 }, 00:11:08.167 "claimed": true, 00:11:08.167 "claim_type": "exclusive_write", 00:11:08.167 "zoned": false, 00:11:08.167 "supported_io_types": { 00:11:08.167 "read": true, 00:11:08.167 "write": true, 00:11:08.167 "unmap": true, 00:11:08.167 "flush": true, 00:11:08.167 "reset": true, 00:11:08.167 "nvme_admin": false, 00:11:08.167 "nvme_io": false, 00:11:08.167 "nvme_io_md": false, 00:11:08.167 "write_zeroes": true, 00:11:08.167 "zcopy": true, 00:11:08.167 "get_zone_info": false, 00:11:08.167 "zone_management": false, 00:11:08.167 "zone_append": false, 00:11:08.167 "compare": false, 00:11:08.167 "compare_and_write": false, 00:11:08.167 "abort": true, 00:11:08.167 "seek_hole": false, 00:11:08.167 "seek_data": false, 00:11:08.167 "copy": true, 00:11:08.167 "nvme_iov_md": false 00:11:08.167 }, 00:11:08.167 "memory_domains": [ 00:11:08.167 { 00:11:08.167 "dma_device_id": "system", 00:11:08.167 "dma_device_type": 1 00:11:08.167 }, 00:11:08.167 { 00:11:08.167 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:08.167 "dma_device_type": 2 00:11:08.167 } 00:11:08.167 ], 00:11:08.167 "driver_specific": {} 00:11:08.167 } 00:11:08.167 ] 00:11:08.167 17:32:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:08.167 17:32:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:11:08.167 17:32:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid1 0 4 00:11:08.167 17:32:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:08.167 17:32:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:08.167 17:32:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:08.167 17:32:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:08.167 17:32:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:08.167 17:32:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:08.167 17:32:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:08.167 17:32:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:08.167 17:32:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:08.167 17:32:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:08.167 17:32:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:08.167 17:32:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:08.167 17:32:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:08.167 17:32:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:08.167 17:32:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:08.167 "name": "Existed_Raid", 00:11:08.167 "uuid": "4829ad26-339b-44a0-a3bf-d814cb326851", 00:11:08.167 "strip_size_kb": 0, 00:11:08.167 "state": "online", 00:11:08.167 "raid_level": "raid1", 00:11:08.167 "superblock": false, 00:11:08.167 "num_base_bdevs": 4, 00:11:08.167 "num_base_bdevs_discovered": 4, 00:11:08.167 "num_base_bdevs_operational": 4, 00:11:08.167 "base_bdevs_list": [ 00:11:08.167 { 00:11:08.167 "name": "NewBaseBdev", 00:11:08.167 "uuid": "f6c97c6b-cf05-4a2f-8744-6665284eafb8", 00:11:08.167 "is_configured": true, 00:11:08.167 "data_offset": 0, 00:11:08.167 "data_size": 65536 00:11:08.167 }, 00:11:08.167 { 00:11:08.167 "name": "BaseBdev2", 00:11:08.167 "uuid": "b98327fe-5d91-4798-a1f2-f12a740e5156", 00:11:08.167 "is_configured": true, 00:11:08.167 "data_offset": 0, 00:11:08.167 "data_size": 65536 00:11:08.167 }, 00:11:08.167 { 00:11:08.167 "name": "BaseBdev3", 00:11:08.167 "uuid": "1bc274d6-5e32-4c98-8904-57eecd247e0b", 00:11:08.167 "is_configured": true, 00:11:08.167 "data_offset": 0, 00:11:08.167 "data_size": 65536 00:11:08.167 }, 00:11:08.167 { 00:11:08.167 "name": "BaseBdev4", 00:11:08.167 "uuid": "c0a4ab88-e694-401c-b920-800a8cef9bf2", 00:11:08.167 "is_configured": true, 00:11:08.167 "data_offset": 0, 00:11:08.167 "data_size": 65536 00:11:08.167 } 00:11:08.167 ] 00:11:08.167 }' 00:11:08.167 17:32:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:08.167 17:32:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:08.736 17:32:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:11:08.736 17:32:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:11:08.736 17:32:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:11:08.736 17:32:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:11:08.736 17:32:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:11:08.736 17:32:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:11:08.736 17:32:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:11:08.736 17:32:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:11:08.736 17:32:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:08.736 17:32:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:08.736 [2024-11-27 17:32:39.767549] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:11:08.736 17:32:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:08.736 17:32:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:11:08.736 "name": "Existed_Raid", 00:11:08.736 "aliases": [ 00:11:08.736 "4829ad26-339b-44a0-a3bf-d814cb326851" 00:11:08.736 ], 00:11:08.736 "product_name": "Raid Volume", 00:11:08.736 "block_size": 512, 00:11:08.736 "num_blocks": 65536, 00:11:08.736 "uuid": "4829ad26-339b-44a0-a3bf-d814cb326851", 00:11:08.736 "assigned_rate_limits": { 00:11:08.736 "rw_ios_per_sec": 0, 00:11:08.736 "rw_mbytes_per_sec": 0, 00:11:08.736 "r_mbytes_per_sec": 0, 00:11:08.736 "w_mbytes_per_sec": 0 00:11:08.736 }, 00:11:08.736 "claimed": false, 00:11:08.736 "zoned": false, 00:11:08.736 "supported_io_types": { 00:11:08.736 "read": true, 00:11:08.736 "write": true, 00:11:08.736 "unmap": false, 00:11:08.736 "flush": false, 00:11:08.736 "reset": true, 00:11:08.736 "nvme_admin": false, 00:11:08.737 "nvme_io": false, 00:11:08.737 "nvme_io_md": false, 00:11:08.737 "write_zeroes": true, 00:11:08.737 "zcopy": false, 00:11:08.737 "get_zone_info": false, 00:11:08.737 "zone_management": false, 00:11:08.737 "zone_append": false, 00:11:08.737 "compare": false, 00:11:08.737 "compare_and_write": false, 00:11:08.737 "abort": false, 00:11:08.737 "seek_hole": false, 00:11:08.737 "seek_data": false, 00:11:08.737 "copy": false, 00:11:08.737 "nvme_iov_md": false 00:11:08.737 }, 00:11:08.737 "memory_domains": [ 00:11:08.737 { 00:11:08.737 "dma_device_id": "system", 00:11:08.737 "dma_device_type": 1 00:11:08.737 }, 00:11:08.737 { 00:11:08.737 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:08.737 "dma_device_type": 2 00:11:08.737 }, 00:11:08.737 { 00:11:08.737 "dma_device_id": "system", 00:11:08.737 "dma_device_type": 1 00:11:08.737 }, 00:11:08.737 { 00:11:08.737 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:08.737 "dma_device_type": 2 00:11:08.737 }, 00:11:08.737 { 00:11:08.737 "dma_device_id": "system", 00:11:08.737 "dma_device_type": 1 00:11:08.737 }, 00:11:08.737 { 00:11:08.737 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:08.737 "dma_device_type": 2 00:11:08.737 }, 00:11:08.737 { 00:11:08.737 "dma_device_id": "system", 00:11:08.737 "dma_device_type": 1 00:11:08.737 }, 00:11:08.737 { 00:11:08.737 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:08.737 "dma_device_type": 2 00:11:08.737 } 00:11:08.737 ], 00:11:08.737 "driver_specific": { 00:11:08.737 "raid": { 00:11:08.737 "uuid": "4829ad26-339b-44a0-a3bf-d814cb326851", 00:11:08.737 "strip_size_kb": 0, 00:11:08.737 "state": "online", 00:11:08.737 "raid_level": "raid1", 00:11:08.737 "superblock": false, 00:11:08.737 "num_base_bdevs": 4, 00:11:08.737 "num_base_bdevs_discovered": 4, 00:11:08.737 "num_base_bdevs_operational": 4, 00:11:08.737 "base_bdevs_list": [ 00:11:08.737 { 00:11:08.737 "name": "NewBaseBdev", 00:11:08.737 "uuid": "f6c97c6b-cf05-4a2f-8744-6665284eafb8", 00:11:08.737 "is_configured": true, 00:11:08.737 "data_offset": 0, 00:11:08.737 "data_size": 65536 00:11:08.737 }, 00:11:08.737 { 00:11:08.737 "name": "BaseBdev2", 00:11:08.737 "uuid": "b98327fe-5d91-4798-a1f2-f12a740e5156", 00:11:08.737 "is_configured": true, 00:11:08.737 "data_offset": 0, 00:11:08.737 "data_size": 65536 00:11:08.737 }, 00:11:08.737 { 00:11:08.737 "name": "BaseBdev3", 00:11:08.737 "uuid": "1bc274d6-5e32-4c98-8904-57eecd247e0b", 00:11:08.737 "is_configured": true, 00:11:08.737 "data_offset": 0, 00:11:08.737 "data_size": 65536 00:11:08.737 }, 00:11:08.737 { 00:11:08.737 "name": "BaseBdev4", 00:11:08.737 "uuid": "c0a4ab88-e694-401c-b920-800a8cef9bf2", 00:11:08.737 "is_configured": true, 00:11:08.737 "data_offset": 0, 00:11:08.737 "data_size": 65536 00:11:08.737 } 00:11:08.737 ] 00:11:08.737 } 00:11:08.737 } 00:11:08.737 }' 00:11:08.737 17:32:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:11:08.737 17:32:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:11:08.737 BaseBdev2 00:11:08.737 BaseBdev3 00:11:08.737 BaseBdev4' 00:11:08.737 17:32:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:08.737 17:32:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:11:08.737 17:32:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:08.737 17:32:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:08.737 17:32:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:11:08.737 17:32:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:08.737 17:32:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:08.737 17:32:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:08.737 17:32:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:08.737 17:32:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:08.737 17:32:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:08.997 17:32:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:11:08.997 17:32:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:08.997 17:32:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:08.997 17:32:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:08.997 17:32:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:08.997 17:32:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:08.997 17:32:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:08.997 17:32:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:08.997 17:32:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:11:08.997 17:32:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:08.997 17:32:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:08.997 17:32:39 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:08.997 17:32:39 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:08.997 17:32:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:08.997 17:32:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:08.997 17:32:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:08.997 17:32:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:11:08.997 17:32:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:08.997 17:32:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:08.997 17:32:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:08.997 17:32:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:08.997 17:32:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:08.997 17:32:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:08.997 17:32:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:11:08.997 17:32:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:08.997 17:32:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:08.997 [2024-11-27 17:32:40.074728] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:11:08.997 [2024-11-27 17:32:40.074795] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:11:08.997 [2024-11-27 17:32:40.074956] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:11:08.997 [2024-11-27 17:32:40.075321] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:11:08.997 [2024-11-27 17:32:40.075380] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name Existed_Raid, state offline 00:11:08.997 17:32:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:08.997 17:32:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 83844 00:11:08.997 17:32:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@950 -- # '[' -z 83844 ']' 00:11:08.997 17:32:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@954 -- # kill -0 83844 00:11:08.997 17:32:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@955 -- # uname 00:11:08.997 17:32:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:11:08.997 17:32:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 83844 00:11:08.997 killing process with pid 83844 00:11:08.997 17:32:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:11:08.997 17:32:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:11:08.998 17:32:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 83844' 00:11:08.998 17:32:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@969 -- # kill 83844 00:11:08.998 [2024-11-27 17:32:40.122946] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:11:08.998 17:32:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@974 -- # wait 83844 00:11:09.257 [2024-11-27 17:32:40.198661] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:11:09.518 17:32:40 bdev_raid.raid_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:11:09.518 00:11:09.518 real 0m9.772s 00:11:09.518 user 0m16.287s 00:11:09.518 sys 0m2.134s 00:11:09.518 ************************************ 00:11:09.518 END TEST raid_state_function_test 00:11:09.518 ************************************ 00:11:09.518 17:32:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:11:09.518 17:32:40 bdev_raid.raid_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:11:09.518 17:32:40 bdev_raid -- bdev/bdev_raid.sh@969 -- # run_test raid_state_function_test_sb raid_state_function_test raid1 4 true 00:11:09.518 17:32:40 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:11:09.518 17:32:40 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:11:09.518 17:32:40 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:11:09.518 ************************************ 00:11:09.518 START TEST raid_state_function_test_sb 00:11:09.518 ************************************ 00:11:09.518 17:32:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1125 -- # raid_state_function_test raid1 4 true 00:11:09.518 17:32:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:11:09.518 17:32:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=4 00:11:09.518 17:32:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:11:09.518 17:32:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:11:09.518 17:32:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:11:09.518 17:32:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:11:09.518 17:32:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:11:09.518 17:32:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:11:09.518 17:32:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:11:09.518 17:32:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:11:09.518 17:32:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:11:09.518 17:32:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:11:09.518 17:32:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:11:09.518 17:32:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:11:09.518 17:32:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:11:09.518 17:32:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev4 00:11:09.518 17:32:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:11:09.518 17:32:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:11:09.518 17:32:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:11:09.518 17:32:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:11:09.518 17:32:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:11:09.518 17:32:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:11:09.518 17:32:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:11:09.518 17:32:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:11:09.518 17:32:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:11:09.518 17:32:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:11:09.518 17:32:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:11:09.518 17:32:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:11:09.518 Process raid pid: 84493 00:11:09.518 17:32:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=84493 00:11:09.518 17:32:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:11:09.518 17:32:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 84493' 00:11:09.518 17:32:40 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 84493 00:11:09.518 17:32:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@831 -- # '[' -z 84493 ']' 00:11:09.518 17:32:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:11:09.518 17:32:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@836 -- # local max_retries=100 00:11:09.518 17:32:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:11:09.518 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:11:09.518 17:32:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@840 -- # xtrace_disable 00:11:09.518 17:32:40 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:09.779 [2024-11-27 17:32:40.742977] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:11:09.779 [2024-11-27 17:32:40.743206] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:11:09.779 [2024-11-27 17:32:40.889966] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:11:09.779 [2024-11-27 17:32:40.958880] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:11:10.040 [2024-11-27 17:32:41.035477] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:10.040 [2024-11-27 17:32:41.035596] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:10.611 17:32:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:11:10.611 17:32:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@864 -- # return 0 00:11:10.611 17:32:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:11:10.611 17:32:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:10.611 17:32:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:10.611 [2024-11-27 17:32:41.570987] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:11:10.611 [2024-11-27 17:32:41.571041] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:11:10.611 [2024-11-27 17:32:41.571055] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:11:10.611 [2024-11-27 17:32:41.571066] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:11:10.611 [2024-11-27 17:32:41.571073] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:11:10.611 [2024-11-27 17:32:41.571087] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:11:10.611 [2024-11-27 17:32:41.571093] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:11:10.611 [2024-11-27 17:32:41.571102] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:11:10.611 17:32:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:10.611 17:32:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:11:10.611 17:32:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:10.611 17:32:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:10.611 17:32:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:10.611 17:32:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:10.611 17:32:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:10.611 17:32:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:10.611 17:32:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:10.611 17:32:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:10.611 17:32:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:10.611 17:32:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:10.611 17:32:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:10.611 17:32:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:10.611 17:32:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:10.611 17:32:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:10.611 17:32:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:10.611 "name": "Existed_Raid", 00:11:10.611 "uuid": "1d73b2cc-fcbf-4a81-a0b9-24c5a70b1a45", 00:11:10.611 "strip_size_kb": 0, 00:11:10.611 "state": "configuring", 00:11:10.611 "raid_level": "raid1", 00:11:10.611 "superblock": true, 00:11:10.611 "num_base_bdevs": 4, 00:11:10.611 "num_base_bdevs_discovered": 0, 00:11:10.611 "num_base_bdevs_operational": 4, 00:11:10.611 "base_bdevs_list": [ 00:11:10.611 { 00:11:10.611 "name": "BaseBdev1", 00:11:10.611 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:10.611 "is_configured": false, 00:11:10.611 "data_offset": 0, 00:11:10.611 "data_size": 0 00:11:10.611 }, 00:11:10.611 { 00:11:10.611 "name": "BaseBdev2", 00:11:10.611 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:10.611 "is_configured": false, 00:11:10.611 "data_offset": 0, 00:11:10.611 "data_size": 0 00:11:10.611 }, 00:11:10.611 { 00:11:10.611 "name": "BaseBdev3", 00:11:10.611 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:10.611 "is_configured": false, 00:11:10.611 "data_offset": 0, 00:11:10.611 "data_size": 0 00:11:10.611 }, 00:11:10.611 { 00:11:10.611 "name": "BaseBdev4", 00:11:10.611 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:10.611 "is_configured": false, 00:11:10.611 "data_offset": 0, 00:11:10.612 "data_size": 0 00:11:10.612 } 00:11:10.612 ] 00:11:10.612 }' 00:11:10.612 17:32:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:10.612 17:32:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:10.872 17:32:41 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:11:10.872 17:32:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:10.872 17:32:41 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:10.872 [2024-11-27 17:32:41.998127] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:11:10.872 [2024-11-27 17:32:41.998226] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:11:10.872 17:32:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:10.872 17:32:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:11:10.872 17:32:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:10.872 17:32:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:10.872 [2024-11-27 17:32:42.010158] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:11:10.872 [2024-11-27 17:32:42.010232] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:11:10.872 [2024-11-27 17:32:42.010260] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:11:10.872 [2024-11-27 17:32:42.010284] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:11:10.872 [2024-11-27 17:32:42.010311] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:11:10.872 [2024-11-27 17:32:42.010339] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:11:10.872 [2024-11-27 17:32:42.010372] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:11:10.872 [2024-11-27 17:32:42.010395] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:11:10.872 17:32:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:10.872 17:32:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:11:10.872 17:32:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:10.872 17:32:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:10.872 [2024-11-27 17:32:42.037262] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:11:10.872 BaseBdev1 00:11:10.872 17:32:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:10.872 17:32:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:11:10.872 17:32:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:11:10.872 17:32:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:11:10.872 17:32:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:11:10.872 17:32:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:11:10.872 17:32:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:11:10.872 17:32:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:11:10.872 17:32:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:10.872 17:32:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:10.872 17:32:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:10.872 17:32:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:11:10.872 17:32:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:10.872 17:32:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:11.132 [ 00:11:11.132 { 00:11:11.132 "name": "BaseBdev1", 00:11:11.132 "aliases": [ 00:11:11.132 "2352465b-833d-48ae-8331-1064c38dfca5" 00:11:11.132 ], 00:11:11.132 "product_name": "Malloc disk", 00:11:11.132 "block_size": 512, 00:11:11.132 "num_blocks": 65536, 00:11:11.132 "uuid": "2352465b-833d-48ae-8331-1064c38dfca5", 00:11:11.132 "assigned_rate_limits": { 00:11:11.132 "rw_ios_per_sec": 0, 00:11:11.132 "rw_mbytes_per_sec": 0, 00:11:11.132 "r_mbytes_per_sec": 0, 00:11:11.132 "w_mbytes_per_sec": 0 00:11:11.132 }, 00:11:11.132 "claimed": true, 00:11:11.132 "claim_type": "exclusive_write", 00:11:11.132 "zoned": false, 00:11:11.132 "supported_io_types": { 00:11:11.132 "read": true, 00:11:11.132 "write": true, 00:11:11.132 "unmap": true, 00:11:11.132 "flush": true, 00:11:11.132 "reset": true, 00:11:11.132 "nvme_admin": false, 00:11:11.132 "nvme_io": false, 00:11:11.132 "nvme_io_md": false, 00:11:11.132 "write_zeroes": true, 00:11:11.132 "zcopy": true, 00:11:11.132 "get_zone_info": false, 00:11:11.132 "zone_management": false, 00:11:11.132 "zone_append": false, 00:11:11.132 "compare": false, 00:11:11.132 "compare_and_write": false, 00:11:11.132 "abort": true, 00:11:11.132 "seek_hole": false, 00:11:11.132 "seek_data": false, 00:11:11.132 "copy": true, 00:11:11.132 "nvme_iov_md": false 00:11:11.132 }, 00:11:11.132 "memory_domains": [ 00:11:11.132 { 00:11:11.132 "dma_device_id": "system", 00:11:11.132 "dma_device_type": 1 00:11:11.132 }, 00:11:11.132 { 00:11:11.132 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:11.132 "dma_device_type": 2 00:11:11.132 } 00:11:11.132 ], 00:11:11.132 "driver_specific": {} 00:11:11.132 } 00:11:11.132 ] 00:11:11.132 17:32:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:11.132 17:32:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:11:11.132 17:32:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:11:11.132 17:32:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:11.132 17:32:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:11.132 17:32:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:11.132 17:32:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:11.132 17:32:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:11.132 17:32:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:11.132 17:32:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:11.132 17:32:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:11.132 17:32:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:11.132 17:32:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:11.132 17:32:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:11.132 17:32:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:11.132 17:32:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:11.132 17:32:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:11.132 17:32:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:11.132 "name": "Existed_Raid", 00:11:11.132 "uuid": "5559f35b-dfaf-46c7-9197-1ec09770f879", 00:11:11.132 "strip_size_kb": 0, 00:11:11.132 "state": "configuring", 00:11:11.132 "raid_level": "raid1", 00:11:11.132 "superblock": true, 00:11:11.132 "num_base_bdevs": 4, 00:11:11.132 "num_base_bdevs_discovered": 1, 00:11:11.132 "num_base_bdevs_operational": 4, 00:11:11.132 "base_bdevs_list": [ 00:11:11.132 { 00:11:11.132 "name": "BaseBdev1", 00:11:11.133 "uuid": "2352465b-833d-48ae-8331-1064c38dfca5", 00:11:11.133 "is_configured": true, 00:11:11.133 "data_offset": 2048, 00:11:11.133 "data_size": 63488 00:11:11.133 }, 00:11:11.133 { 00:11:11.133 "name": "BaseBdev2", 00:11:11.133 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:11.133 "is_configured": false, 00:11:11.133 "data_offset": 0, 00:11:11.133 "data_size": 0 00:11:11.133 }, 00:11:11.133 { 00:11:11.133 "name": "BaseBdev3", 00:11:11.133 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:11.133 "is_configured": false, 00:11:11.133 "data_offset": 0, 00:11:11.133 "data_size": 0 00:11:11.133 }, 00:11:11.133 { 00:11:11.133 "name": "BaseBdev4", 00:11:11.133 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:11.133 "is_configured": false, 00:11:11.133 "data_offset": 0, 00:11:11.133 "data_size": 0 00:11:11.133 } 00:11:11.133 ] 00:11:11.133 }' 00:11:11.133 17:32:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:11.133 17:32:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:11.393 17:32:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:11:11.393 17:32:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:11.393 17:32:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:11.393 [2024-11-27 17:32:42.500467] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:11:11.393 [2024-11-27 17:32:42.500521] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:11:11.393 17:32:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:11.393 17:32:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:11:11.393 17:32:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:11.393 17:32:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:11.393 [2024-11-27 17:32:42.512513] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:11:11.393 [2024-11-27 17:32:42.514690] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:11:11.393 [2024-11-27 17:32:42.514777] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:11:11.393 [2024-11-27 17:32:42.514820] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:11:11.393 [2024-11-27 17:32:42.514843] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:11:11.393 [2024-11-27 17:32:42.514861] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:11:11.393 [2024-11-27 17:32:42.514881] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:11:11.393 17:32:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:11.393 17:32:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:11:11.393 17:32:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:11:11.393 17:32:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:11:11.393 17:32:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:11.393 17:32:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:11.393 17:32:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:11.394 17:32:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:11.394 17:32:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:11.394 17:32:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:11.394 17:32:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:11.394 17:32:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:11.394 17:32:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:11.394 17:32:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:11.394 17:32:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:11.394 17:32:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:11.394 17:32:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:11.394 17:32:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:11.394 17:32:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:11.394 "name": "Existed_Raid", 00:11:11.394 "uuid": "d2f0bcd7-fdf9-4827-a7f7-33139962b64b", 00:11:11.394 "strip_size_kb": 0, 00:11:11.394 "state": "configuring", 00:11:11.394 "raid_level": "raid1", 00:11:11.394 "superblock": true, 00:11:11.394 "num_base_bdevs": 4, 00:11:11.394 "num_base_bdevs_discovered": 1, 00:11:11.394 "num_base_bdevs_operational": 4, 00:11:11.394 "base_bdevs_list": [ 00:11:11.394 { 00:11:11.394 "name": "BaseBdev1", 00:11:11.394 "uuid": "2352465b-833d-48ae-8331-1064c38dfca5", 00:11:11.394 "is_configured": true, 00:11:11.394 "data_offset": 2048, 00:11:11.394 "data_size": 63488 00:11:11.394 }, 00:11:11.394 { 00:11:11.394 "name": "BaseBdev2", 00:11:11.394 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:11.394 "is_configured": false, 00:11:11.394 "data_offset": 0, 00:11:11.394 "data_size": 0 00:11:11.394 }, 00:11:11.394 { 00:11:11.394 "name": "BaseBdev3", 00:11:11.394 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:11.394 "is_configured": false, 00:11:11.394 "data_offset": 0, 00:11:11.394 "data_size": 0 00:11:11.394 }, 00:11:11.394 { 00:11:11.394 "name": "BaseBdev4", 00:11:11.394 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:11.394 "is_configured": false, 00:11:11.394 "data_offset": 0, 00:11:11.394 "data_size": 0 00:11:11.394 } 00:11:11.394 ] 00:11:11.394 }' 00:11:11.394 17:32:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:11.394 17:32:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:11.966 17:32:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:11:11.966 17:32:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:11.966 17:32:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:11.966 [2024-11-27 17:32:42.962589] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:11:11.966 BaseBdev2 00:11:11.966 17:32:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:11.966 17:32:42 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:11:11.966 17:32:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:11:11.966 17:32:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:11:11.966 17:32:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:11:11.966 17:32:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:11:11.966 17:32:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:11:11.966 17:32:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:11:11.966 17:32:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:11.966 17:32:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:11.966 17:32:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:11.966 17:32:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:11:11.966 17:32:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:11.966 17:32:42 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:11.966 [ 00:11:11.966 { 00:11:11.966 "name": "BaseBdev2", 00:11:11.966 "aliases": [ 00:11:11.966 "53695494-2ae1-49d6-a90a-e0f2ecf5e49b" 00:11:11.966 ], 00:11:11.966 "product_name": "Malloc disk", 00:11:11.966 "block_size": 512, 00:11:11.966 "num_blocks": 65536, 00:11:11.966 "uuid": "53695494-2ae1-49d6-a90a-e0f2ecf5e49b", 00:11:11.966 "assigned_rate_limits": { 00:11:11.966 "rw_ios_per_sec": 0, 00:11:11.966 "rw_mbytes_per_sec": 0, 00:11:11.966 "r_mbytes_per_sec": 0, 00:11:11.966 "w_mbytes_per_sec": 0 00:11:11.966 }, 00:11:11.966 "claimed": true, 00:11:11.966 "claim_type": "exclusive_write", 00:11:11.966 "zoned": false, 00:11:11.966 "supported_io_types": { 00:11:11.966 "read": true, 00:11:11.966 "write": true, 00:11:11.966 "unmap": true, 00:11:11.966 "flush": true, 00:11:11.966 "reset": true, 00:11:11.966 "nvme_admin": false, 00:11:11.966 "nvme_io": false, 00:11:11.966 "nvme_io_md": false, 00:11:11.966 "write_zeroes": true, 00:11:11.966 "zcopy": true, 00:11:11.966 "get_zone_info": false, 00:11:11.966 "zone_management": false, 00:11:11.966 "zone_append": false, 00:11:11.966 "compare": false, 00:11:11.966 "compare_and_write": false, 00:11:11.966 "abort": true, 00:11:11.966 "seek_hole": false, 00:11:11.966 "seek_data": false, 00:11:11.966 "copy": true, 00:11:11.966 "nvme_iov_md": false 00:11:11.966 }, 00:11:11.966 "memory_domains": [ 00:11:11.966 { 00:11:11.966 "dma_device_id": "system", 00:11:11.966 "dma_device_type": 1 00:11:11.966 }, 00:11:11.966 { 00:11:11.966 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:11.966 "dma_device_type": 2 00:11:11.966 } 00:11:11.966 ], 00:11:11.966 "driver_specific": {} 00:11:11.966 } 00:11:11.966 ] 00:11:11.966 17:32:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:11.966 17:32:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:11:11.966 17:32:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:11:11.966 17:32:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:11:11.966 17:32:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:11:11.966 17:32:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:11.966 17:32:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:11.966 17:32:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:11.966 17:32:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:11.966 17:32:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:11.966 17:32:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:11.966 17:32:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:11.966 17:32:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:11.966 17:32:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:11.966 17:32:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:11.966 17:32:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:11.966 17:32:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:11.966 17:32:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:11.966 17:32:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:11.966 17:32:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:11.966 "name": "Existed_Raid", 00:11:11.966 "uuid": "d2f0bcd7-fdf9-4827-a7f7-33139962b64b", 00:11:11.966 "strip_size_kb": 0, 00:11:11.966 "state": "configuring", 00:11:11.966 "raid_level": "raid1", 00:11:11.966 "superblock": true, 00:11:11.966 "num_base_bdevs": 4, 00:11:11.966 "num_base_bdevs_discovered": 2, 00:11:11.966 "num_base_bdevs_operational": 4, 00:11:11.966 "base_bdevs_list": [ 00:11:11.966 { 00:11:11.966 "name": "BaseBdev1", 00:11:11.966 "uuid": "2352465b-833d-48ae-8331-1064c38dfca5", 00:11:11.966 "is_configured": true, 00:11:11.966 "data_offset": 2048, 00:11:11.966 "data_size": 63488 00:11:11.966 }, 00:11:11.966 { 00:11:11.966 "name": "BaseBdev2", 00:11:11.966 "uuid": "53695494-2ae1-49d6-a90a-e0f2ecf5e49b", 00:11:11.966 "is_configured": true, 00:11:11.966 "data_offset": 2048, 00:11:11.966 "data_size": 63488 00:11:11.966 }, 00:11:11.966 { 00:11:11.966 "name": "BaseBdev3", 00:11:11.966 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:11.966 "is_configured": false, 00:11:11.966 "data_offset": 0, 00:11:11.966 "data_size": 0 00:11:11.966 }, 00:11:11.966 { 00:11:11.966 "name": "BaseBdev4", 00:11:11.966 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:11.966 "is_configured": false, 00:11:11.966 "data_offset": 0, 00:11:11.966 "data_size": 0 00:11:11.966 } 00:11:11.966 ] 00:11:11.966 }' 00:11:11.966 17:32:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:11.966 17:32:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:12.227 17:32:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:11:12.227 17:32:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:12.227 17:32:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:12.492 [2024-11-27 17:32:43.422573] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:11:12.492 BaseBdev3 00:11:12.492 17:32:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:12.492 17:32:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:11:12.492 17:32:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:11:12.492 17:32:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:11:12.492 17:32:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:11:12.492 17:32:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:11:12.492 17:32:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:11:12.492 17:32:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:11:12.492 17:32:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:12.492 17:32:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:12.492 17:32:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:12.492 17:32:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:11:12.492 17:32:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:12.492 17:32:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:12.492 [ 00:11:12.492 { 00:11:12.492 "name": "BaseBdev3", 00:11:12.492 "aliases": [ 00:11:12.492 "13eea6dd-d2c6-45a3-bedd-84ef82407cba" 00:11:12.492 ], 00:11:12.492 "product_name": "Malloc disk", 00:11:12.492 "block_size": 512, 00:11:12.492 "num_blocks": 65536, 00:11:12.492 "uuid": "13eea6dd-d2c6-45a3-bedd-84ef82407cba", 00:11:12.492 "assigned_rate_limits": { 00:11:12.492 "rw_ios_per_sec": 0, 00:11:12.492 "rw_mbytes_per_sec": 0, 00:11:12.492 "r_mbytes_per_sec": 0, 00:11:12.492 "w_mbytes_per_sec": 0 00:11:12.492 }, 00:11:12.492 "claimed": true, 00:11:12.492 "claim_type": "exclusive_write", 00:11:12.492 "zoned": false, 00:11:12.492 "supported_io_types": { 00:11:12.492 "read": true, 00:11:12.492 "write": true, 00:11:12.492 "unmap": true, 00:11:12.492 "flush": true, 00:11:12.492 "reset": true, 00:11:12.492 "nvme_admin": false, 00:11:12.492 "nvme_io": false, 00:11:12.492 "nvme_io_md": false, 00:11:12.492 "write_zeroes": true, 00:11:12.492 "zcopy": true, 00:11:12.492 "get_zone_info": false, 00:11:12.492 "zone_management": false, 00:11:12.493 "zone_append": false, 00:11:12.493 "compare": false, 00:11:12.493 "compare_and_write": false, 00:11:12.493 "abort": true, 00:11:12.493 "seek_hole": false, 00:11:12.493 "seek_data": false, 00:11:12.493 "copy": true, 00:11:12.493 "nvme_iov_md": false 00:11:12.493 }, 00:11:12.493 "memory_domains": [ 00:11:12.493 { 00:11:12.493 "dma_device_id": "system", 00:11:12.493 "dma_device_type": 1 00:11:12.493 }, 00:11:12.493 { 00:11:12.493 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:12.493 "dma_device_type": 2 00:11:12.493 } 00:11:12.493 ], 00:11:12.493 "driver_specific": {} 00:11:12.493 } 00:11:12.493 ] 00:11:12.493 17:32:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:12.493 17:32:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:11:12.493 17:32:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:11:12.493 17:32:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:11:12.493 17:32:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:11:12.493 17:32:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:12.493 17:32:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:12.493 17:32:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:12.493 17:32:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:12.493 17:32:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:12.493 17:32:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:12.493 17:32:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:12.493 17:32:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:12.493 17:32:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:12.493 17:32:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:12.493 17:32:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:12.493 17:32:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:12.493 17:32:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:12.493 17:32:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:12.493 17:32:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:12.493 "name": "Existed_Raid", 00:11:12.493 "uuid": "d2f0bcd7-fdf9-4827-a7f7-33139962b64b", 00:11:12.493 "strip_size_kb": 0, 00:11:12.493 "state": "configuring", 00:11:12.493 "raid_level": "raid1", 00:11:12.493 "superblock": true, 00:11:12.493 "num_base_bdevs": 4, 00:11:12.493 "num_base_bdevs_discovered": 3, 00:11:12.493 "num_base_bdevs_operational": 4, 00:11:12.493 "base_bdevs_list": [ 00:11:12.493 { 00:11:12.493 "name": "BaseBdev1", 00:11:12.493 "uuid": "2352465b-833d-48ae-8331-1064c38dfca5", 00:11:12.493 "is_configured": true, 00:11:12.493 "data_offset": 2048, 00:11:12.493 "data_size": 63488 00:11:12.493 }, 00:11:12.493 { 00:11:12.493 "name": "BaseBdev2", 00:11:12.493 "uuid": "53695494-2ae1-49d6-a90a-e0f2ecf5e49b", 00:11:12.493 "is_configured": true, 00:11:12.493 "data_offset": 2048, 00:11:12.493 "data_size": 63488 00:11:12.493 }, 00:11:12.493 { 00:11:12.493 "name": "BaseBdev3", 00:11:12.493 "uuid": "13eea6dd-d2c6-45a3-bedd-84ef82407cba", 00:11:12.493 "is_configured": true, 00:11:12.493 "data_offset": 2048, 00:11:12.493 "data_size": 63488 00:11:12.493 }, 00:11:12.493 { 00:11:12.493 "name": "BaseBdev4", 00:11:12.493 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:12.493 "is_configured": false, 00:11:12.493 "data_offset": 0, 00:11:12.493 "data_size": 0 00:11:12.493 } 00:11:12.493 ] 00:11:12.493 }' 00:11:12.493 17:32:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:12.493 17:32:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:12.757 17:32:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:11:12.757 17:32:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:12.757 17:32:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:12.757 [2024-11-27 17:32:43.906622] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:11:12.757 [2024-11-27 17:32:43.906952] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:11:12.757 [2024-11-27 17:32:43.906972] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:11:12.757 [2024-11-27 17:32:43.907297] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:11:12.757 BaseBdev4 00:11:12.757 [2024-11-27 17:32:43.907475] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:11:12.757 [2024-11-27 17:32:43.907490] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:11:12.757 [2024-11-27 17:32:43.907630] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:12.757 17:32:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:12.757 17:32:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev4 00:11:12.757 17:32:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev4 00:11:12.757 17:32:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:11:12.757 17:32:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:11:12.757 17:32:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:11:12.757 17:32:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:11:12.757 17:32:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:11:12.757 17:32:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:12.757 17:32:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:12.757 17:32:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:12.757 17:32:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:11:12.757 17:32:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:12.757 17:32:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:12.757 [ 00:11:12.757 { 00:11:12.757 "name": "BaseBdev4", 00:11:12.757 "aliases": [ 00:11:12.757 "38891f9a-26e6-4eb7-aec8-ad9bfbd93957" 00:11:12.757 ], 00:11:12.757 "product_name": "Malloc disk", 00:11:12.757 "block_size": 512, 00:11:12.757 "num_blocks": 65536, 00:11:12.757 "uuid": "38891f9a-26e6-4eb7-aec8-ad9bfbd93957", 00:11:12.757 "assigned_rate_limits": { 00:11:12.757 "rw_ios_per_sec": 0, 00:11:12.757 "rw_mbytes_per_sec": 0, 00:11:12.757 "r_mbytes_per_sec": 0, 00:11:12.757 "w_mbytes_per_sec": 0 00:11:12.757 }, 00:11:12.757 "claimed": true, 00:11:12.757 "claim_type": "exclusive_write", 00:11:12.757 "zoned": false, 00:11:12.757 "supported_io_types": { 00:11:12.757 "read": true, 00:11:12.757 "write": true, 00:11:12.757 "unmap": true, 00:11:12.757 "flush": true, 00:11:12.757 "reset": true, 00:11:12.757 "nvme_admin": false, 00:11:12.757 "nvme_io": false, 00:11:12.757 "nvme_io_md": false, 00:11:12.757 "write_zeroes": true, 00:11:12.757 "zcopy": true, 00:11:12.757 "get_zone_info": false, 00:11:12.757 "zone_management": false, 00:11:12.757 "zone_append": false, 00:11:12.757 "compare": false, 00:11:12.757 "compare_and_write": false, 00:11:12.757 "abort": true, 00:11:12.757 "seek_hole": false, 00:11:12.757 "seek_data": false, 00:11:12.757 "copy": true, 00:11:12.757 "nvme_iov_md": false 00:11:12.757 }, 00:11:12.757 "memory_domains": [ 00:11:12.757 { 00:11:12.757 "dma_device_id": "system", 00:11:12.757 "dma_device_type": 1 00:11:12.757 }, 00:11:12.757 { 00:11:12.757 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:12.757 "dma_device_type": 2 00:11:12.757 } 00:11:12.757 ], 00:11:12.757 "driver_specific": {} 00:11:12.757 } 00:11:12.757 ] 00:11:12.757 17:32:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:12.757 17:32:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:11:12.757 17:32:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:11:12.757 17:32:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:11:12.757 17:32:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 4 00:11:13.017 17:32:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:13.017 17:32:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:13.017 17:32:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:13.017 17:32:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:13.017 17:32:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:13.017 17:32:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:13.017 17:32:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:13.017 17:32:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:13.017 17:32:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:13.017 17:32:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:13.017 17:32:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:13.017 17:32:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:13.017 17:32:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:13.017 17:32:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:13.017 17:32:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:13.017 "name": "Existed_Raid", 00:11:13.017 "uuid": "d2f0bcd7-fdf9-4827-a7f7-33139962b64b", 00:11:13.017 "strip_size_kb": 0, 00:11:13.017 "state": "online", 00:11:13.017 "raid_level": "raid1", 00:11:13.017 "superblock": true, 00:11:13.017 "num_base_bdevs": 4, 00:11:13.017 "num_base_bdevs_discovered": 4, 00:11:13.017 "num_base_bdevs_operational": 4, 00:11:13.017 "base_bdevs_list": [ 00:11:13.017 { 00:11:13.017 "name": "BaseBdev1", 00:11:13.017 "uuid": "2352465b-833d-48ae-8331-1064c38dfca5", 00:11:13.017 "is_configured": true, 00:11:13.017 "data_offset": 2048, 00:11:13.017 "data_size": 63488 00:11:13.017 }, 00:11:13.017 { 00:11:13.017 "name": "BaseBdev2", 00:11:13.017 "uuid": "53695494-2ae1-49d6-a90a-e0f2ecf5e49b", 00:11:13.017 "is_configured": true, 00:11:13.017 "data_offset": 2048, 00:11:13.017 "data_size": 63488 00:11:13.017 }, 00:11:13.017 { 00:11:13.017 "name": "BaseBdev3", 00:11:13.017 "uuid": "13eea6dd-d2c6-45a3-bedd-84ef82407cba", 00:11:13.017 "is_configured": true, 00:11:13.017 "data_offset": 2048, 00:11:13.017 "data_size": 63488 00:11:13.017 }, 00:11:13.017 { 00:11:13.017 "name": "BaseBdev4", 00:11:13.017 "uuid": "38891f9a-26e6-4eb7-aec8-ad9bfbd93957", 00:11:13.017 "is_configured": true, 00:11:13.017 "data_offset": 2048, 00:11:13.017 "data_size": 63488 00:11:13.017 } 00:11:13.017 ] 00:11:13.017 }' 00:11:13.017 17:32:43 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:13.017 17:32:43 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:13.278 17:32:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:11:13.278 17:32:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:11:13.278 17:32:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:11:13.278 17:32:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:11:13.278 17:32:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:11:13.278 17:32:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:11:13.278 17:32:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:11:13.278 17:32:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:11:13.278 17:32:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:13.278 17:32:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:13.278 [2024-11-27 17:32:44.358239] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:11:13.278 17:32:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:13.278 17:32:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:11:13.278 "name": "Existed_Raid", 00:11:13.278 "aliases": [ 00:11:13.278 "d2f0bcd7-fdf9-4827-a7f7-33139962b64b" 00:11:13.278 ], 00:11:13.278 "product_name": "Raid Volume", 00:11:13.278 "block_size": 512, 00:11:13.278 "num_blocks": 63488, 00:11:13.278 "uuid": "d2f0bcd7-fdf9-4827-a7f7-33139962b64b", 00:11:13.278 "assigned_rate_limits": { 00:11:13.278 "rw_ios_per_sec": 0, 00:11:13.278 "rw_mbytes_per_sec": 0, 00:11:13.278 "r_mbytes_per_sec": 0, 00:11:13.278 "w_mbytes_per_sec": 0 00:11:13.278 }, 00:11:13.278 "claimed": false, 00:11:13.278 "zoned": false, 00:11:13.278 "supported_io_types": { 00:11:13.278 "read": true, 00:11:13.278 "write": true, 00:11:13.278 "unmap": false, 00:11:13.278 "flush": false, 00:11:13.278 "reset": true, 00:11:13.278 "nvme_admin": false, 00:11:13.278 "nvme_io": false, 00:11:13.278 "nvme_io_md": false, 00:11:13.278 "write_zeroes": true, 00:11:13.278 "zcopy": false, 00:11:13.278 "get_zone_info": false, 00:11:13.278 "zone_management": false, 00:11:13.278 "zone_append": false, 00:11:13.278 "compare": false, 00:11:13.278 "compare_and_write": false, 00:11:13.278 "abort": false, 00:11:13.278 "seek_hole": false, 00:11:13.278 "seek_data": false, 00:11:13.278 "copy": false, 00:11:13.278 "nvme_iov_md": false 00:11:13.278 }, 00:11:13.278 "memory_domains": [ 00:11:13.278 { 00:11:13.278 "dma_device_id": "system", 00:11:13.278 "dma_device_type": 1 00:11:13.278 }, 00:11:13.278 { 00:11:13.278 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:13.278 "dma_device_type": 2 00:11:13.278 }, 00:11:13.278 { 00:11:13.278 "dma_device_id": "system", 00:11:13.278 "dma_device_type": 1 00:11:13.278 }, 00:11:13.278 { 00:11:13.278 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:13.278 "dma_device_type": 2 00:11:13.278 }, 00:11:13.278 { 00:11:13.278 "dma_device_id": "system", 00:11:13.278 "dma_device_type": 1 00:11:13.278 }, 00:11:13.278 { 00:11:13.278 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:13.278 "dma_device_type": 2 00:11:13.278 }, 00:11:13.278 { 00:11:13.278 "dma_device_id": "system", 00:11:13.278 "dma_device_type": 1 00:11:13.278 }, 00:11:13.278 { 00:11:13.278 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:13.278 "dma_device_type": 2 00:11:13.278 } 00:11:13.278 ], 00:11:13.278 "driver_specific": { 00:11:13.278 "raid": { 00:11:13.278 "uuid": "d2f0bcd7-fdf9-4827-a7f7-33139962b64b", 00:11:13.278 "strip_size_kb": 0, 00:11:13.278 "state": "online", 00:11:13.278 "raid_level": "raid1", 00:11:13.278 "superblock": true, 00:11:13.278 "num_base_bdevs": 4, 00:11:13.278 "num_base_bdevs_discovered": 4, 00:11:13.278 "num_base_bdevs_operational": 4, 00:11:13.278 "base_bdevs_list": [ 00:11:13.278 { 00:11:13.278 "name": "BaseBdev1", 00:11:13.278 "uuid": "2352465b-833d-48ae-8331-1064c38dfca5", 00:11:13.278 "is_configured": true, 00:11:13.278 "data_offset": 2048, 00:11:13.278 "data_size": 63488 00:11:13.278 }, 00:11:13.278 { 00:11:13.278 "name": "BaseBdev2", 00:11:13.278 "uuid": "53695494-2ae1-49d6-a90a-e0f2ecf5e49b", 00:11:13.278 "is_configured": true, 00:11:13.278 "data_offset": 2048, 00:11:13.278 "data_size": 63488 00:11:13.278 }, 00:11:13.278 { 00:11:13.278 "name": "BaseBdev3", 00:11:13.279 "uuid": "13eea6dd-d2c6-45a3-bedd-84ef82407cba", 00:11:13.279 "is_configured": true, 00:11:13.279 "data_offset": 2048, 00:11:13.279 "data_size": 63488 00:11:13.279 }, 00:11:13.279 { 00:11:13.279 "name": "BaseBdev4", 00:11:13.279 "uuid": "38891f9a-26e6-4eb7-aec8-ad9bfbd93957", 00:11:13.279 "is_configured": true, 00:11:13.279 "data_offset": 2048, 00:11:13.279 "data_size": 63488 00:11:13.279 } 00:11:13.279 ] 00:11:13.279 } 00:11:13.279 } 00:11:13.279 }' 00:11:13.279 17:32:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:11:13.279 17:32:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:11:13.279 BaseBdev2 00:11:13.279 BaseBdev3 00:11:13.279 BaseBdev4' 00:11:13.279 17:32:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:13.539 17:32:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:11:13.539 17:32:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:13.539 17:32:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:11:13.539 17:32:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:13.539 17:32:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:13.539 17:32:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:13.539 17:32:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:13.539 17:32:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:13.539 17:32:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:13.539 17:32:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:13.539 17:32:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:11:13.539 17:32:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:13.539 17:32:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:13.539 17:32:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:13.539 17:32:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:13.539 17:32:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:13.539 17:32:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:13.539 17:32:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:13.539 17:32:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:11:13.539 17:32:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:13.539 17:32:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:13.539 17:32:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:13.539 17:32:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:13.539 17:32:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:13.539 17:32:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:13.539 17:32:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:13.539 17:32:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:11:13.539 17:32:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:13.539 17:32:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:13.539 17:32:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:13.539 17:32:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:13.539 17:32:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:13.539 17:32:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:13.539 17:32:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:11:13.539 17:32:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:13.539 17:32:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:13.539 [2024-11-27 17:32:44.665385] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:11:13.539 17:32:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:13.539 17:32:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:11:13.539 17:32:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:11:13.539 17:32:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:11:13.539 17:32:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@199 -- # return 0 00:11:13.539 17:32:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:11:13.539 17:32:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 3 00:11:13.539 17:32:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:13.539 17:32:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:13.539 17:32:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:13.539 17:32:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:13.539 17:32:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:11:13.539 17:32:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:13.539 17:32:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:13.539 17:32:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:13.539 17:32:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:13.539 17:32:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:13.539 17:32:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:13.539 17:32:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:13.539 17:32:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:13.539 17:32:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:13.799 17:32:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:13.799 "name": "Existed_Raid", 00:11:13.799 "uuid": "d2f0bcd7-fdf9-4827-a7f7-33139962b64b", 00:11:13.799 "strip_size_kb": 0, 00:11:13.799 "state": "online", 00:11:13.799 "raid_level": "raid1", 00:11:13.799 "superblock": true, 00:11:13.799 "num_base_bdevs": 4, 00:11:13.799 "num_base_bdevs_discovered": 3, 00:11:13.799 "num_base_bdevs_operational": 3, 00:11:13.799 "base_bdevs_list": [ 00:11:13.799 { 00:11:13.799 "name": null, 00:11:13.799 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:13.799 "is_configured": false, 00:11:13.799 "data_offset": 0, 00:11:13.799 "data_size": 63488 00:11:13.799 }, 00:11:13.799 { 00:11:13.799 "name": "BaseBdev2", 00:11:13.799 "uuid": "53695494-2ae1-49d6-a90a-e0f2ecf5e49b", 00:11:13.799 "is_configured": true, 00:11:13.799 "data_offset": 2048, 00:11:13.799 "data_size": 63488 00:11:13.799 }, 00:11:13.799 { 00:11:13.799 "name": "BaseBdev3", 00:11:13.799 "uuid": "13eea6dd-d2c6-45a3-bedd-84ef82407cba", 00:11:13.799 "is_configured": true, 00:11:13.799 "data_offset": 2048, 00:11:13.799 "data_size": 63488 00:11:13.799 }, 00:11:13.799 { 00:11:13.799 "name": "BaseBdev4", 00:11:13.799 "uuid": "38891f9a-26e6-4eb7-aec8-ad9bfbd93957", 00:11:13.799 "is_configured": true, 00:11:13.799 "data_offset": 2048, 00:11:13.799 "data_size": 63488 00:11:13.799 } 00:11:13.799 ] 00:11:13.799 }' 00:11:13.799 17:32:44 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:13.799 17:32:44 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:14.059 17:32:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:11:14.059 17:32:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:11:14.059 17:32:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:11:14.059 17:32:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:14.059 17:32:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:14.059 17:32:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:14.059 17:32:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:14.059 17:32:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:11:14.059 17:32:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:11:14.059 17:32:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:11:14.059 17:32:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:14.059 17:32:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:14.059 [2024-11-27 17:32:45.141035] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:11:14.059 17:32:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:14.059 17:32:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:11:14.059 17:32:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:11:14.059 17:32:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:14.059 17:32:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:11:14.059 17:32:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:14.059 17:32:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:14.059 17:32:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:14.059 17:32:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:11:14.059 17:32:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:11:14.059 17:32:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:11:14.059 17:32:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:14.059 17:32:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:14.059 [2024-11-27 17:32:45.217531] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:11:14.059 17:32:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:14.059 17:32:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:11:14.059 17:32:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:11:14.059 17:32:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:14.059 17:32:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:11:14.059 17:32:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:14.059 17:32:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:14.320 17:32:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:14.320 17:32:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:11:14.320 17:32:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:11:14.320 17:32:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev4 00:11:14.320 17:32:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:14.320 17:32:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:14.320 [2024-11-27 17:32:45.293590] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev4 00:11:14.320 [2024-11-27 17:32:45.293771] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:11:14.320 [2024-11-27 17:32:45.314539] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:11:14.320 [2024-11-27 17:32:45.314661] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:11:14.320 [2024-11-27 17:32:45.314728] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:11:14.320 17:32:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:14.320 17:32:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:11:14.320 17:32:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:11:14.320 17:32:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:14.321 17:32:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:11:14.321 17:32:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:14.321 17:32:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:14.321 17:32:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:14.321 17:32:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:11:14.321 17:32:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:11:14.321 17:32:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 4 -gt 2 ']' 00:11:14.321 17:32:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:11:14.321 17:32:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:11:14.321 17:32:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:11:14.321 17:32:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:14.321 17:32:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:14.321 BaseBdev2 00:11:14.321 17:32:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:14.321 17:32:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:11:14.321 17:32:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:11:14.321 17:32:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:11:14.321 17:32:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:11:14.321 17:32:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:11:14.321 17:32:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:11:14.321 17:32:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:11:14.321 17:32:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:14.321 17:32:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:14.321 17:32:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:14.321 17:32:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:11:14.321 17:32:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:14.321 17:32:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:14.321 [ 00:11:14.321 { 00:11:14.321 "name": "BaseBdev2", 00:11:14.321 "aliases": [ 00:11:14.321 "9ba989a7-2c30-4601-bcba-32542d502bb2" 00:11:14.321 ], 00:11:14.321 "product_name": "Malloc disk", 00:11:14.321 "block_size": 512, 00:11:14.321 "num_blocks": 65536, 00:11:14.321 "uuid": "9ba989a7-2c30-4601-bcba-32542d502bb2", 00:11:14.321 "assigned_rate_limits": { 00:11:14.321 "rw_ios_per_sec": 0, 00:11:14.321 "rw_mbytes_per_sec": 0, 00:11:14.321 "r_mbytes_per_sec": 0, 00:11:14.321 "w_mbytes_per_sec": 0 00:11:14.321 }, 00:11:14.321 "claimed": false, 00:11:14.321 "zoned": false, 00:11:14.321 "supported_io_types": { 00:11:14.321 "read": true, 00:11:14.321 "write": true, 00:11:14.321 "unmap": true, 00:11:14.321 "flush": true, 00:11:14.321 "reset": true, 00:11:14.321 "nvme_admin": false, 00:11:14.321 "nvme_io": false, 00:11:14.321 "nvme_io_md": false, 00:11:14.321 "write_zeroes": true, 00:11:14.321 "zcopy": true, 00:11:14.321 "get_zone_info": false, 00:11:14.321 "zone_management": false, 00:11:14.321 "zone_append": false, 00:11:14.321 "compare": false, 00:11:14.321 "compare_and_write": false, 00:11:14.321 "abort": true, 00:11:14.321 "seek_hole": false, 00:11:14.321 "seek_data": false, 00:11:14.321 "copy": true, 00:11:14.321 "nvme_iov_md": false 00:11:14.321 }, 00:11:14.321 "memory_domains": [ 00:11:14.321 { 00:11:14.321 "dma_device_id": "system", 00:11:14.321 "dma_device_type": 1 00:11:14.321 }, 00:11:14.321 { 00:11:14.321 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:14.321 "dma_device_type": 2 00:11:14.321 } 00:11:14.321 ], 00:11:14.321 "driver_specific": {} 00:11:14.321 } 00:11:14.321 ] 00:11:14.321 17:32:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:14.321 17:32:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:11:14.321 17:32:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:11:14.321 17:32:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:11:14.321 17:32:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:11:14.321 17:32:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:14.321 17:32:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:14.321 BaseBdev3 00:11:14.321 17:32:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:14.321 17:32:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:11:14.321 17:32:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:11:14.321 17:32:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:11:14.321 17:32:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:11:14.321 17:32:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:11:14.321 17:32:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:11:14.321 17:32:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:11:14.321 17:32:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:14.321 17:32:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:14.321 17:32:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:14.321 17:32:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:11:14.321 17:32:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:14.321 17:32:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:14.321 [ 00:11:14.321 { 00:11:14.321 "name": "BaseBdev3", 00:11:14.321 "aliases": [ 00:11:14.321 "1720c9c5-ed38-462f-8449-c28508d32fa5" 00:11:14.321 ], 00:11:14.321 "product_name": "Malloc disk", 00:11:14.321 "block_size": 512, 00:11:14.321 "num_blocks": 65536, 00:11:14.321 "uuid": "1720c9c5-ed38-462f-8449-c28508d32fa5", 00:11:14.321 "assigned_rate_limits": { 00:11:14.321 "rw_ios_per_sec": 0, 00:11:14.321 "rw_mbytes_per_sec": 0, 00:11:14.321 "r_mbytes_per_sec": 0, 00:11:14.321 "w_mbytes_per_sec": 0 00:11:14.321 }, 00:11:14.321 "claimed": false, 00:11:14.321 "zoned": false, 00:11:14.321 "supported_io_types": { 00:11:14.321 "read": true, 00:11:14.321 "write": true, 00:11:14.321 "unmap": true, 00:11:14.321 "flush": true, 00:11:14.321 "reset": true, 00:11:14.321 "nvme_admin": false, 00:11:14.321 "nvme_io": false, 00:11:14.321 "nvme_io_md": false, 00:11:14.321 "write_zeroes": true, 00:11:14.321 "zcopy": true, 00:11:14.321 "get_zone_info": false, 00:11:14.321 "zone_management": false, 00:11:14.321 "zone_append": false, 00:11:14.321 "compare": false, 00:11:14.321 "compare_and_write": false, 00:11:14.321 "abort": true, 00:11:14.321 "seek_hole": false, 00:11:14.321 "seek_data": false, 00:11:14.321 "copy": true, 00:11:14.321 "nvme_iov_md": false 00:11:14.321 }, 00:11:14.321 "memory_domains": [ 00:11:14.321 { 00:11:14.321 "dma_device_id": "system", 00:11:14.321 "dma_device_type": 1 00:11:14.321 }, 00:11:14.321 { 00:11:14.321 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:14.321 "dma_device_type": 2 00:11:14.321 } 00:11:14.321 ], 00:11:14.321 "driver_specific": {} 00:11:14.321 } 00:11:14.321 ] 00:11:14.321 17:32:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:14.321 17:32:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:11:14.321 17:32:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:11:14.321 17:32:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:11:14.321 17:32:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:11:14.321 17:32:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:14.321 17:32:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:14.321 BaseBdev4 00:11:14.321 17:32:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:14.321 17:32:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev4 00:11:14.321 17:32:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev4 00:11:14.321 17:32:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:11:14.321 17:32:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:11:14.321 17:32:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:11:14.321 17:32:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:11:14.321 17:32:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:11:14.321 17:32:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:14.321 17:32:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:14.582 17:32:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:14.582 17:32:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:11:14.582 17:32:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:14.582 17:32:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:14.582 [ 00:11:14.582 { 00:11:14.582 "name": "BaseBdev4", 00:11:14.582 "aliases": [ 00:11:14.582 "ff131e9c-8342-44aa-ad9b-51c1db5af0d8" 00:11:14.582 ], 00:11:14.582 "product_name": "Malloc disk", 00:11:14.582 "block_size": 512, 00:11:14.582 "num_blocks": 65536, 00:11:14.582 "uuid": "ff131e9c-8342-44aa-ad9b-51c1db5af0d8", 00:11:14.582 "assigned_rate_limits": { 00:11:14.582 "rw_ios_per_sec": 0, 00:11:14.582 "rw_mbytes_per_sec": 0, 00:11:14.582 "r_mbytes_per_sec": 0, 00:11:14.582 "w_mbytes_per_sec": 0 00:11:14.582 }, 00:11:14.582 "claimed": false, 00:11:14.582 "zoned": false, 00:11:14.582 "supported_io_types": { 00:11:14.582 "read": true, 00:11:14.582 "write": true, 00:11:14.582 "unmap": true, 00:11:14.582 "flush": true, 00:11:14.582 "reset": true, 00:11:14.582 "nvme_admin": false, 00:11:14.582 "nvme_io": false, 00:11:14.582 "nvme_io_md": false, 00:11:14.582 "write_zeroes": true, 00:11:14.582 "zcopy": true, 00:11:14.582 "get_zone_info": false, 00:11:14.582 "zone_management": false, 00:11:14.582 "zone_append": false, 00:11:14.582 "compare": false, 00:11:14.582 "compare_and_write": false, 00:11:14.582 "abort": true, 00:11:14.582 "seek_hole": false, 00:11:14.582 "seek_data": false, 00:11:14.582 "copy": true, 00:11:14.582 "nvme_iov_md": false 00:11:14.582 }, 00:11:14.582 "memory_domains": [ 00:11:14.582 { 00:11:14.582 "dma_device_id": "system", 00:11:14.582 "dma_device_type": 1 00:11:14.582 }, 00:11:14.582 { 00:11:14.582 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:14.582 "dma_device_type": 2 00:11:14.582 } 00:11:14.582 ], 00:11:14.582 "driver_specific": {} 00:11:14.582 } 00:11:14.582 ] 00:11:14.582 17:32:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:14.582 17:32:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:11:14.582 17:32:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:11:14.582 17:32:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:11:14.582 17:32:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:11:14.582 17:32:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:14.582 17:32:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:14.582 [2024-11-27 17:32:45.550017] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:11:14.582 [2024-11-27 17:32:45.550109] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:11:14.582 [2024-11-27 17:32:45.550172] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:11:14.582 [2024-11-27 17:32:45.552294] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:11:14.582 [2024-11-27 17:32:45.552399] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:11:14.582 17:32:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:14.582 17:32:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:11:14.582 17:32:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:14.582 17:32:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:14.582 17:32:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:14.582 17:32:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:14.582 17:32:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:14.582 17:32:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:14.582 17:32:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:14.582 17:32:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:14.582 17:32:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:14.582 17:32:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:14.582 17:32:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:14.582 17:32:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:14.582 17:32:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:14.582 17:32:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:14.582 17:32:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:14.582 "name": "Existed_Raid", 00:11:14.582 "uuid": "3c24ae85-615c-40fe-95e5-629669d53de6", 00:11:14.582 "strip_size_kb": 0, 00:11:14.582 "state": "configuring", 00:11:14.582 "raid_level": "raid1", 00:11:14.582 "superblock": true, 00:11:14.582 "num_base_bdevs": 4, 00:11:14.582 "num_base_bdevs_discovered": 3, 00:11:14.582 "num_base_bdevs_operational": 4, 00:11:14.582 "base_bdevs_list": [ 00:11:14.582 { 00:11:14.582 "name": "BaseBdev1", 00:11:14.582 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:14.582 "is_configured": false, 00:11:14.582 "data_offset": 0, 00:11:14.582 "data_size": 0 00:11:14.582 }, 00:11:14.582 { 00:11:14.582 "name": "BaseBdev2", 00:11:14.582 "uuid": "9ba989a7-2c30-4601-bcba-32542d502bb2", 00:11:14.583 "is_configured": true, 00:11:14.583 "data_offset": 2048, 00:11:14.583 "data_size": 63488 00:11:14.583 }, 00:11:14.583 { 00:11:14.583 "name": "BaseBdev3", 00:11:14.583 "uuid": "1720c9c5-ed38-462f-8449-c28508d32fa5", 00:11:14.583 "is_configured": true, 00:11:14.583 "data_offset": 2048, 00:11:14.583 "data_size": 63488 00:11:14.583 }, 00:11:14.583 { 00:11:14.583 "name": "BaseBdev4", 00:11:14.583 "uuid": "ff131e9c-8342-44aa-ad9b-51c1db5af0d8", 00:11:14.583 "is_configured": true, 00:11:14.583 "data_offset": 2048, 00:11:14.583 "data_size": 63488 00:11:14.583 } 00:11:14.583 ] 00:11:14.583 }' 00:11:14.583 17:32:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:14.583 17:32:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:14.843 17:32:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:11:14.843 17:32:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:14.843 17:32:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:14.843 [2024-11-27 17:32:45.937349] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:11:14.843 17:32:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:14.843 17:32:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:11:14.843 17:32:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:14.843 17:32:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:14.843 17:32:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:14.843 17:32:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:14.843 17:32:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:14.843 17:32:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:14.843 17:32:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:14.843 17:32:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:14.843 17:32:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:14.843 17:32:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:14.843 17:32:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:14.843 17:32:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:14.843 17:32:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:14.843 17:32:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:14.843 17:32:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:14.843 "name": "Existed_Raid", 00:11:14.843 "uuid": "3c24ae85-615c-40fe-95e5-629669d53de6", 00:11:14.843 "strip_size_kb": 0, 00:11:14.843 "state": "configuring", 00:11:14.843 "raid_level": "raid1", 00:11:14.843 "superblock": true, 00:11:14.843 "num_base_bdevs": 4, 00:11:14.843 "num_base_bdevs_discovered": 2, 00:11:14.843 "num_base_bdevs_operational": 4, 00:11:14.843 "base_bdevs_list": [ 00:11:14.843 { 00:11:14.843 "name": "BaseBdev1", 00:11:14.843 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:14.843 "is_configured": false, 00:11:14.843 "data_offset": 0, 00:11:14.843 "data_size": 0 00:11:14.843 }, 00:11:14.843 { 00:11:14.843 "name": null, 00:11:14.843 "uuid": "9ba989a7-2c30-4601-bcba-32542d502bb2", 00:11:14.843 "is_configured": false, 00:11:14.843 "data_offset": 0, 00:11:14.843 "data_size": 63488 00:11:14.843 }, 00:11:14.843 { 00:11:14.843 "name": "BaseBdev3", 00:11:14.843 "uuid": "1720c9c5-ed38-462f-8449-c28508d32fa5", 00:11:14.843 "is_configured": true, 00:11:14.843 "data_offset": 2048, 00:11:14.843 "data_size": 63488 00:11:14.843 }, 00:11:14.843 { 00:11:14.843 "name": "BaseBdev4", 00:11:14.843 "uuid": "ff131e9c-8342-44aa-ad9b-51c1db5af0d8", 00:11:14.843 "is_configured": true, 00:11:14.843 "data_offset": 2048, 00:11:14.843 "data_size": 63488 00:11:14.843 } 00:11:14.843 ] 00:11:14.843 }' 00:11:14.843 17:32:45 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:14.843 17:32:45 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:15.444 17:32:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:11:15.444 17:32:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:15.444 17:32:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:15.444 17:32:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:15.444 17:32:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:15.444 17:32:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:11:15.444 17:32:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:11:15.444 17:32:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:15.444 17:32:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:15.444 [2024-11-27 17:32:46.401232] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:11:15.444 BaseBdev1 00:11:15.444 17:32:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:15.444 17:32:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:11:15.444 17:32:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:11:15.444 17:32:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:11:15.444 17:32:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:11:15.444 17:32:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:11:15.444 17:32:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:11:15.444 17:32:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:11:15.444 17:32:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:15.444 17:32:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:15.444 17:32:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:15.444 17:32:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:11:15.444 17:32:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:15.444 17:32:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:15.444 [ 00:11:15.444 { 00:11:15.444 "name": "BaseBdev1", 00:11:15.444 "aliases": [ 00:11:15.444 "9f178a95-1965-4357-98c3-300d891db811" 00:11:15.444 ], 00:11:15.444 "product_name": "Malloc disk", 00:11:15.444 "block_size": 512, 00:11:15.444 "num_blocks": 65536, 00:11:15.444 "uuid": "9f178a95-1965-4357-98c3-300d891db811", 00:11:15.444 "assigned_rate_limits": { 00:11:15.444 "rw_ios_per_sec": 0, 00:11:15.444 "rw_mbytes_per_sec": 0, 00:11:15.444 "r_mbytes_per_sec": 0, 00:11:15.444 "w_mbytes_per_sec": 0 00:11:15.444 }, 00:11:15.444 "claimed": true, 00:11:15.444 "claim_type": "exclusive_write", 00:11:15.444 "zoned": false, 00:11:15.444 "supported_io_types": { 00:11:15.444 "read": true, 00:11:15.444 "write": true, 00:11:15.444 "unmap": true, 00:11:15.444 "flush": true, 00:11:15.445 "reset": true, 00:11:15.445 "nvme_admin": false, 00:11:15.445 "nvme_io": false, 00:11:15.445 "nvme_io_md": false, 00:11:15.445 "write_zeroes": true, 00:11:15.445 "zcopy": true, 00:11:15.445 "get_zone_info": false, 00:11:15.445 "zone_management": false, 00:11:15.445 "zone_append": false, 00:11:15.445 "compare": false, 00:11:15.445 "compare_and_write": false, 00:11:15.445 "abort": true, 00:11:15.445 "seek_hole": false, 00:11:15.445 "seek_data": false, 00:11:15.445 "copy": true, 00:11:15.445 "nvme_iov_md": false 00:11:15.445 }, 00:11:15.445 "memory_domains": [ 00:11:15.445 { 00:11:15.445 "dma_device_id": "system", 00:11:15.445 "dma_device_type": 1 00:11:15.445 }, 00:11:15.445 { 00:11:15.445 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:15.445 "dma_device_type": 2 00:11:15.445 } 00:11:15.445 ], 00:11:15.445 "driver_specific": {} 00:11:15.445 } 00:11:15.445 ] 00:11:15.445 17:32:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:15.445 17:32:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:11:15.445 17:32:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:11:15.445 17:32:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:15.445 17:32:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:15.445 17:32:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:15.445 17:32:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:15.445 17:32:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:15.445 17:32:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:15.445 17:32:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:15.445 17:32:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:15.445 17:32:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:15.445 17:32:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:15.445 17:32:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:15.445 17:32:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:15.445 17:32:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:15.445 17:32:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:15.445 17:32:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:15.445 "name": "Existed_Raid", 00:11:15.445 "uuid": "3c24ae85-615c-40fe-95e5-629669d53de6", 00:11:15.445 "strip_size_kb": 0, 00:11:15.445 "state": "configuring", 00:11:15.445 "raid_level": "raid1", 00:11:15.445 "superblock": true, 00:11:15.445 "num_base_bdevs": 4, 00:11:15.445 "num_base_bdevs_discovered": 3, 00:11:15.445 "num_base_bdevs_operational": 4, 00:11:15.445 "base_bdevs_list": [ 00:11:15.445 { 00:11:15.445 "name": "BaseBdev1", 00:11:15.445 "uuid": "9f178a95-1965-4357-98c3-300d891db811", 00:11:15.445 "is_configured": true, 00:11:15.445 "data_offset": 2048, 00:11:15.445 "data_size": 63488 00:11:15.445 }, 00:11:15.445 { 00:11:15.445 "name": null, 00:11:15.445 "uuid": "9ba989a7-2c30-4601-bcba-32542d502bb2", 00:11:15.445 "is_configured": false, 00:11:15.445 "data_offset": 0, 00:11:15.445 "data_size": 63488 00:11:15.445 }, 00:11:15.445 { 00:11:15.445 "name": "BaseBdev3", 00:11:15.445 "uuid": "1720c9c5-ed38-462f-8449-c28508d32fa5", 00:11:15.445 "is_configured": true, 00:11:15.445 "data_offset": 2048, 00:11:15.445 "data_size": 63488 00:11:15.445 }, 00:11:15.445 { 00:11:15.445 "name": "BaseBdev4", 00:11:15.445 "uuid": "ff131e9c-8342-44aa-ad9b-51c1db5af0d8", 00:11:15.445 "is_configured": true, 00:11:15.445 "data_offset": 2048, 00:11:15.445 "data_size": 63488 00:11:15.445 } 00:11:15.445 ] 00:11:15.445 }' 00:11:15.445 17:32:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:15.445 17:32:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:15.706 17:32:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:11:15.706 17:32:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:15.706 17:32:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:15.706 17:32:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:15.706 17:32:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:15.706 17:32:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:11:15.706 17:32:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:11:15.706 17:32:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:15.706 17:32:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:15.706 [2024-11-27 17:32:46.860490] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:11:15.706 17:32:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:15.706 17:32:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:11:15.706 17:32:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:15.706 17:32:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:15.706 17:32:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:15.706 17:32:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:15.706 17:32:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:15.706 17:32:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:15.706 17:32:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:15.706 17:32:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:15.706 17:32:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:15.706 17:32:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:15.706 17:32:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:15.706 17:32:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:15.706 17:32:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:15.706 17:32:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:15.966 17:32:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:15.966 "name": "Existed_Raid", 00:11:15.966 "uuid": "3c24ae85-615c-40fe-95e5-629669d53de6", 00:11:15.966 "strip_size_kb": 0, 00:11:15.966 "state": "configuring", 00:11:15.966 "raid_level": "raid1", 00:11:15.966 "superblock": true, 00:11:15.966 "num_base_bdevs": 4, 00:11:15.966 "num_base_bdevs_discovered": 2, 00:11:15.966 "num_base_bdevs_operational": 4, 00:11:15.966 "base_bdevs_list": [ 00:11:15.966 { 00:11:15.966 "name": "BaseBdev1", 00:11:15.966 "uuid": "9f178a95-1965-4357-98c3-300d891db811", 00:11:15.966 "is_configured": true, 00:11:15.966 "data_offset": 2048, 00:11:15.966 "data_size": 63488 00:11:15.966 }, 00:11:15.966 { 00:11:15.966 "name": null, 00:11:15.966 "uuid": "9ba989a7-2c30-4601-bcba-32542d502bb2", 00:11:15.966 "is_configured": false, 00:11:15.966 "data_offset": 0, 00:11:15.966 "data_size": 63488 00:11:15.966 }, 00:11:15.966 { 00:11:15.966 "name": null, 00:11:15.966 "uuid": "1720c9c5-ed38-462f-8449-c28508d32fa5", 00:11:15.966 "is_configured": false, 00:11:15.966 "data_offset": 0, 00:11:15.966 "data_size": 63488 00:11:15.966 }, 00:11:15.966 { 00:11:15.966 "name": "BaseBdev4", 00:11:15.966 "uuid": "ff131e9c-8342-44aa-ad9b-51c1db5af0d8", 00:11:15.966 "is_configured": true, 00:11:15.966 "data_offset": 2048, 00:11:15.966 "data_size": 63488 00:11:15.966 } 00:11:15.966 ] 00:11:15.966 }' 00:11:15.966 17:32:46 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:15.966 17:32:46 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:16.226 17:32:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:11:16.226 17:32:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:16.226 17:32:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:16.226 17:32:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:16.226 17:32:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:16.226 17:32:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:11:16.226 17:32:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:11:16.226 17:32:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:16.226 17:32:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:16.226 [2024-11-27 17:32:47.339708] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:11:16.226 17:32:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:16.226 17:32:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:11:16.226 17:32:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:16.226 17:32:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:16.226 17:32:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:16.226 17:32:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:16.226 17:32:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:16.226 17:32:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:16.226 17:32:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:16.226 17:32:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:16.226 17:32:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:16.226 17:32:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:16.226 17:32:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:16.226 17:32:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:16.226 17:32:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:16.226 17:32:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:16.226 17:32:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:16.226 "name": "Existed_Raid", 00:11:16.226 "uuid": "3c24ae85-615c-40fe-95e5-629669d53de6", 00:11:16.226 "strip_size_kb": 0, 00:11:16.226 "state": "configuring", 00:11:16.226 "raid_level": "raid1", 00:11:16.226 "superblock": true, 00:11:16.226 "num_base_bdevs": 4, 00:11:16.226 "num_base_bdevs_discovered": 3, 00:11:16.226 "num_base_bdevs_operational": 4, 00:11:16.226 "base_bdevs_list": [ 00:11:16.227 { 00:11:16.227 "name": "BaseBdev1", 00:11:16.227 "uuid": "9f178a95-1965-4357-98c3-300d891db811", 00:11:16.227 "is_configured": true, 00:11:16.227 "data_offset": 2048, 00:11:16.227 "data_size": 63488 00:11:16.227 }, 00:11:16.227 { 00:11:16.227 "name": null, 00:11:16.227 "uuid": "9ba989a7-2c30-4601-bcba-32542d502bb2", 00:11:16.227 "is_configured": false, 00:11:16.227 "data_offset": 0, 00:11:16.227 "data_size": 63488 00:11:16.227 }, 00:11:16.227 { 00:11:16.227 "name": "BaseBdev3", 00:11:16.227 "uuid": "1720c9c5-ed38-462f-8449-c28508d32fa5", 00:11:16.227 "is_configured": true, 00:11:16.227 "data_offset": 2048, 00:11:16.227 "data_size": 63488 00:11:16.227 }, 00:11:16.227 { 00:11:16.227 "name": "BaseBdev4", 00:11:16.227 "uuid": "ff131e9c-8342-44aa-ad9b-51c1db5af0d8", 00:11:16.227 "is_configured": true, 00:11:16.227 "data_offset": 2048, 00:11:16.227 "data_size": 63488 00:11:16.227 } 00:11:16.227 ] 00:11:16.227 }' 00:11:16.227 17:32:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:16.227 17:32:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:16.797 17:32:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:11:16.797 17:32:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:16.797 17:32:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:16.797 17:32:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:16.797 17:32:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:16.797 17:32:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:11:16.797 17:32:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:11:16.797 17:32:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:16.797 17:32:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:16.797 [2024-11-27 17:32:47.810893] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:11:16.797 17:32:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:16.797 17:32:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:11:16.797 17:32:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:16.797 17:32:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:16.797 17:32:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:16.797 17:32:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:16.797 17:32:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:16.797 17:32:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:16.797 17:32:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:16.797 17:32:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:16.797 17:32:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:16.797 17:32:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:16.797 17:32:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:16.797 17:32:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:16.797 17:32:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:16.797 17:32:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:16.797 17:32:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:16.797 "name": "Existed_Raid", 00:11:16.797 "uuid": "3c24ae85-615c-40fe-95e5-629669d53de6", 00:11:16.797 "strip_size_kb": 0, 00:11:16.797 "state": "configuring", 00:11:16.797 "raid_level": "raid1", 00:11:16.797 "superblock": true, 00:11:16.797 "num_base_bdevs": 4, 00:11:16.797 "num_base_bdevs_discovered": 2, 00:11:16.797 "num_base_bdevs_operational": 4, 00:11:16.797 "base_bdevs_list": [ 00:11:16.797 { 00:11:16.797 "name": null, 00:11:16.797 "uuid": "9f178a95-1965-4357-98c3-300d891db811", 00:11:16.797 "is_configured": false, 00:11:16.797 "data_offset": 0, 00:11:16.797 "data_size": 63488 00:11:16.797 }, 00:11:16.797 { 00:11:16.797 "name": null, 00:11:16.797 "uuid": "9ba989a7-2c30-4601-bcba-32542d502bb2", 00:11:16.797 "is_configured": false, 00:11:16.797 "data_offset": 0, 00:11:16.797 "data_size": 63488 00:11:16.797 }, 00:11:16.797 { 00:11:16.797 "name": "BaseBdev3", 00:11:16.797 "uuid": "1720c9c5-ed38-462f-8449-c28508d32fa5", 00:11:16.797 "is_configured": true, 00:11:16.797 "data_offset": 2048, 00:11:16.797 "data_size": 63488 00:11:16.797 }, 00:11:16.797 { 00:11:16.797 "name": "BaseBdev4", 00:11:16.797 "uuid": "ff131e9c-8342-44aa-ad9b-51c1db5af0d8", 00:11:16.797 "is_configured": true, 00:11:16.797 "data_offset": 2048, 00:11:16.797 "data_size": 63488 00:11:16.797 } 00:11:16.797 ] 00:11:16.797 }' 00:11:16.797 17:32:47 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:16.797 17:32:47 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:17.366 17:32:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:17.366 17:32:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:17.366 17:32:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:17.366 17:32:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:11:17.366 17:32:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:17.366 17:32:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:11:17.367 17:32:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:11:17.367 17:32:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:17.367 17:32:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:17.367 [2024-11-27 17:32:48.313511] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:11:17.367 17:32:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:17.367 17:32:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 4 00:11:17.367 17:32:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:17.367 17:32:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:17.367 17:32:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:17.367 17:32:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:17.367 17:32:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:17.367 17:32:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:17.367 17:32:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:17.367 17:32:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:17.367 17:32:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:17.367 17:32:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:17.367 17:32:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:17.367 17:32:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:17.367 17:32:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:17.367 17:32:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:17.367 17:32:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:17.367 "name": "Existed_Raid", 00:11:17.367 "uuid": "3c24ae85-615c-40fe-95e5-629669d53de6", 00:11:17.367 "strip_size_kb": 0, 00:11:17.367 "state": "configuring", 00:11:17.367 "raid_level": "raid1", 00:11:17.367 "superblock": true, 00:11:17.367 "num_base_bdevs": 4, 00:11:17.367 "num_base_bdevs_discovered": 3, 00:11:17.367 "num_base_bdevs_operational": 4, 00:11:17.367 "base_bdevs_list": [ 00:11:17.367 { 00:11:17.367 "name": null, 00:11:17.367 "uuid": "9f178a95-1965-4357-98c3-300d891db811", 00:11:17.367 "is_configured": false, 00:11:17.367 "data_offset": 0, 00:11:17.367 "data_size": 63488 00:11:17.367 }, 00:11:17.367 { 00:11:17.367 "name": "BaseBdev2", 00:11:17.367 "uuid": "9ba989a7-2c30-4601-bcba-32542d502bb2", 00:11:17.367 "is_configured": true, 00:11:17.367 "data_offset": 2048, 00:11:17.367 "data_size": 63488 00:11:17.367 }, 00:11:17.367 { 00:11:17.367 "name": "BaseBdev3", 00:11:17.367 "uuid": "1720c9c5-ed38-462f-8449-c28508d32fa5", 00:11:17.367 "is_configured": true, 00:11:17.367 "data_offset": 2048, 00:11:17.367 "data_size": 63488 00:11:17.367 }, 00:11:17.367 { 00:11:17.367 "name": "BaseBdev4", 00:11:17.367 "uuid": "ff131e9c-8342-44aa-ad9b-51c1db5af0d8", 00:11:17.367 "is_configured": true, 00:11:17.367 "data_offset": 2048, 00:11:17.367 "data_size": 63488 00:11:17.367 } 00:11:17.367 ] 00:11:17.367 }' 00:11:17.367 17:32:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:17.367 17:32:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:17.627 17:32:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:17.627 17:32:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:11:17.627 17:32:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:17.627 17:32:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:17.627 17:32:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:17.627 17:32:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:11:17.627 17:32:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:17.627 17:32:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:11:17.627 17:32:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:17.627 17:32:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:17.888 17:32:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:17.888 17:32:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 9f178a95-1965-4357-98c3-300d891db811 00:11:17.888 17:32:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:17.888 17:32:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:17.888 [2024-11-27 17:32:48.873447] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:11:17.888 [2024-11-27 17:32:48.873654] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:11:17.888 [2024-11-27 17:32:48.873671] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:11:17.888 [2024-11-27 17:32:48.873947] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002a10 00:11:17.888 NewBaseBdev 00:11:17.888 [2024-11-27 17:32:48.874098] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:11:17.888 [2024-11-27 17:32:48.874115] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001c80 00:11:17.888 [2024-11-27 17:32:48.874245] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:17.888 17:32:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:17.888 17:32:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:11:17.888 17:32:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=NewBaseBdev 00:11:17.888 17:32:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:11:17.888 17:32:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:11:17.888 17:32:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:11:17.888 17:32:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:11:17.888 17:32:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:11:17.888 17:32:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:17.888 17:32:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:17.888 17:32:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:17.888 17:32:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:11:17.888 17:32:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:17.888 17:32:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:17.888 [ 00:11:17.888 { 00:11:17.888 "name": "NewBaseBdev", 00:11:17.888 "aliases": [ 00:11:17.888 "9f178a95-1965-4357-98c3-300d891db811" 00:11:17.888 ], 00:11:17.888 "product_name": "Malloc disk", 00:11:17.888 "block_size": 512, 00:11:17.888 "num_blocks": 65536, 00:11:17.888 "uuid": "9f178a95-1965-4357-98c3-300d891db811", 00:11:17.888 "assigned_rate_limits": { 00:11:17.888 "rw_ios_per_sec": 0, 00:11:17.888 "rw_mbytes_per_sec": 0, 00:11:17.888 "r_mbytes_per_sec": 0, 00:11:17.888 "w_mbytes_per_sec": 0 00:11:17.888 }, 00:11:17.888 "claimed": true, 00:11:17.888 "claim_type": "exclusive_write", 00:11:17.888 "zoned": false, 00:11:17.888 "supported_io_types": { 00:11:17.888 "read": true, 00:11:17.888 "write": true, 00:11:17.888 "unmap": true, 00:11:17.888 "flush": true, 00:11:17.888 "reset": true, 00:11:17.888 "nvme_admin": false, 00:11:17.888 "nvme_io": false, 00:11:17.888 "nvme_io_md": false, 00:11:17.888 "write_zeroes": true, 00:11:17.888 "zcopy": true, 00:11:17.888 "get_zone_info": false, 00:11:17.888 "zone_management": false, 00:11:17.888 "zone_append": false, 00:11:17.888 "compare": false, 00:11:17.888 "compare_and_write": false, 00:11:17.888 "abort": true, 00:11:17.888 "seek_hole": false, 00:11:17.888 "seek_data": false, 00:11:17.888 "copy": true, 00:11:17.888 "nvme_iov_md": false 00:11:17.888 }, 00:11:17.888 "memory_domains": [ 00:11:17.888 { 00:11:17.888 "dma_device_id": "system", 00:11:17.888 "dma_device_type": 1 00:11:17.888 }, 00:11:17.888 { 00:11:17.888 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:17.888 "dma_device_type": 2 00:11:17.888 } 00:11:17.888 ], 00:11:17.888 "driver_specific": {} 00:11:17.888 } 00:11:17.888 ] 00:11:17.888 17:32:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:17.888 17:32:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:11:17.888 17:32:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid1 0 4 00:11:17.888 17:32:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:11:17.888 17:32:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:17.888 17:32:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:17.888 17:32:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:17.888 17:32:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:17.888 17:32:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:17.888 17:32:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:17.888 17:32:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:17.888 17:32:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:17.888 17:32:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:17.888 17:32:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:11:17.888 17:32:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:17.888 17:32:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:17.888 17:32:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:17.888 17:32:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:17.888 "name": "Existed_Raid", 00:11:17.888 "uuid": "3c24ae85-615c-40fe-95e5-629669d53de6", 00:11:17.888 "strip_size_kb": 0, 00:11:17.888 "state": "online", 00:11:17.888 "raid_level": "raid1", 00:11:17.888 "superblock": true, 00:11:17.888 "num_base_bdevs": 4, 00:11:17.888 "num_base_bdevs_discovered": 4, 00:11:17.888 "num_base_bdevs_operational": 4, 00:11:17.888 "base_bdevs_list": [ 00:11:17.888 { 00:11:17.888 "name": "NewBaseBdev", 00:11:17.888 "uuid": "9f178a95-1965-4357-98c3-300d891db811", 00:11:17.888 "is_configured": true, 00:11:17.888 "data_offset": 2048, 00:11:17.888 "data_size": 63488 00:11:17.888 }, 00:11:17.889 { 00:11:17.889 "name": "BaseBdev2", 00:11:17.889 "uuid": "9ba989a7-2c30-4601-bcba-32542d502bb2", 00:11:17.889 "is_configured": true, 00:11:17.889 "data_offset": 2048, 00:11:17.889 "data_size": 63488 00:11:17.889 }, 00:11:17.889 { 00:11:17.889 "name": "BaseBdev3", 00:11:17.889 "uuid": "1720c9c5-ed38-462f-8449-c28508d32fa5", 00:11:17.889 "is_configured": true, 00:11:17.889 "data_offset": 2048, 00:11:17.889 "data_size": 63488 00:11:17.889 }, 00:11:17.889 { 00:11:17.889 "name": "BaseBdev4", 00:11:17.889 "uuid": "ff131e9c-8342-44aa-ad9b-51c1db5af0d8", 00:11:17.889 "is_configured": true, 00:11:17.889 "data_offset": 2048, 00:11:17.889 "data_size": 63488 00:11:17.889 } 00:11:17.889 ] 00:11:17.889 }' 00:11:17.889 17:32:48 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:17.889 17:32:48 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:18.480 17:32:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:11:18.480 17:32:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:11:18.480 17:32:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:11:18.480 17:32:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:11:18.480 17:32:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:11:18.480 17:32:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:11:18.480 17:32:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:11:18.480 17:32:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:11:18.480 17:32:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:18.480 17:32:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:18.480 [2024-11-27 17:32:49.384929] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:11:18.480 17:32:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:18.480 17:32:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:11:18.480 "name": "Existed_Raid", 00:11:18.480 "aliases": [ 00:11:18.480 "3c24ae85-615c-40fe-95e5-629669d53de6" 00:11:18.480 ], 00:11:18.480 "product_name": "Raid Volume", 00:11:18.480 "block_size": 512, 00:11:18.480 "num_blocks": 63488, 00:11:18.480 "uuid": "3c24ae85-615c-40fe-95e5-629669d53de6", 00:11:18.480 "assigned_rate_limits": { 00:11:18.480 "rw_ios_per_sec": 0, 00:11:18.480 "rw_mbytes_per_sec": 0, 00:11:18.480 "r_mbytes_per_sec": 0, 00:11:18.480 "w_mbytes_per_sec": 0 00:11:18.480 }, 00:11:18.480 "claimed": false, 00:11:18.480 "zoned": false, 00:11:18.480 "supported_io_types": { 00:11:18.480 "read": true, 00:11:18.480 "write": true, 00:11:18.480 "unmap": false, 00:11:18.480 "flush": false, 00:11:18.480 "reset": true, 00:11:18.480 "nvme_admin": false, 00:11:18.480 "nvme_io": false, 00:11:18.480 "nvme_io_md": false, 00:11:18.480 "write_zeroes": true, 00:11:18.480 "zcopy": false, 00:11:18.480 "get_zone_info": false, 00:11:18.480 "zone_management": false, 00:11:18.480 "zone_append": false, 00:11:18.480 "compare": false, 00:11:18.480 "compare_and_write": false, 00:11:18.480 "abort": false, 00:11:18.480 "seek_hole": false, 00:11:18.480 "seek_data": false, 00:11:18.480 "copy": false, 00:11:18.480 "nvme_iov_md": false 00:11:18.480 }, 00:11:18.480 "memory_domains": [ 00:11:18.480 { 00:11:18.480 "dma_device_id": "system", 00:11:18.480 "dma_device_type": 1 00:11:18.480 }, 00:11:18.480 { 00:11:18.480 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:18.480 "dma_device_type": 2 00:11:18.480 }, 00:11:18.480 { 00:11:18.480 "dma_device_id": "system", 00:11:18.480 "dma_device_type": 1 00:11:18.480 }, 00:11:18.480 { 00:11:18.480 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:18.480 "dma_device_type": 2 00:11:18.480 }, 00:11:18.480 { 00:11:18.480 "dma_device_id": "system", 00:11:18.480 "dma_device_type": 1 00:11:18.480 }, 00:11:18.480 { 00:11:18.480 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:18.480 "dma_device_type": 2 00:11:18.480 }, 00:11:18.480 { 00:11:18.480 "dma_device_id": "system", 00:11:18.480 "dma_device_type": 1 00:11:18.480 }, 00:11:18.480 { 00:11:18.480 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:18.480 "dma_device_type": 2 00:11:18.480 } 00:11:18.480 ], 00:11:18.480 "driver_specific": { 00:11:18.480 "raid": { 00:11:18.480 "uuid": "3c24ae85-615c-40fe-95e5-629669d53de6", 00:11:18.480 "strip_size_kb": 0, 00:11:18.480 "state": "online", 00:11:18.480 "raid_level": "raid1", 00:11:18.480 "superblock": true, 00:11:18.480 "num_base_bdevs": 4, 00:11:18.481 "num_base_bdevs_discovered": 4, 00:11:18.481 "num_base_bdevs_operational": 4, 00:11:18.481 "base_bdevs_list": [ 00:11:18.481 { 00:11:18.481 "name": "NewBaseBdev", 00:11:18.481 "uuid": "9f178a95-1965-4357-98c3-300d891db811", 00:11:18.481 "is_configured": true, 00:11:18.481 "data_offset": 2048, 00:11:18.481 "data_size": 63488 00:11:18.481 }, 00:11:18.481 { 00:11:18.481 "name": "BaseBdev2", 00:11:18.481 "uuid": "9ba989a7-2c30-4601-bcba-32542d502bb2", 00:11:18.481 "is_configured": true, 00:11:18.481 "data_offset": 2048, 00:11:18.481 "data_size": 63488 00:11:18.481 }, 00:11:18.481 { 00:11:18.481 "name": "BaseBdev3", 00:11:18.481 "uuid": "1720c9c5-ed38-462f-8449-c28508d32fa5", 00:11:18.481 "is_configured": true, 00:11:18.481 "data_offset": 2048, 00:11:18.481 "data_size": 63488 00:11:18.481 }, 00:11:18.481 { 00:11:18.481 "name": "BaseBdev4", 00:11:18.481 "uuid": "ff131e9c-8342-44aa-ad9b-51c1db5af0d8", 00:11:18.481 "is_configured": true, 00:11:18.481 "data_offset": 2048, 00:11:18.481 "data_size": 63488 00:11:18.481 } 00:11:18.481 ] 00:11:18.481 } 00:11:18.481 } 00:11:18.481 }' 00:11:18.481 17:32:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:11:18.481 17:32:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:11:18.481 BaseBdev2 00:11:18.481 BaseBdev3 00:11:18.481 BaseBdev4' 00:11:18.481 17:32:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:18.481 17:32:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:11:18.481 17:32:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:18.481 17:32:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:18.481 17:32:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:11:18.481 17:32:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:18.481 17:32:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:18.481 17:32:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:18.481 17:32:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:18.481 17:32:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:18.481 17:32:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:18.481 17:32:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:11:18.481 17:32:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:18.481 17:32:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:18.481 17:32:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:18.481 17:32:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:18.481 17:32:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:18.481 17:32:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:18.481 17:32:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:18.481 17:32:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:18.481 17:32:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:11:18.481 17:32:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:18.481 17:32:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:18.481 17:32:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:18.481 17:32:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:18.481 17:32:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:18.481 17:32:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:18.481 17:32:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:11:18.481 17:32:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:18.481 17:32:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:18.481 17:32:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:18.481 17:32:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:18.741 17:32:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:18.741 17:32:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:18.741 17:32:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:11:18.741 17:32:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:18.741 17:32:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:18.741 [2024-11-27 17:32:49.688060] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:11:18.741 [2024-11-27 17:32:49.688098] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:11:18.741 [2024-11-27 17:32:49.688194] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:11:18.741 [2024-11-27 17:32:49.688485] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:11:18.741 [2024-11-27 17:32:49.688503] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name Existed_Raid, state offline 00:11:18.741 17:32:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:18.741 17:32:49 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 84493 00:11:18.741 17:32:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@950 -- # '[' -z 84493 ']' 00:11:18.741 17:32:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@954 -- # kill -0 84493 00:11:18.741 17:32:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@955 -- # uname 00:11:18.741 17:32:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:11:18.741 17:32:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 84493 00:11:18.741 17:32:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:11:18.741 17:32:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:11:18.741 17:32:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@968 -- # echo 'killing process with pid 84493' 00:11:18.741 killing process with pid 84493 00:11:18.741 17:32:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@969 -- # kill 84493 00:11:18.741 [2024-11-27 17:32:49.737152] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:11:18.741 17:32:49 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@974 -- # wait 84493 00:11:18.741 [2024-11-27 17:32:49.814543] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:11:19.001 17:32:50 bdev_raid.raid_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:11:19.001 00:11:19.001 real 0m9.533s 00:11:19.001 user 0m16.004s 00:11:19.001 sys 0m2.049s 00:11:19.001 17:32:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@1126 -- # xtrace_disable 00:11:19.001 17:32:50 bdev_raid.raid_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:19.001 ************************************ 00:11:19.001 END TEST raid_state_function_test_sb 00:11:19.001 ************************************ 00:11:19.261 17:32:50 bdev_raid -- bdev/bdev_raid.sh@970 -- # run_test raid_superblock_test raid_superblock_test raid1 4 00:11:19.261 17:32:50 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:11:19.261 17:32:50 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:11:19.261 17:32:50 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:11:19.261 ************************************ 00:11:19.261 START TEST raid_superblock_test 00:11:19.261 ************************************ 00:11:19.261 17:32:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1125 -- # raid_superblock_test raid1 4 00:11:19.261 17:32:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=raid1 00:11:19.261 17:32:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=4 00:11:19.261 17:32:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:11:19.261 17:32:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:11:19.261 17:32:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:11:19.261 17:32:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:11:19.261 17:32:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:11:19.261 17:32:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:11:19.261 17:32:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:11:19.261 17:32:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:11:19.261 17:32:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:11:19.261 17:32:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:11:19.261 17:32:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:11:19.261 17:32:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' raid1 '!=' raid1 ']' 00:11:19.261 17:32:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@408 -- # strip_size=0 00:11:19.261 17:32:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=85147 00:11:19.261 17:32:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:11:19.261 17:32:50 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 85147 00:11:19.261 17:32:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@831 -- # '[' -z 85147 ']' 00:11:19.261 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:11:19.261 17:32:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:11:19.261 17:32:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:11:19.261 17:32:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:11:19.261 17:32:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:11:19.261 17:32:50 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:19.261 [2024-11-27 17:32:50.345663] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:11:19.261 [2024-11-27 17:32:50.346650] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid85147 ] 00:11:19.521 [2024-11-27 17:32:50.513718] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:11:19.521 [2024-11-27 17:32:50.581849] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:11:19.521 [2024-11-27 17:32:50.658392] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:19.521 [2024-11-27 17:32:50.658433] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:20.091 17:32:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:11:20.091 17:32:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@864 -- # return 0 00:11:20.091 17:32:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:11:20.091 17:32:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:11:20.091 17:32:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:11:20.091 17:32:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:11:20.091 17:32:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:11:20.091 17:32:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:11:20.091 17:32:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:11:20.091 17:32:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:11:20.091 17:32:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:11:20.091 17:32:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:20.091 17:32:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:20.091 malloc1 00:11:20.091 17:32:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:20.091 17:32:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:11:20.091 17:32:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:20.091 17:32:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:20.091 [2024-11-27 17:32:51.193021] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:11:20.091 [2024-11-27 17:32:51.193146] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:20.091 [2024-11-27 17:32:51.193186] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:11:20.091 [2024-11-27 17:32:51.193238] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:20.091 [2024-11-27 17:32:51.195672] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:20.091 [2024-11-27 17:32:51.195748] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:11:20.091 pt1 00:11:20.092 17:32:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:20.092 17:32:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:11:20.092 17:32:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:11:20.092 17:32:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:11:20.092 17:32:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:11:20.092 17:32:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:11:20.092 17:32:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:11:20.092 17:32:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:11:20.092 17:32:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:11:20.092 17:32:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:11:20.092 17:32:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:20.092 17:32:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:20.092 malloc2 00:11:20.092 17:32:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:20.092 17:32:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:11:20.092 17:32:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:20.092 17:32:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:20.092 [2024-11-27 17:32:51.247070] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:11:20.092 [2024-11-27 17:32:51.247286] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:20.092 [2024-11-27 17:32:51.247365] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:11:20.092 [2024-11-27 17:32:51.247446] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:20.092 [2024-11-27 17:32:51.252211] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:20.092 pt2 00:11:20.092 [2024-11-27 17:32:51.252351] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:11:20.092 17:32:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:20.092 17:32:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:11:20.092 17:32:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:11:20.092 17:32:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc3 00:11:20.092 17:32:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt3 00:11:20.092 17:32:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000003 00:11:20.092 17:32:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:11:20.092 17:32:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:11:20.092 17:32:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:11:20.092 17:32:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc3 00:11:20.092 17:32:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:20.092 17:32:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:20.092 malloc3 00:11:20.092 17:32:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:20.092 17:32:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:11:20.092 17:32:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:20.092 17:32:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:20.352 [2024-11-27 17:32:51.283736] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:11:20.352 [2024-11-27 17:32:51.283851] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:20.352 [2024-11-27 17:32:51.283913] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:11:20.352 [2024-11-27 17:32:51.283945] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:20.353 [2024-11-27 17:32:51.286343] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:20.353 [2024-11-27 17:32:51.286414] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:11:20.353 pt3 00:11:20.353 17:32:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:20.353 17:32:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:11:20.353 17:32:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:11:20.353 17:32:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc4 00:11:20.353 17:32:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt4 00:11:20.353 17:32:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000004 00:11:20.353 17:32:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:11:20.353 17:32:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:11:20.353 17:32:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:11:20.353 17:32:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc4 00:11:20.353 17:32:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:20.353 17:32:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:20.353 malloc4 00:11:20.353 17:32:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:20.353 17:32:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:11:20.353 17:32:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:20.353 17:32:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:20.353 [2024-11-27 17:32:51.322133] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:11:20.353 [2024-11-27 17:32:51.322202] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:20.353 [2024-11-27 17:32:51.322218] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:11:20.353 [2024-11-27 17:32:51.322233] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:20.353 [2024-11-27 17:32:51.324641] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:20.353 [2024-11-27 17:32:51.324685] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:11:20.353 pt4 00:11:20.353 17:32:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:20.353 17:32:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:11:20.353 17:32:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:11:20.353 17:32:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''pt1 pt2 pt3 pt4'\''' -n raid_bdev1 -s 00:11:20.353 17:32:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:20.353 17:32:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:20.353 [2024-11-27 17:32:51.334181] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:11:20.353 [2024-11-27 17:32:51.336188] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:11:20.353 [2024-11-27 17:32:51.336249] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:11:20.353 [2024-11-27 17:32:51.336289] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:11:20.353 [2024-11-27 17:32:51.336436] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:11:20.353 [2024-11-27 17:32:51.336451] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:11:20.353 [2024-11-27 17:32:51.336676] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:11:20.353 [2024-11-27 17:32:51.336820] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:11:20.353 [2024-11-27 17:32:51.336830] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:11:20.353 [2024-11-27 17:32:51.336952] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:20.353 17:32:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:20.353 17:32:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:11:20.353 17:32:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:20.353 17:32:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:20.353 17:32:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:20.353 17:32:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:20.353 17:32:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:20.353 17:32:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:20.353 17:32:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:20.353 17:32:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:20.353 17:32:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:20.353 17:32:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:20.353 17:32:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:20.353 17:32:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:20.353 17:32:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:20.353 17:32:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:20.353 17:32:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:20.353 "name": "raid_bdev1", 00:11:20.353 "uuid": "273fd7ed-7d4a-49bf-990d-f56b6d0dd53d", 00:11:20.353 "strip_size_kb": 0, 00:11:20.353 "state": "online", 00:11:20.353 "raid_level": "raid1", 00:11:20.353 "superblock": true, 00:11:20.353 "num_base_bdevs": 4, 00:11:20.353 "num_base_bdevs_discovered": 4, 00:11:20.353 "num_base_bdevs_operational": 4, 00:11:20.353 "base_bdevs_list": [ 00:11:20.353 { 00:11:20.353 "name": "pt1", 00:11:20.353 "uuid": "00000000-0000-0000-0000-000000000001", 00:11:20.353 "is_configured": true, 00:11:20.353 "data_offset": 2048, 00:11:20.353 "data_size": 63488 00:11:20.353 }, 00:11:20.353 { 00:11:20.353 "name": "pt2", 00:11:20.353 "uuid": "00000000-0000-0000-0000-000000000002", 00:11:20.353 "is_configured": true, 00:11:20.353 "data_offset": 2048, 00:11:20.353 "data_size": 63488 00:11:20.353 }, 00:11:20.353 { 00:11:20.353 "name": "pt3", 00:11:20.353 "uuid": "00000000-0000-0000-0000-000000000003", 00:11:20.353 "is_configured": true, 00:11:20.353 "data_offset": 2048, 00:11:20.353 "data_size": 63488 00:11:20.353 }, 00:11:20.353 { 00:11:20.353 "name": "pt4", 00:11:20.353 "uuid": "00000000-0000-0000-0000-000000000004", 00:11:20.353 "is_configured": true, 00:11:20.353 "data_offset": 2048, 00:11:20.353 "data_size": 63488 00:11:20.353 } 00:11:20.353 ] 00:11:20.353 }' 00:11:20.353 17:32:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:20.353 17:32:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:20.614 17:32:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:11:20.614 17:32:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:11:20.614 17:32:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:11:20.614 17:32:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:11:20.614 17:32:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:11:20.614 17:32:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:11:20.614 17:32:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:11:20.614 17:32:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:11:20.614 17:32:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:20.614 17:32:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:20.614 [2024-11-27 17:32:51.789718] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:11:20.874 17:32:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:20.874 17:32:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:11:20.874 "name": "raid_bdev1", 00:11:20.874 "aliases": [ 00:11:20.874 "273fd7ed-7d4a-49bf-990d-f56b6d0dd53d" 00:11:20.874 ], 00:11:20.874 "product_name": "Raid Volume", 00:11:20.874 "block_size": 512, 00:11:20.874 "num_blocks": 63488, 00:11:20.874 "uuid": "273fd7ed-7d4a-49bf-990d-f56b6d0dd53d", 00:11:20.874 "assigned_rate_limits": { 00:11:20.874 "rw_ios_per_sec": 0, 00:11:20.874 "rw_mbytes_per_sec": 0, 00:11:20.874 "r_mbytes_per_sec": 0, 00:11:20.874 "w_mbytes_per_sec": 0 00:11:20.874 }, 00:11:20.874 "claimed": false, 00:11:20.874 "zoned": false, 00:11:20.874 "supported_io_types": { 00:11:20.874 "read": true, 00:11:20.874 "write": true, 00:11:20.874 "unmap": false, 00:11:20.874 "flush": false, 00:11:20.874 "reset": true, 00:11:20.874 "nvme_admin": false, 00:11:20.874 "nvme_io": false, 00:11:20.874 "nvme_io_md": false, 00:11:20.874 "write_zeroes": true, 00:11:20.874 "zcopy": false, 00:11:20.874 "get_zone_info": false, 00:11:20.874 "zone_management": false, 00:11:20.874 "zone_append": false, 00:11:20.874 "compare": false, 00:11:20.874 "compare_and_write": false, 00:11:20.874 "abort": false, 00:11:20.874 "seek_hole": false, 00:11:20.874 "seek_data": false, 00:11:20.874 "copy": false, 00:11:20.874 "nvme_iov_md": false 00:11:20.874 }, 00:11:20.874 "memory_domains": [ 00:11:20.874 { 00:11:20.874 "dma_device_id": "system", 00:11:20.874 "dma_device_type": 1 00:11:20.874 }, 00:11:20.874 { 00:11:20.874 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:20.874 "dma_device_type": 2 00:11:20.874 }, 00:11:20.874 { 00:11:20.874 "dma_device_id": "system", 00:11:20.874 "dma_device_type": 1 00:11:20.874 }, 00:11:20.874 { 00:11:20.874 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:20.874 "dma_device_type": 2 00:11:20.874 }, 00:11:20.874 { 00:11:20.874 "dma_device_id": "system", 00:11:20.874 "dma_device_type": 1 00:11:20.874 }, 00:11:20.874 { 00:11:20.874 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:20.874 "dma_device_type": 2 00:11:20.874 }, 00:11:20.874 { 00:11:20.874 "dma_device_id": "system", 00:11:20.874 "dma_device_type": 1 00:11:20.874 }, 00:11:20.874 { 00:11:20.874 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:20.874 "dma_device_type": 2 00:11:20.874 } 00:11:20.874 ], 00:11:20.874 "driver_specific": { 00:11:20.874 "raid": { 00:11:20.874 "uuid": "273fd7ed-7d4a-49bf-990d-f56b6d0dd53d", 00:11:20.874 "strip_size_kb": 0, 00:11:20.874 "state": "online", 00:11:20.874 "raid_level": "raid1", 00:11:20.874 "superblock": true, 00:11:20.874 "num_base_bdevs": 4, 00:11:20.874 "num_base_bdevs_discovered": 4, 00:11:20.874 "num_base_bdevs_operational": 4, 00:11:20.874 "base_bdevs_list": [ 00:11:20.874 { 00:11:20.874 "name": "pt1", 00:11:20.874 "uuid": "00000000-0000-0000-0000-000000000001", 00:11:20.874 "is_configured": true, 00:11:20.874 "data_offset": 2048, 00:11:20.874 "data_size": 63488 00:11:20.874 }, 00:11:20.874 { 00:11:20.874 "name": "pt2", 00:11:20.874 "uuid": "00000000-0000-0000-0000-000000000002", 00:11:20.874 "is_configured": true, 00:11:20.874 "data_offset": 2048, 00:11:20.874 "data_size": 63488 00:11:20.874 }, 00:11:20.874 { 00:11:20.874 "name": "pt3", 00:11:20.874 "uuid": "00000000-0000-0000-0000-000000000003", 00:11:20.874 "is_configured": true, 00:11:20.874 "data_offset": 2048, 00:11:20.874 "data_size": 63488 00:11:20.874 }, 00:11:20.874 { 00:11:20.874 "name": "pt4", 00:11:20.874 "uuid": "00000000-0000-0000-0000-000000000004", 00:11:20.874 "is_configured": true, 00:11:20.874 "data_offset": 2048, 00:11:20.874 "data_size": 63488 00:11:20.874 } 00:11:20.874 ] 00:11:20.874 } 00:11:20.874 } 00:11:20.874 }' 00:11:20.874 17:32:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:11:20.874 17:32:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:11:20.874 pt2 00:11:20.874 pt3 00:11:20.874 pt4' 00:11:20.874 17:32:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:20.874 17:32:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:11:20.874 17:32:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:20.874 17:32:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:11:20.874 17:32:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:20.874 17:32:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:20.874 17:32:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:20.874 17:32:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:20.874 17:32:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:20.874 17:32:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:20.874 17:32:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:20.874 17:32:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:11:20.874 17:32:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:20.874 17:32:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:20.874 17:32:51 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:20.874 17:32:51 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:20.874 17:32:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:20.874 17:32:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:20.874 17:32:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:20.874 17:32:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:11:20.874 17:32:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:20.874 17:32:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:20.874 17:32:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:20.874 17:32:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:20.874 17:32:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:20.874 17:32:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:20.874 17:32:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:20.874 17:32:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt4 00:11:20.874 17:32:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:20.874 17:32:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:20.874 17:32:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:21.135 17:32:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:21.135 17:32:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:21.135 17:32:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:21.135 17:32:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:11:21.135 17:32:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:11:21.135 17:32:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:21.135 17:32:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:21.135 [2024-11-27 17:32:52.117031] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:11:21.135 17:32:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:21.135 17:32:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=273fd7ed-7d4a-49bf-990d-f56b6d0dd53d 00:11:21.135 17:32:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z 273fd7ed-7d4a-49bf-990d-f56b6d0dd53d ']' 00:11:21.135 17:32:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:11:21.135 17:32:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:21.135 17:32:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:21.135 [2024-11-27 17:32:52.164689] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:11:21.135 [2024-11-27 17:32:52.164757] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:11:21.135 [2024-11-27 17:32:52.164879] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:11:21.135 [2024-11-27 17:32:52.165019] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:11:21.135 [2024-11-27 17:32:52.165063] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:11:21.135 17:32:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:21.135 17:32:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:21.135 17:32:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:21.135 17:32:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:21.135 17:32:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:11:21.135 17:32:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:21.135 17:32:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:11:21.135 17:32:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:11:21.135 17:32:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:11:21.135 17:32:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:11:21.135 17:32:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:21.135 17:32:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:21.135 17:32:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:21.135 17:32:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:11:21.135 17:32:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:11:21.135 17:32:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:21.135 17:32:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:21.135 17:32:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:21.135 17:32:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:11:21.135 17:32:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt3 00:11:21.135 17:32:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:21.135 17:32:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:21.135 17:32:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:21.135 17:32:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:11:21.135 17:32:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt4 00:11:21.135 17:32:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:21.135 17:32:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:21.135 17:32:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:21.135 17:32:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:11:21.135 17:32:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:11:21.135 17:32:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:21.135 17:32:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:21.135 17:32:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:21.135 17:32:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:11:21.135 17:32:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:11:21.135 17:32:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@650 -- # local es=0 00:11:21.135 17:32:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:11:21.135 17:32:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:11:21.135 17:32:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:11:21.135 17:32:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:11:21.135 17:32:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:11:21.135 17:32:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:11:21.135 17:32:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:21.135 17:32:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:21.395 [2024-11-27 17:32:52.328427] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:11:21.395 [2024-11-27 17:32:52.330642] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:11:21.395 [2024-11-27 17:32:52.330771] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc3 is claimed 00:11:21.395 [2024-11-27 17:32:52.330815] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc4 is claimed 00:11:21.395 [2024-11-27 17:32:52.330874] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:11:21.395 [2024-11-27 17:32:52.330920] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:11:21.395 [2024-11-27 17:32:52.330940] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc3 00:11:21.395 [2024-11-27 17:32:52.330957] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc4 00:11:21.395 [2024-11-27 17:32:52.330972] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:11:21.395 [2024-11-27 17:32:52.330983] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state configuring 00:11:21.395 request: 00:11:21.395 { 00:11:21.395 "name": "raid_bdev1", 00:11:21.395 "raid_level": "raid1", 00:11:21.395 "base_bdevs": [ 00:11:21.395 "malloc1", 00:11:21.395 "malloc2", 00:11:21.395 "malloc3", 00:11:21.395 "malloc4" 00:11:21.395 ], 00:11:21.395 "superblock": false, 00:11:21.395 "method": "bdev_raid_create", 00:11:21.395 "req_id": 1 00:11:21.395 } 00:11:21.395 Got JSON-RPC error response 00:11:21.395 response: 00:11:21.395 { 00:11:21.395 "code": -17, 00:11:21.395 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:11:21.395 } 00:11:21.395 17:32:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:11:21.395 17:32:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@653 -- # es=1 00:11:21.395 17:32:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:11:21.395 17:32:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:11:21.395 17:32:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:11:21.395 17:32:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:11:21.395 17:32:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:21.395 17:32:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:21.395 17:32:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:21.395 17:32:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:21.395 17:32:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:11:21.395 17:32:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:11:21.396 17:32:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:11:21.396 17:32:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:21.396 17:32:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:21.396 [2024-11-27 17:32:52.380288] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:11:21.396 [2024-11-27 17:32:52.380368] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:21.396 [2024-11-27 17:32:52.380408] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:11:21.396 [2024-11-27 17:32:52.380435] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:21.396 [2024-11-27 17:32:52.382959] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:21.396 [2024-11-27 17:32:52.383026] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:11:21.396 [2024-11-27 17:32:52.383136] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:11:21.396 [2024-11-27 17:32:52.383215] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:11:21.396 pt1 00:11:21.396 17:32:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:21.396 17:32:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 4 00:11:21.396 17:32:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:21.396 17:32:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:21.396 17:32:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:21.396 17:32:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:21.396 17:32:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:21.396 17:32:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:21.396 17:32:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:21.396 17:32:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:21.396 17:32:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:21.396 17:32:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:21.396 17:32:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:21.396 17:32:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:21.396 17:32:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:21.396 17:32:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:21.396 17:32:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:21.396 "name": "raid_bdev1", 00:11:21.396 "uuid": "273fd7ed-7d4a-49bf-990d-f56b6d0dd53d", 00:11:21.396 "strip_size_kb": 0, 00:11:21.396 "state": "configuring", 00:11:21.396 "raid_level": "raid1", 00:11:21.396 "superblock": true, 00:11:21.396 "num_base_bdevs": 4, 00:11:21.396 "num_base_bdevs_discovered": 1, 00:11:21.396 "num_base_bdevs_operational": 4, 00:11:21.396 "base_bdevs_list": [ 00:11:21.396 { 00:11:21.396 "name": "pt1", 00:11:21.396 "uuid": "00000000-0000-0000-0000-000000000001", 00:11:21.396 "is_configured": true, 00:11:21.396 "data_offset": 2048, 00:11:21.396 "data_size": 63488 00:11:21.396 }, 00:11:21.396 { 00:11:21.396 "name": null, 00:11:21.396 "uuid": "00000000-0000-0000-0000-000000000002", 00:11:21.396 "is_configured": false, 00:11:21.396 "data_offset": 2048, 00:11:21.396 "data_size": 63488 00:11:21.396 }, 00:11:21.396 { 00:11:21.396 "name": null, 00:11:21.396 "uuid": "00000000-0000-0000-0000-000000000003", 00:11:21.396 "is_configured": false, 00:11:21.396 "data_offset": 2048, 00:11:21.396 "data_size": 63488 00:11:21.396 }, 00:11:21.396 { 00:11:21.396 "name": null, 00:11:21.396 "uuid": "00000000-0000-0000-0000-000000000004", 00:11:21.396 "is_configured": false, 00:11:21.396 "data_offset": 2048, 00:11:21.396 "data_size": 63488 00:11:21.396 } 00:11:21.396 ] 00:11:21.396 }' 00:11:21.396 17:32:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:21.396 17:32:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:21.965 17:32:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 4 -gt 2 ']' 00:11:21.965 17:32:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@472 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:11:21.965 17:32:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:21.965 17:32:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:21.965 [2024-11-27 17:32:52.863509] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:11:21.965 [2024-11-27 17:32:52.863608] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:21.965 [2024-11-27 17:32:52.863657] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009c80 00:11:21.965 [2024-11-27 17:32:52.863687] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:21.965 [2024-11-27 17:32:52.864188] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:21.965 [2024-11-27 17:32:52.864242] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:11:21.965 [2024-11-27 17:32:52.864352] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:11:21.965 [2024-11-27 17:32:52.864416] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:11:21.965 pt2 00:11:21.965 17:32:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:21.965 17:32:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@473 -- # rpc_cmd bdev_passthru_delete pt2 00:11:21.965 17:32:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:21.965 17:32:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:21.965 [2024-11-27 17:32:52.871510] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: pt2 00:11:21.965 17:32:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:21.965 17:32:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@474 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 4 00:11:21.965 17:32:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:21.965 17:32:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:21.965 17:32:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:21.965 17:32:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:21.965 17:32:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:21.965 17:32:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:21.965 17:32:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:21.965 17:32:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:21.965 17:32:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:21.965 17:32:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:21.965 17:32:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:21.965 17:32:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:21.965 17:32:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:21.965 17:32:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:21.965 17:32:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:21.965 "name": "raid_bdev1", 00:11:21.965 "uuid": "273fd7ed-7d4a-49bf-990d-f56b6d0dd53d", 00:11:21.965 "strip_size_kb": 0, 00:11:21.965 "state": "configuring", 00:11:21.965 "raid_level": "raid1", 00:11:21.965 "superblock": true, 00:11:21.965 "num_base_bdevs": 4, 00:11:21.965 "num_base_bdevs_discovered": 1, 00:11:21.965 "num_base_bdevs_operational": 4, 00:11:21.965 "base_bdevs_list": [ 00:11:21.965 { 00:11:21.965 "name": "pt1", 00:11:21.965 "uuid": "00000000-0000-0000-0000-000000000001", 00:11:21.965 "is_configured": true, 00:11:21.965 "data_offset": 2048, 00:11:21.965 "data_size": 63488 00:11:21.965 }, 00:11:21.965 { 00:11:21.965 "name": null, 00:11:21.965 "uuid": "00000000-0000-0000-0000-000000000002", 00:11:21.965 "is_configured": false, 00:11:21.965 "data_offset": 0, 00:11:21.965 "data_size": 63488 00:11:21.965 }, 00:11:21.965 { 00:11:21.965 "name": null, 00:11:21.965 "uuid": "00000000-0000-0000-0000-000000000003", 00:11:21.965 "is_configured": false, 00:11:21.965 "data_offset": 2048, 00:11:21.965 "data_size": 63488 00:11:21.965 }, 00:11:21.965 { 00:11:21.965 "name": null, 00:11:21.965 "uuid": "00000000-0000-0000-0000-000000000004", 00:11:21.965 "is_configured": false, 00:11:21.965 "data_offset": 2048, 00:11:21.965 "data_size": 63488 00:11:21.965 } 00:11:21.965 ] 00:11:21.965 }' 00:11:21.965 17:32:52 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:21.965 17:32:52 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:22.225 17:32:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:11:22.225 17:32:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:11:22.225 17:32:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:11:22.225 17:32:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:22.225 17:32:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:22.225 [2024-11-27 17:32:53.334797] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:11:22.225 [2024-11-27 17:32:53.334882] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:22.225 [2024-11-27 17:32:53.334907] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009f80 00:11:22.225 [2024-11-27 17:32:53.334921] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:22.225 [2024-11-27 17:32:53.335434] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:22.225 [2024-11-27 17:32:53.335462] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:11:22.225 [2024-11-27 17:32:53.335553] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:11:22.225 [2024-11-27 17:32:53.335583] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:11:22.225 pt2 00:11:22.225 17:32:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:22.225 17:32:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:11:22.225 17:32:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:11:22.225 17:32:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:11:22.225 17:32:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:22.225 17:32:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:22.225 [2024-11-27 17:32:53.346684] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:11:22.225 [2024-11-27 17:32:53.346753] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:22.225 [2024-11-27 17:32:53.346782] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:11:22.225 [2024-11-27 17:32:53.346794] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:22.225 [2024-11-27 17:32:53.347207] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:22.225 [2024-11-27 17:32:53.347227] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:11:22.225 [2024-11-27 17:32:53.347295] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:11:22.225 [2024-11-27 17:32:53.347318] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:11:22.225 pt3 00:11:22.225 17:32:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:22.225 17:32:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:11:22.225 17:32:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:11:22.225 17:32:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:11:22.225 17:32:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:22.225 17:32:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:22.225 [2024-11-27 17:32:53.358664] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:11:22.225 [2024-11-27 17:32:53.358740] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:22.225 [2024-11-27 17:32:53.358758] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a580 00:11:22.225 [2024-11-27 17:32:53.358768] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:22.225 [2024-11-27 17:32:53.359100] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:22.225 [2024-11-27 17:32:53.359117] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:11:22.225 [2024-11-27 17:32:53.359188] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt4 00:11:22.225 [2024-11-27 17:32:53.359210] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:11:22.225 [2024-11-27 17:32:53.359329] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:11:22.225 [2024-11-27 17:32:53.359343] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:11:22.225 [2024-11-27 17:32:53.359580] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:11:22.225 [2024-11-27 17:32:53.359739] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:11:22.225 [2024-11-27 17:32:53.359749] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:11:22.225 [2024-11-27 17:32:53.359856] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:22.225 pt4 00:11:22.225 17:32:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:22.225 17:32:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:11:22.225 17:32:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:11:22.225 17:32:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:11:22.225 17:32:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:22.225 17:32:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:22.225 17:32:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:22.225 17:32:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:22.225 17:32:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:22.225 17:32:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:22.225 17:32:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:22.225 17:32:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:22.225 17:32:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:22.225 17:32:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:22.225 17:32:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:22.225 17:32:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:22.225 17:32:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:22.225 17:32:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:22.485 17:32:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:22.485 "name": "raid_bdev1", 00:11:22.485 "uuid": "273fd7ed-7d4a-49bf-990d-f56b6d0dd53d", 00:11:22.485 "strip_size_kb": 0, 00:11:22.485 "state": "online", 00:11:22.485 "raid_level": "raid1", 00:11:22.485 "superblock": true, 00:11:22.485 "num_base_bdevs": 4, 00:11:22.485 "num_base_bdevs_discovered": 4, 00:11:22.485 "num_base_bdevs_operational": 4, 00:11:22.485 "base_bdevs_list": [ 00:11:22.485 { 00:11:22.485 "name": "pt1", 00:11:22.485 "uuid": "00000000-0000-0000-0000-000000000001", 00:11:22.485 "is_configured": true, 00:11:22.485 "data_offset": 2048, 00:11:22.485 "data_size": 63488 00:11:22.485 }, 00:11:22.485 { 00:11:22.485 "name": "pt2", 00:11:22.485 "uuid": "00000000-0000-0000-0000-000000000002", 00:11:22.485 "is_configured": true, 00:11:22.485 "data_offset": 2048, 00:11:22.485 "data_size": 63488 00:11:22.485 }, 00:11:22.485 { 00:11:22.485 "name": "pt3", 00:11:22.485 "uuid": "00000000-0000-0000-0000-000000000003", 00:11:22.485 "is_configured": true, 00:11:22.485 "data_offset": 2048, 00:11:22.485 "data_size": 63488 00:11:22.485 }, 00:11:22.485 { 00:11:22.485 "name": "pt4", 00:11:22.485 "uuid": "00000000-0000-0000-0000-000000000004", 00:11:22.485 "is_configured": true, 00:11:22.485 "data_offset": 2048, 00:11:22.485 "data_size": 63488 00:11:22.485 } 00:11:22.485 ] 00:11:22.485 }' 00:11:22.485 17:32:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:22.485 17:32:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:22.744 17:32:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:11:22.745 17:32:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:11:22.745 17:32:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:11:22.745 17:32:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:11:22.745 17:32:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:11:22.745 17:32:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:11:22.745 17:32:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:11:22.745 17:32:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:22.745 17:32:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:22.745 17:32:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:11:22.745 [2024-11-27 17:32:53.830300] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:11:22.745 17:32:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:22.745 17:32:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:11:22.745 "name": "raid_bdev1", 00:11:22.745 "aliases": [ 00:11:22.745 "273fd7ed-7d4a-49bf-990d-f56b6d0dd53d" 00:11:22.745 ], 00:11:22.745 "product_name": "Raid Volume", 00:11:22.745 "block_size": 512, 00:11:22.745 "num_blocks": 63488, 00:11:22.745 "uuid": "273fd7ed-7d4a-49bf-990d-f56b6d0dd53d", 00:11:22.745 "assigned_rate_limits": { 00:11:22.745 "rw_ios_per_sec": 0, 00:11:22.745 "rw_mbytes_per_sec": 0, 00:11:22.745 "r_mbytes_per_sec": 0, 00:11:22.745 "w_mbytes_per_sec": 0 00:11:22.745 }, 00:11:22.745 "claimed": false, 00:11:22.745 "zoned": false, 00:11:22.745 "supported_io_types": { 00:11:22.745 "read": true, 00:11:22.745 "write": true, 00:11:22.745 "unmap": false, 00:11:22.745 "flush": false, 00:11:22.745 "reset": true, 00:11:22.745 "nvme_admin": false, 00:11:22.745 "nvme_io": false, 00:11:22.745 "nvme_io_md": false, 00:11:22.745 "write_zeroes": true, 00:11:22.745 "zcopy": false, 00:11:22.745 "get_zone_info": false, 00:11:22.745 "zone_management": false, 00:11:22.745 "zone_append": false, 00:11:22.745 "compare": false, 00:11:22.745 "compare_and_write": false, 00:11:22.745 "abort": false, 00:11:22.745 "seek_hole": false, 00:11:22.745 "seek_data": false, 00:11:22.745 "copy": false, 00:11:22.745 "nvme_iov_md": false 00:11:22.745 }, 00:11:22.745 "memory_domains": [ 00:11:22.745 { 00:11:22.745 "dma_device_id": "system", 00:11:22.745 "dma_device_type": 1 00:11:22.745 }, 00:11:22.745 { 00:11:22.745 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:22.745 "dma_device_type": 2 00:11:22.745 }, 00:11:22.745 { 00:11:22.745 "dma_device_id": "system", 00:11:22.745 "dma_device_type": 1 00:11:22.745 }, 00:11:22.745 { 00:11:22.745 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:22.745 "dma_device_type": 2 00:11:22.745 }, 00:11:22.745 { 00:11:22.745 "dma_device_id": "system", 00:11:22.745 "dma_device_type": 1 00:11:22.745 }, 00:11:22.745 { 00:11:22.745 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:22.745 "dma_device_type": 2 00:11:22.745 }, 00:11:22.745 { 00:11:22.745 "dma_device_id": "system", 00:11:22.745 "dma_device_type": 1 00:11:22.745 }, 00:11:22.745 { 00:11:22.745 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:11:22.745 "dma_device_type": 2 00:11:22.745 } 00:11:22.745 ], 00:11:22.745 "driver_specific": { 00:11:22.745 "raid": { 00:11:22.745 "uuid": "273fd7ed-7d4a-49bf-990d-f56b6d0dd53d", 00:11:22.745 "strip_size_kb": 0, 00:11:22.745 "state": "online", 00:11:22.745 "raid_level": "raid1", 00:11:22.745 "superblock": true, 00:11:22.745 "num_base_bdevs": 4, 00:11:22.745 "num_base_bdevs_discovered": 4, 00:11:22.745 "num_base_bdevs_operational": 4, 00:11:22.745 "base_bdevs_list": [ 00:11:22.745 { 00:11:22.745 "name": "pt1", 00:11:22.745 "uuid": "00000000-0000-0000-0000-000000000001", 00:11:22.745 "is_configured": true, 00:11:22.745 "data_offset": 2048, 00:11:22.745 "data_size": 63488 00:11:22.745 }, 00:11:22.745 { 00:11:22.745 "name": "pt2", 00:11:22.745 "uuid": "00000000-0000-0000-0000-000000000002", 00:11:22.745 "is_configured": true, 00:11:22.745 "data_offset": 2048, 00:11:22.745 "data_size": 63488 00:11:22.745 }, 00:11:22.745 { 00:11:22.745 "name": "pt3", 00:11:22.745 "uuid": "00000000-0000-0000-0000-000000000003", 00:11:22.745 "is_configured": true, 00:11:22.745 "data_offset": 2048, 00:11:22.745 "data_size": 63488 00:11:22.745 }, 00:11:22.745 { 00:11:22.745 "name": "pt4", 00:11:22.745 "uuid": "00000000-0000-0000-0000-000000000004", 00:11:22.745 "is_configured": true, 00:11:22.745 "data_offset": 2048, 00:11:22.745 "data_size": 63488 00:11:22.745 } 00:11:22.745 ] 00:11:22.745 } 00:11:22.745 } 00:11:22.745 }' 00:11:22.745 17:32:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:11:22.745 17:32:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:11:22.745 pt2 00:11:22.745 pt3 00:11:22.745 pt4' 00:11:22.745 17:32:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:23.005 17:32:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:11:23.005 17:32:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:23.005 17:32:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:23.005 17:32:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:11:23.005 17:32:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:23.005 17:32:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:23.005 17:32:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:23.005 17:32:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:23.005 17:32:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:23.005 17:32:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:23.005 17:32:53 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:11:23.005 17:32:53 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:23.005 17:32:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:23.005 17:32:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:23.005 17:32:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:23.005 17:32:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:23.006 17:32:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:23.006 17:32:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:23.006 17:32:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:11:23.006 17:32:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:23.006 17:32:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:23.006 17:32:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:23.006 17:32:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:23.006 17:32:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:23.006 17:32:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:23.006 17:32:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:11:23.006 17:32:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt4 00:11:23.006 17:32:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:11:23.006 17:32:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:23.006 17:32:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:23.006 17:32:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:23.006 17:32:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:11:23.006 17:32:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:11:23.006 17:32:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:11:23.006 17:32:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:23.006 17:32:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:23.006 17:32:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:11:23.006 [2024-11-27 17:32:54.145656] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:11:23.006 17:32:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:23.006 17:32:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' 273fd7ed-7d4a-49bf-990d-f56b6d0dd53d '!=' 273fd7ed-7d4a-49bf-990d-f56b6d0dd53d ']' 00:11:23.006 17:32:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy raid1 00:11:23.006 17:32:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:11:23.006 17:32:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@199 -- # return 0 00:11:23.006 17:32:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@493 -- # rpc_cmd bdev_passthru_delete pt1 00:11:23.006 17:32:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:23.006 17:32:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:23.006 [2024-11-27 17:32:54.193307] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: pt1 00:11:23.265 17:32:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:23.265 17:32:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@496 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:11:23.265 17:32:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:23.265 17:32:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:23.265 17:32:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:23.265 17:32:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:23.265 17:32:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:11:23.265 17:32:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:23.265 17:32:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:23.265 17:32:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:23.265 17:32:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:23.265 17:32:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:23.265 17:32:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:23.265 17:32:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:23.265 17:32:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:23.265 17:32:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:23.265 17:32:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:23.265 "name": "raid_bdev1", 00:11:23.265 "uuid": "273fd7ed-7d4a-49bf-990d-f56b6d0dd53d", 00:11:23.265 "strip_size_kb": 0, 00:11:23.265 "state": "online", 00:11:23.265 "raid_level": "raid1", 00:11:23.265 "superblock": true, 00:11:23.265 "num_base_bdevs": 4, 00:11:23.265 "num_base_bdevs_discovered": 3, 00:11:23.265 "num_base_bdevs_operational": 3, 00:11:23.265 "base_bdevs_list": [ 00:11:23.265 { 00:11:23.265 "name": null, 00:11:23.265 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:23.265 "is_configured": false, 00:11:23.265 "data_offset": 0, 00:11:23.265 "data_size": 63488 00:11:23.265 }, 00:11:23.266 { 00:11:23.266 "name": "pt2", 00:11:23.266 "uuid": "00000000-0000-0000-0000-000000000002", 00:11:23.266 "is_configured": true, 00:11:23.266 "data_offset": 2048, 00:11:23.266 "data_size": 63488 00:11:23.266 }, 00:11:23.266 { 00:11:23.266 "name": "pt3", 00:11:23.266 "uuid": "00000000-0000-0000-0000-000000000003", 00:11:23.266 "is_configured": true, 00:11:23.266 "data_offset": 2048, 00:11:23.266 "data_size": 63488 00:11:23.266 }, 00:11:23.266 { 00:11:23.266 "name": "pt4", 00:11:23.266 "uuid": "00000000-0000-0000-0000-000000000004", 00:11:23.266 "is_configured": true, 00:11:23.266 "data_offset": 2048, 00:11:23.266 "data_size": 63488 00:11:23.266 } 00:11:23.266 ] 00:11:23.266 }' 00:11:23.266 17:32:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:23.266 17:32:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:23.525 17:32:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@499 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:11:23.525 17:32:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:23.525 17:32:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:23.525 [2024-11-27 17:32:54.640500] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:11:23.525 [2024-11-27 17:32:54.640569] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:11:23.525 [2024-11-27 17:32:54.640676] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:11:23.525 [2024-11-27 17:32:54.640769] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:11:23.525 [2024-11-27 17:32:54.640816] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:11:23.525 17:32:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:23.525 17:32:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:23.525 17:32:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:23.525 17:32:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:23.525 17:32:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # jq -r '.[]' 00:11:23.525 17:32:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:23.525 17:32:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@500 -- # raid_bdev= 00:11:23.525 17:32:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@501 -- # '[' -n '' ']' 00:11:23.525 17:32:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i = 1 )) 00:11:23.525 17:32:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:11:23.525 17:32:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt2 00:11:23.525 17:32:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:23.525 17:32:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:23.525 17:32:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:23.525 17:32:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:11:23.525 17:32:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:11:23.525 17:32:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt3 00:11:23.525 17:32:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:23.525 17:32:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:23.786 17:32:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:23.786 17:32:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:11:23.786 17:32:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:11:23.786 17:32:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt4 00:11:23.786 17:32:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:23.786 17:32:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:23.786 17:32:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:23.786 17:32:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:11:23.786 17:32:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:11:23.786 17:32:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i = 1 )) 00:11:23.786 17:32:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:11:23.786 17:32:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@512 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:11:23.786 17:32:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:23.786 17:32:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:23.786 [2024-11-27 17:32:54.736321] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:11:23.786 [2024-11-27 17:32:54.736374] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:23.786 [2024-11-27 17:32:54.736411] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a880 00:11:23.786 [2024-11-27 17:32:54.736423] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:23.786 [2024-11-27 17:32:54.739001] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:23.786 [2024-11-27 17:32:54.739039] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:11:23.786 [2024-11-27 17:32:54.739114] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:11:23.786 [2024-11-27 17:32:54.739164] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:11:23.786 pt2 00:11:23.786 17:32:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:23.786 17:32:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@515 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 3 00:11:23.786 17:32:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:23.786 17:32:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:23.786 17:32:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:23.786 17:32:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:23.786 17:32:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:11:23.786 17:32:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:23.786 17:32:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:23.786 17:32:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:23.786 17:32:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:23.786 17:32:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:23.786 17:32:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:23.786 17:32:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:23.786 17:32:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:23.786 17:32:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:23.786 17:32:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:23.786 "name": "raid_bdev1", 00:11:23.786 "uuid": "273fd7ed-7d4a-49bf-990d-f56b6d0dd53d", 00:11:23.786 "strip_size_kb": 0, 00:11:23.786 "state": "configuring", 00:11:23.786 "raid_level": "raid1", 00:11:23.786 "superblock": true, 00:11:23.786 "num_base_bdevs": 4, 00:11:23.786 "num_base_bdevs_discovered": 1, 00:11:23.786 "num_base_bdevs_operational": 3, 00:11:23.786 "base_bdevs_list": [ 00:11:23.786 { 00:11:23.786 "name": null, 00:11:23.786 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:23.786 "is_configured": false, 00:11:23.786 "data_offset": 2048, 00:11:23.786 "data_size": 63488 00:11:23.786 }, 00:11:23.786 { 00:11:23.786 "name": "pt2", 00:11:23.786 "uuid": "00000000-0000-0000-0000-000000000002", 00:11:23.786 "is_configured": true, 00:11:23.786 "data_offset": 2048, 00:11:23.786 "data_size": 63488 00:11:23.786 }, 00:11:23.786 { 00:11:23.786 "name": null, 00:11:23.786 "uuid": "00000000-0000-0000-0000-000000000003", 00:11:23.786 "is_configured": false, 00:11:23.786 "data_offset": 2048, 00:11:23.786 "data_size": 63488 00:11:23.786 }, 00:11:23.786 { 00:11:23.786 "name": null, 00:11:23.786 "uuid": "00000000-0000-0000-0000-000000000004", 00:11:23.786 "is_configured": false, 00:11:23.786 "data_offset": 2048, 00:11:23.786 "data_size": 63488 00:11:23.786 } 00:11:23.786 ] 00:11:23.786 }' 00:11:23.786 17:32:54 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:23.786 17:32:54 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:24.046 17:32:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i++ )) 00:11:24.046 17:32:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:11:24.046 17:32:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@512 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:11:24.046 17:32:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:24.046 17:32:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:24.046 [2024-11-27 17:32:55.163624] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:11:24.046 [2024-11-27 17:32:55.163752] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:24.046 [2024-11-27 17:32:55.163791] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ae80 00:11:24.046 [2024-11-27 17:32:55.163827] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:24.046 [2024-11-27 17:32:55.164313] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:24.046 [2024-11-27 17:32:55.164372] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:11:24.046 [2024-11-27 17:32:55.164482] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:11:24.046 [2024-11-27 17:32:55.164545] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:11:24.046 pt3 00:11:24.046 17:32:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:24.046 17:32:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@515 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 3 00:11:24.046 17:32:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:24.046 17:32:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:24.046 17:32:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:24.046 17:32:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:24.046 17:32:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:11:24.046 17:32:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:24.046 17:32:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:24.046 17:32:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:24.046 17:32:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:24.046 17:32:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:24.046 17:32:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:24.046 17:32:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:24.046 17:32:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:24.046 17:32:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:24.046 17:32:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:24.046 "name": "raid_bdev1", 00:11:24.046 "uuid": "273fd7ed-7d4a-49bf-990d-f56b6d0dd53d", 00:11:24.046 "strip_size_kb": 0, 00:11:24.046 "state": "configuring", 00:11:24.046 "raid_level": "raid1", 00:11:24.046 "superblock": true, 00:11:24.046 "num_base_bdevs": 4, 00:11:24.046 "num_base_bdevs_discovered": 2, 00:11:24.046 "num_base_bdevs_operational": 3, 00:11:24.046 "base_bdevs_list": [ 00:11:24.046 { 00:11:24.046 "name": null, 00:11:24.046 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:24.046 "is_configured": false, 00:11:24.046 "data_offset": 2048, 00:11:24.046 "data_size": 63488 00:11:24.046 }, 00:11:24.046 { 00:11:24.046 "name": "pt2", 00:11:24.046 "uuid": "00000000-0000-0000-0000-000000000002", 00:11:24.046 "is_configured": true, 00:11:24.046 "data_offset": 2048, 00:11:24.046 "data_size": 63488 00:11:24.046 }, 00:11:24.046 { 00:11:24.046 "name": "pt3", 00:11:24.046 "uuid": "00000000-0000-0000-0000-000000000003", 00:11:24.046 "is_configured": true, 00:11:24.046 "data_offset": 2048, 00:11:24.046 "data_size": 63488 00:11:24.046 }, 00:11:24.046 { 00:11:24.046 "name": null, 00:11:24.046 "uuid": "00000000-0000-0000-0000-000000000004", 00:11:24.046 "is_configured": false, 00:11:24.046 "data_offset": 2048, 00:11:24.046 "data_size": 63488 00:11:24.046 } 00:11:24.046 ] 00:11:24.046 }' 00:11:24.046 17:32:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:24.046 17:32:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:24.616 17:32:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i++ )) 00:11:24.616 17:32:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:11:24.616 17:32:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@519 -- # i=3 00:11:24.616 17:32:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@520 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:11:24.616 17:32:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:24.616 17:32:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:24.616 [2024-11-27 17:32:55.610870] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:11:24.616 [2024-11-27 17:32:55.610951] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:24.616 [2024-11-27 17:32:55.610976] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b180 00:11:24.616 [2024-11-27 17:32:55.610987] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:24.616 [2024-11-27 17:32:55.611467] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:24.616 [2024-11-27 17:32:55.611488] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:11:24.616 [2024-11-27 17:32:55.611568] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt4 00:11:24.616 [2024-11-27 17:32:55.611594] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:11:24.616 [2024-11-27 17:32:55.611702] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:11:24.616 [2024-11-27 17:32:55.611714] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:11:24.616 [2024-11-27 17:32:55.611966] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000026d0 00:11:24.616 [2024-11-27 17:32:55.612102] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:11:24.616 [2024-11-27 17:32:55.612112] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001c80 00:11:24.616 [2024-11-27 17:32:55.612238] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:24.616 pt4 00:11:24.616 17:32:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:24.616 17:32:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@523 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:11:24.616 17:32:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:24.616 17:32:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:24.616 17:32:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:24.616 17:32:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:24.616 17:32:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:11:24.616 17:32:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:24.616 17:32:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:24.616 17:32:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:24.616 17:32:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:24.616 17:32:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:24.616 17:32:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:24.616 17:32:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:24.616 17:32:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:24.616 17:32:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:24.616 17:32:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:24.616 "name": "raid_bdev1", 00:11:24.616 "uuid": "273fd7ed-7d4a-49bf-990d-f56b6d0dd53d", 00:11:24.616 "strip_size_kb": 0, 00:11:24.616 "state": "online", 00:11:24.616 "raid_level": "raid1", 00:11:24.616 "superblock": true, 00:11:24.616 "num_base_bdevs": 4, 00:11:24.616 "num_base_bdevs_discovered": 3, 00:11:24.616 "num_base_bdevs_operational": 3, 00:11:24.616 "base_bdevs_list": [ 00:11:24.616 { 00:11:24.616 "name": null, 00:11:24.616 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:24.616 "is_configured": false, 00:11:24.616 "data_offset": 2048, 00:11:24.616 "data_size": 63488 00:11:24.616 }, 00:11:24.616 { 00:11:24.616 "name": "pt2", 00:11:24.616 "uuid": "00000000-0000-0000-0000-000000000002", 00:11:24.616 "is_configured": true, 00:11:24.616 "data_offset": 2048, 00:11:24.616 "data_size": 63488 00:11:24.616 }, 00:11:24.616 { 00:11:24.616 "name": "pt3", 00:11:24.616 "uuid": "00000000-0000-0000-0000-000000000003", 00:11:24.616 "is_configured": true, 00:11:24.616 "data_offset": 2048, 00:11:24.616 "data_size": 63488 00:11:24.616 }, 00:11:24.616 { 00:11:24.616 "name": "pt4", 00:11:24.616 "uuid": "00000000-0000-0000-0000-000000000004", 00:11:24.616 "is_configured": true, 00:11:24.616 "data_offset": 2048, 00:11:24.616 "data_size": 63488 00:11:24.616 } 00:11:24.616 ] 00:11:24.616 }' 00:11:24.616 17:32:55 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:24.616 17:32:55 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:24.877 17:32:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@526 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:11:24.877 17:32:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:24.877 17:32:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:24.877 [2024-11-27 17:32:56.058236] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:11:24.877 [2024-11-27 17:32:56.058327] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:11:24.877 [2024-11-27 17:32:56.058469] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:11:24.877 [2024-11-27 17:32:56.058577] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:11:24.877 [2024-11-27 17:32:56.058627] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name raid_bdev1, state offline 00:11:24.877 17:32:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:24.877 17:32:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # jq -r '.[]' 00:11:24.877 17:32:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:25.137 17:32:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:25.137 17:32:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:25.137 17:32:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:25.137 17:32:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@527 -- # raid_bdev= 00:11:25.137 17:32:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@528 -- # '[' -n '' ']' 00:11:25.137 17:32:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@532 -- # '[' 4 -gt 2 ']' 00:11:25.137 17:32:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@534 -- # i=3 00:11:25.137 17:32:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@535 -- # rpc_cmd bdev_passthru_delete pt4 00:11:25.137 17:32:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:25.137 17:32:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:25.137 17:32:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:25.137 17:32:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@540 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:11:25.137 17:32:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:25.137 17:32:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:25.137 [2024-11-27 17:32:56.114052] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:11:25.137 [2024-11-27 17:32:56.114111] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:25.137 [2024-11-27 17:32:56.114136] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b480 00:11:25.137 [2024-11-27 17:32:56.114158] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:25.137 [2024-11-27 17:32:56.116757] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:25.138 [2024-11-27 17:32:56.116795] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:11:25.138 [2024-11-27 17:32:56.116877] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:11:25.138 [2024-11-27 17:32:56.116927] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:11:25.138 [2024-11-27 17:32:56.117057] bdev_raid.c:3675:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev pt2 (4) greater than existing raid bdev raid_bdev1 (2) 00:11:25.138 [2024-11-27 17:32:56.117080] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:11:25.138 [2024-11-27 17:32:56.117097] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002000 name raid_bdev1, state configuring 00:11:25.138 [2024-11-27 17:32:56.117131] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:11:25.138 [2024-11-27 17:32:56.117250] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:11:25.138 pt1 00:11:25.138 17:32:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:25.138 17:32:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@542 -- # '[' 4 -gt 2 ']' 00:11:25.138 17:32:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@545 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 3 00:11:25.138 17:32:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:25.138 17:32:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:11:25.138 17:32:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:25.138 17:32:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:25.138 17:32:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:11:25.138 17:32:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:25.138 17:32:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:25.138 17:32:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:25.138 17:32:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:25.138 17:32:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:25.138 17:32:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:25.138 17:32:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:25.138 17:32:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:25.138 17:32:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:25.138 17:32:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:25.138 "name": "raid_bdev1", 00:11:25.138 "uuid": "273fd7ed-7d4a-49bf-990d-f56b6d0dd53d", 00:11:25.138 "strip_size_kb": 0, 00:11:25.138 "state": "configuring", 00:11:25.138 "raid_level": "raid1", 00:11:25.138 "superblock": true, 00:11:25.138 "num_base_bdevs": 4, 00:11:25.138 "num_base_bdevs_discovered": 2, 00:11:25.138 "num_base_bdevs_operational": 3, 00:11:25.138 "base_bdevs_list": [ 00:11:25.138 { 00:11:25.138 "name": null, 00:11:25.138 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:25.138 "is_configured": false, 00:11:25.138 "data_offset": 2048, 00:11:25.138 "data_size": 63488 00:11:25.138 }, 00:11:25.138 { 00:11:25.138 "name": "pt2", 00:11:25.138 "uuid": "00000000-0000-0000-0000-000000000002", 00:11:25.138 "is_configured": true, 00:11:25.138 "data_offset": 2048, 00:11:25.138 "data_size": 63488 00:11:25.138 }, 00:11:25.138 { 00:11:25.138 "name": "pt3", 00:11:25.138 "uuid": "00000000-0000-0000-0000-000000000003", 00:11:25.138 "is_configured": true, 00:11:25.138 "data_offset": 2048, 00:11:25.138 "data_size": 63488 00:11:25.138 }, 00:11:25.138 { 00:11:25.138 "name": null, 00:11:25.138 "uuid": "00000000-0000-0000-0000-000000000004", 00:11:25.138 "is_configured": false, 00:11:25.138 "data_offset": 2048, 00:11:25.138 "data_size": 63488 00:11:25.138 } 00:11:25.138 ] 00:11:25.138 }' 00:11:25.138 17:32:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:25.138 17:32:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:25.398 17:32:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@546 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:11:25.398 17:32:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@546 -- # rpc_cmd bdev_raid_get_bdevs configuring 00:11:25.398 17:32:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:25.398 17:32:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:25.398 17:32:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:25.398 17:32:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@546 -- # [[ false == \f\a\l\s\e ]] 00:11:25.398 17:32:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@549 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:11:25.398 17:32:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:25.398 17:32:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:25.398 [2024-11-27 17:32:56.581262] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:11:25.398 [2024-11-27 17:32:56.581371] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:25.398 [2024-11-27 17:32:56.581397] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ba80 00:11:25.398 [2024-11-27 17:32:56.581409] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:25.398 [2024-11-27 17:32:56.581881] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:25.398 [2024-11-27 17:32:56.581902] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:11:25.398 [2024-11-27 17:32:56.581975] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt4 00:11:25.398 [2024-11-27 17:32:56.582002] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:11:25.398 [2024-11-27 17:32:56.582128] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000002380 00:11:25.398 [2024-11-27 17:32:56.582143] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:11:25.398 [2024-11-27 17:32:56.582429] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000027a0 00:11:25.398 [2024-11-27 17:32:56.582563] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000002380 00:11:25.398 [2024-11-27 17:32:56.582577] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000002380 00:11:25.398 [2024-11-27 17:32:56.582699] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:25.658 pt4 00:11:25.658 17:32:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:25.658 17:32:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@554 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:11:25.658 17:32:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:25.658 17:32:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:25.658 17:32:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:25.658 17:32:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:25.658 17:32:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:11:25.658 17:32:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:25.658 17:32:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:25.658 17:32:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:25.658 17:32:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:25.658 17:32:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:25.658 17:32:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:25.658 17:32:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:25.658 17:32:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:25.658 17:32:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:25.658 17:32:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:25.658 "name": "raid_bdev1", 00:11:25.658 "uuid": "273fd7ed-7d4a-49bf-990d-f56b6d0dd53d", 00:11:25.658 "strip_size_kb": 0, 00:11:25.658 "state": "online", 00:11:25.658 "raid_level": "raid1", 00:11:25.658 "superblock": true, 00:11:25.658 "num_base_bdevs": 4, 00:11:25.658 "num_base_bdevs_discovered": 3, 00:11:25.658 "num_base_bdevs_operational": 3, 00:11:25.658 "base_bdevs_list": [ 00:11:25.658 { 00:11:25.658 "name": null, 00:11:25.658 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:25.658 "is_configured": false, 00:11:25.658 "data_offset": 2048, 00:11:25.658 "data_size": 63488 00:11:25.658 }, 00:11:25.658 { 00:11:25.658 "name": "pt2", 00:11:25.658 "uuid": "00000000-0000-0000-0000-000000000002", 00:11:25.658 "is_configured": true, 00:11:25.658 "data_offset": 2048, 00:11:25.658 "data_size": 63488 00:11:25.658 }, 00:11:25.658 { 00:11:25.659 "name": "pt3", 00:11:25.659 "uuid": "00000000-0000-0000-0000-000000000003", 00:11:25.659 "is_configured": true, 00:11:25.659 "data_offset": 2048, 00:11:25.659 "data_size": 63488 00:11:25.659 }, 00:11:25.659 { 00:11:25.659 "name": "pt4", 00:11:25.659 "uuid": "00000000-0000-0000-0000-000000000004", 00:11:25.659 "is_configured": true, 00:11:25.659 "data_offset": 2048, 00:11:25.659 "data_size": 63488 00:11:25.659 } 00:11:25.659 ] 00:11:25.659 }' 00:11:25.659 17:32:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:25.659 17:32:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:25.919 17:32:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # rpc_cmd bdev_raid_get_bdevs online 00:11:25.919 17:32:56 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:11:25.919 17:32:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:25.919 17:32:56 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:25.919 17:32:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:25.919 17:32:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@555 -- # [[ false == \f\a\l\s\e ]] 00:11:25.919 17:32:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # jq -r '.[] | .uuid' 00:11:25.919 17:32:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:11:25.919 17:32:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:25.919 17:32:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:25.919 [2024-11-27 17:32:57.048774] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:11:25.919 17:32:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:25.919 17:32:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@558 -- # '[' 273fd7ed-7d4a-49bf-990d-f56b6d0dd53d '!=' 273fd7ed-7d4a-49bf-990d-f56b6d0dd53d ']' 00:11:25.919 17:32:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 85147 00:11:25.919 17:32:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@950 -- # '[' -z 85147 ']' 00:11:25.919 17:32:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@954 -- # kill -0 85147 00:11:25.919 17:32:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@955 -- # uname 00:11:25.919 17:32:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:11:25.919 17:32:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 85147 00:11:26.179 17:32:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:11:26.179 17:32:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:11:26.179 17:32:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 85147' 00:11:26.179 killing process with pid 85147 00:11:26.179 17:32:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@969 -- # kill 85147 00:11:26.179 [2024-11-27 17:32:57.111445] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:11:26.179 [2024-11-27 17:32:57.111553] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:11:26.179 [2024-11-27 17:32:57.111653] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:11:26.179 17:32:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@974 -- # wait 85147 00:11:26.179 [2024-11-27 17:32:57.111663] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002380 name raid_bdev1, state offline 00:11:26.179 [2024-11-27 17:32:57.192406] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:11:26.440 17:32:57 bdev_raid.raid_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:11:26.440 00:11:26.440 real 0m7.306s 00:11:26.440 user 0m12.055s 00:11:26.440 sys 0m1.585s 00:11:26.440 17:32:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:11:26.440 17:32:57 bdev_raid.raid_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:11:26.440 ************************************ 00:11:26.440 END TEST raid_superblock_test 00:11:26.440 ************************************ 00:11:26.440 17:32:57 bdev_raid -- bdev/bdev_raid.sh@971 -- # run_test raid_read_error_test raid_io_error_test raid1 4 read 00:11:26.440 17:32:57 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:11:26.440 17:32:57 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:11:26.440 17:32:57 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:11:26.700 ************************************ 00:11:26.700 START TEST raid_read_error_test 00:11:26.700 ************************************ 00:11:26.700 17:32:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1125 -- # raid_io_error_test raid1 4 read 00:11:26.700 17:32:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid1 00:11:26.700 17:32:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=4 00:11:26.700 17:32:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=read 00:11:26.700 17:32:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:11:26.700 17:32:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:11:26.700 17:32:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:11:26.700 17:32:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:11:26.700 17:32:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:11:26.700 17:32:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:11:26.700 17:32:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:11:26.700 17:32:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:11:26.700 17:32:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:11:26.700 17:32:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:11:26.700 17:32:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:11:26.700 17:32:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev4 00:11:26.700 17:32:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:11:26.700 17:32:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:11:26.700 17:32:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:11:26.700 17:32:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:11:26.700 17:32:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:11:26.700 17:32:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:11:26.700 17:32:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:11:26.700 17:32:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:11:26.700 17:32:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:11:26.700 17:32:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid1 '!=' raid1 ']' 00:11:26.700 17:32:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@804 -- # strip_size=0 00:11:26.700 17:32:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:11:26.700 17:32:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.eZcH5c8Ex9 00:11:26.700 17:32:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=85623 00:11:26.700 17:32:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:11:26.700 17:32:57 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 85623 00:11:26.700 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:11:26.700 17:32:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@831 -- # '[' -z 85623 ']' 00:11:26.700 17:32:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:11:26.700 17:32:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:11:26.700 17:32:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:11:26.700 17:32:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:11:26.700 17:32:57 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:26.700 [2024-11-27 17:32:57.744043] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:11:26.700 [2024-11-27 17:32:57.744217] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid85623 ] 00:11:26.960 [2024-11-27 17:32:57.891102] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:11:26.960 [2024-11-27 17:32:57.959060] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:11:26.960 [2024-11-27 17:32:58.034630] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:26.960 [2024-11-27 17:32:58.034684] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:27.530 17:32:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:11:27.530 17:32:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@864 -- # return 0 00:11:27.530 17:32:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:11:27.530 17:32:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:11:27.530 17:32:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:27.530 17:32:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:27.530 BaseBdev1_malloc 00:11:27.531 17:32:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:27.531 17:32:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:11:27.531 17:32:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:27.531 17:32:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:27.531 true 00:11:27.531 17:32:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:27.531 17:32:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:11:27.531 17:32:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:27.531 17:32:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:27.531 [2024-11-27 17:32:58.604751] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:11:27.531 [2024-11-27 17:32:58.604877] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:27.531 [2024-11-27 17:32:58.604907] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:11:27.531 [2024-11-27 17:32:58.604916] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:27.531 [2024-11-27 17:32:58.607331] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:27.531 [2024-11-27 17:32:58.607365] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:11:27.531 BaseBdev1 00:11:27.531 17:32:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:27.531 17:32:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:11:27.531 17:32:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:11:27.531 17:32:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:27.531 17:32:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:27.531 BaseBdev2_malloc 00:11:27.531 17:32:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:27.531 17:32:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:11:27.531 17:32:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:27.531 17:32:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:27.531 true 00:11:27.531 17:32:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:27.531 17:32:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:11:27.531 17:32:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:27.531 17:32:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:27.531 [2024-11-27 17:32:58.661308] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:11:27.531 [2024-11-27 17:32:58.661362] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:27.531 [2024-11-27 17:32:58.661383] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:11:27.531 [2024-11-27 17:32:58.661391] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:27.531 [2024-11-27 17:32:58.663711] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:27.531 [2024-11-27 17:32:58.663783] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:11:27.531 BaseBdev2 00:11:27.531 17:32:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:27.531 17:32:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:11:27.531 17:32:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:11:27.531 17:32:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:27.531 17:32:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:27.531 BaseBdev3_malloc 00:11:27.531 17:32:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:27.531 17:32:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:11:27.531 17:32:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:27.531 17:32:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:27.531 true 00:11:27.531 17:32:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:27.531 17:32:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:11:27.531 17:32:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:27.531 17:32:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:27.531 [2024-11-27 17:32:58.707573] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:11:27.531 [2024-11-27 17:32:58.707617] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:27.531 [2024-11-27 17:32:58.707653] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008780 00:11:27.531 [2024-11-27 17:32:58.707661] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:27.531 [2024-11-27 17:32:58.709996] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:27.531 [2024-11-27 17:32:58.710027] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:11:27.531 BaseBdev3 00:11:27.531 17:32:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:27.531 17:32:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:11:27.531 17:32:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:11:27.531 17:32:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:27.531 17:32:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:27.791 BaseBdev4_malloc 00:11:27.791 17:32:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:27.791 17:32:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev4_malloc 00:11:27.791 17:32:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:27.791 17:32:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:27.791 true 00:11:27.791 17:32:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:27.791 17:32:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev4_malloc -p BaseBdev4 00:11:27.791 17:32:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:27.791 17:32:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:27.791 [2024-11-27 17:32:58.754077] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev4_malloc 00:11:27.791 [2024-11-27 17:32:58.754123] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:27.791 [2024-11-27 17:32:58.754158] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:11:27.791 [2024-11-27 17:32:58.754182] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:27.791 [2024-11-27 17:32:58.756520] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:27.791 [2024-11-27 17:32:58.756608] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:11:27.791 BaseBdev4 00:11:27.791 17:32:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:27.791 17:32:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 -s 00:11:27.791 17:32:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:27.791 17:32:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:27.791 [2024-11-27 17:32:58.766122] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:11:27.791 [2024-11-27 17:32:58.768244] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:11:27.791 [2024-11-27 17:32:58.768315] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:11:27.791 [2024-11-27 17:32:58.768372] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:11:27.791 [2024-11-27 17:32:58.768559] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000002000 00:11:27.791 [2024-11-27 17:32:58.768592] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:11:27.791 [2024-11-27 17:32:58.768870] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:11:27.791 [2024-11-27 17:32:58.769037] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000002000 00:11:27.791 [2024-11-27 17:32:58.769050] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000002000 00:11:27.791 [2024-11-27 17:32:58.769190] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:27.791 17:32:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:27.791 17:32:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:11:27.791 17:32:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:27.791 17:32:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:27.791 17:32:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:27.791 17:32:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:27.791 17:32:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:27.791 17:32:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:27.791 17:32:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:27.791 17:32:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:27.791 17:32:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:27.791 17:32:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:27.791 17:32:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:27.791 17:32:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:27.791 17:32:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:27.791 17:32:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:27.791 17:32:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:27.791 "name": "raid_bdev1", 00:11:27.791 "uuid": "79dacd20-218e-40bb-8ebc-239ee0e193dc", 00:11:27.791 "strip_size_kb": 0, 00:11:27.791 "state": "online", 00:11:27.791 "raid_level": "raid1", 00:11:27.791 "superblock": true, 00:11:27.791 "num_base_bdevs": 4, 00:11:27.791 "num_base_bdevs_discovered": 4, 00:11:27.791 "num_base_bdevs_operational": 4, 00:11:27.791 "base_bdevs_list": [ 00:11:27.791 { 00:11:27.791 "name": "BaseBdev1", 00:11:27.791 "uuid": "1bd127bd-2102-50b8-9f55-620cdc4252f1", 00:11:27.791 "is_configured": true, 00:11:27.791 "data_offset": 2048, 00:11:27.791 "data_size": 63488 00:11:27.791 }, 00:11:27.791 { 00:11:27.791 "name": "BaseBdev2", 00:11:27.791 "uuid": "3c45de1a-8c59-52e5-b125-4663dee45492", 00:11:27.791 "is_configured": true, 00:11:27.791 "data_offset": 2048, 00:11:27.791 "data_size": 63488 00:11:27.791 }, 00:11:27.791 { 00:11:27.791 "name": "BaseBdev3", 00:11:27.791 "uuid": "0b43da25-7967-5133-9a71-b9c053da8132", 00:11:27.791 "is_configured": true, 00:11:27.791 "data_offset": 2048, 00:11:27.791 "data_size": 63488 00:11:27.791 }, 00:11:27.791 { 00:11:27.791 "name": "BaseBdev4", 00:11:27.791 "uuid": "a751d3bd-fb69-5299-8532-29165e262a3f", 00:11:27.791 "is_configured": true, 00:11:27.791 "data_offset": 2048, 00:11:27.791 "data_size": 63488 00:11:27.791 } 00:11:27.791 ] 00:11:27.791 }' 00:11:27.791 17:32:58 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:27.791 17:32:58 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:28.051 17:32:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:11:28.051 17:32:59 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:11:28.311 [2024-11-27 17:32:59.249695] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000026d0 00:11:29.251 17:33:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc read failure 00:11:29.251 17:33:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:29.251 17:33:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:29.251 17:33:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:29.251 17:33:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:11:29.251 17:33:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid1 = \r\a\i\d\1 ]] 00:11:29.251 17:33:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@832 -- # [[ read = \w\r\i\t\e ]] 00:11:29.251 17:33:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@835 -- # expected_num_base_bdevs=4 00:11:29.251 17:33:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:11:29.251 17:33:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:29.251 17:33:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:29.251 17:33:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:29.251 17:33:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:29.251 17:33:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:29.251 17:33:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:29.251 17:33:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:29.251 17:33:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:29.251 17:33:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:29.251 17:33:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:29.251 17:33:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:29.251 17:33:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:29.251 17:33:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:29.251 17:33:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:29.251 17:33:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:29.251 "name": "raid_bdev1", 00:11:29.251 "uuid": "79dacd20-218e-40bb-8ebc-239ee0e193dc", 00:11:29.251 "strip_size_kb": 0, 00:11:29.251 "state": "online", 00:11:29.251 "raid_level": "raid1", 00:11:29.251 "superblock": true, 00:11:29.251 "num_base_bdevs": 4, 00:11:29.251 "num_base_bdevs_discovered": 4, 00:11:29.251 "num_base_bdevs_operational": 4, 00:11:29.251 "base_bdevs_list": [ 00:11:29.251 { 00:11:29.251 "name": "BaseBdev1", 00:11:29.251 "uuid": "1bd127bd-2102-50b8-9f55-620cdc4252f1", 00:11:29.251 "is_configured": true, 00:11:29.251 "data_offset": 2048, 00:11:29.251 "data_size": 63488 00:11:29.251 }, 00:11:29.251 { 00:11:29.251 "name": "BaseBdev2", 00:11:29.251 "uuid": "3c45de1a-8c59-52e5-b125-4663dee45492", 00:11:29.251 "is_configured": true, 00:11:29.251 "data_offset": 2048, 00:11:29.251 "data_size": 63488 00:11:29.251 }, 00:11:29.251 { 00:11:29.251 "name": "BaseBdev3", 00:11:29.251 "uuid": "0b43da25-7967-5133-9a71-b9c053da8132", 00:11:29.251 "is_configured": true, 00:11:29.251 "data_offset": 2048, 00:11:29.251 "data_size": 63488 00:11:29.251 }, 00:11:29.251 { 00:11:29.251 "name": "BaseBdev4", 00:11:29.251 "uuid": "a751d3bd-fb69-5299-8532-29165e262a3f", 00:11:29.251 "is_configured": true, 00:11:29.251 "data_offset": 2048, 00:11:29.251 "data_size": 63488 00:11:29.251 } 00:11:29.251 ] 00:11:29.251 }' 00:11:29.251 17:33:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:29.251 17:33:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:29.512 17:33:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:11:29.512 17:33:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:29.512 17:33:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:29.512 [2024-11-27 17:33:00.639393] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:11:29.512 [2024-11-27 17:33:00.639430] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:11:29.512 [2024-11-27 17:33:00.642101] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:11:29.512 [2024-11-27 17:33:00.642176] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:29.512 [2024-11-27 17:33:00.642327] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:11:29.512 [2024-11-27 17:33:00.642338] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002000 name raid_bdev1, state offline 00:11:29.512 { 00:11:29.512 "results": [ 00:11:29.512 { 00:11:29.512 "job": "raid_bdev1", 00:11:29.512 "core_mask": "0x1", 00:11:29.512 "workload": "randrw", 00:11:29.512 "percentage": 50, 00:11:29.512 "status": "finished", 00:11:29.512 "queue_depth": 1, 00:11:29.512 "io_size": 131072, 00:11:29.512 "runtime": 1.390312, 00:11:29.512 "iops": 8600.227862522945, 00:11:29.512 "mibps": 1075.028482815368, 00:11:29.512 "io_failed": 0, 00:11:29.512 "io_timeout": 0, 00:11:29.512 "avg_latency_us": 113.75114246720325, 00:11:29.512 "min_latency_us": 22.46986899563319, 00:11:29.512 "max_latency_us": 1552.5449781659388 00:11:29.512 } 00:11:29.512 ], 00:11:29.512 "core_count": 1 00:11:29.512 } 00:11:29.512 17:33:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:29.512 17:33:00 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 85623 00:11:29.512 17:33:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@950 -- # '[' -z 85623 ']' 00:11:29.512 17:33:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@954 -- # kill -0 85623 00:11:29.512 17:33:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@955 -- # uname 00:11:29.512 17:33:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:11:29.512 17:33:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 85623 00:11:29.512 17:33:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:11:29.512 17:33:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:11:29.512 17:33:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 85623' 00:11:29.512 killing process with pid 85623 00:11:29.512 17:33:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@969 -- # kill 85623 00:11:29.512 [2024-11-27 17:33:00.678010] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:11:29.512 17:33:00 bdev_raid.raid_read_error_test -- common/autotest_common.sh@974 -- # wait 85623 00:11:29.772 [2024-11-27 17:33:00.745369] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:11:30.047 17:33:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.eZcH5c8Ex9 00:11:30.047 17:33:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:11:30.047 17:33:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:11:30.047 17:33:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.00 00:11:30.047 17:33:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid1 00:11:30.047 17:33:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:11:30.047 17:33:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@199 -- # return 0 00:11:30.047 17:33:01 bdev_raid.raid_read_error_test -- bdev/bdev_raid.sh@847 -- # [[ 0.00 = \0\.\0\0 ]] 00:11:30.047 ************************************ 00:11:30.047 END TEST raid_read_error_test 00:11:30.047 ************************************ 00:11:30.047 00:11:30.047 real 0m3.490s 00:11:30.047 user 0m4.201s 00:11:30.047 sys 0m0.654s 00:11:30.047 17:33:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:11:30.047 17:33:01 bdev_raid.raid_read_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:30.047 17:33:01 bdev_raid -- bdev/bdev_raid.sh@972 -- # run_test raid_write_error_test raid_io_error_test raid1 4 write 00:11:30.047 17:33:01 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:11:30.047 17:33:01 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:11:30.047 17:33:01 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:11:30.047 ************************************ 00:11:30.047 START TEST raid_write_error_test 00:11:30.047 ************************************ 00:11:30.047 17:33:01 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1125 -- # raid_io_error_test raid1 4 write 00:11:30.047 17:33:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@790 -- # local raid_level=raid1 00:11:30.047 17:33:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@791 -- # local num_base_bdevs=4 00:11:30.047 17:33:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@792 -- # local error_io_type=write 00:11:30.047 17:33:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i = 1 )) 00:11:30.047 17:33:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:11:30.047 17:33:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev1 00:11:30.047 17:33:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:11:30.047 17:33:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:11:30.047 17:33:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev2 00:11:30.047 17:33:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:11:30.047 17:33:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:11:30.047 17:33:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev3 00:11:30.047 17:33:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:11:30.047 17:33:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:11:30.047 17:33:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # echo BaseBdev4 00:11:30.047 17:33:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i++ )) 00:11:30.047 17:33:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # (( i <= num_base_bdevs )) 00:11:30.047 17:33:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:11:30.047 17:33:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@793 -- # local base_bdevs 00:11:30.047 17:33:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@794 -- # local raid_bdev_name=raid_bdev1 00:11:30.047 17:33:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@795 -- # local strip_size 00:11:30.047 17:33:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@796 -- # local create_arg 00:11:30.047 17:33:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@797 -- # local bdevperf_log 00:11:30.047 17:33:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@798 -- # local fail_per_s 00:11:30.047 17:33:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@800 -- # '[' raid1 '!=' raid1 ']' 00:11:30.047 17:33:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@804 -- # strip_size=0 00:11:30.047 17:33:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # mktemp -p /raidtest 00:11:30.047 17:33:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@807 -- # bdevperf_log=/raidtest/tmp.nt2LJ5PjWR 00:11:30.047 17:33:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@810 -- # raid_pid=85752 00:11:30.047 17:33:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@811 -- # waitforlisten 85752 00:11:30.047 17:33:01 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@809 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 128k -q 1 -z -f -L bdev_raid 00:11:30.047 17:33:01 bdev_raid.raid_write_error_test -- common/autotest_common.sh@831 -- # '[' -z 85752 ']' 00:11:30.047 17:33:01 bdev_raid.raid_write_error_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:11:30.047 17:33:01 bdev_raid.raid_write_error_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:11:30.047 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:11:30.047 17:33:01 bdev_raid.raid_write_error_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:11:30.047 17:33:01 bdev_raid.raid_write_error_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:11:30.319 17:33:01 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:30.319 [2024-11-27 17:33:01.310071] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:11:30.319 [2024-11-27 17:33:01.310237] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid85752 ] 00:11:30.319 [2024-11-27 17:33:01.454724] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:11:30.579 [2024-11-27 17:33:01.524439] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:11:30.579 [2024-11-27 17:33:01.601138] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:30.579 [2024-11-27 17:33:01.601183] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:31.150 17:33:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:11:31.150 17:33:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@864 -- # return 0 00:11:31.150 17:33:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:11:31.150 17:33:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:11:31.150 17:33:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:31.150 17:33:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:31.150 BaseBdev1_malloc 00:11:31.150 17:33:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:31.150 17:33:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev1_malloc 00:11:31.150 17:33:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:31.150 17:33:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:31.150 true 00:11:31.150 17:33:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:31.150 17:33:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev1_malloc -p BaseBdev1 00:11:31.150 17:33:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:31.150 17:33:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:31.150 [2024-11-27 17:33:02.179838] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev1_malloc 00:11:31.150 [2024-11-27 17:33:02.179899] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:31.150 [2024-11-27 17:33:02.179922] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006980 00:11:31.150 [2024-11-27 17:33:02.179931] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:31.150 [2024-11-27 17:33:02.182349] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:31.150 [2024-11-27 17:33:02.182396] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:11:31.150 BaseBdev1 00:11:31.150 17:33:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:31.150 17:33:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:11:31.150 17:33:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:11:31.150 17:33:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:31.150 17:33:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:31.150 BaseBdev2_malloc 00:11:31.150 17:33:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:31.150 17:33:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev2_malloc 00:11:31.150 17:33:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:31.150 17:33:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:31.150 true 00:11:31.150 17:33:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:31.150 17:33:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev2_malloc -p BaseBdev2 00:11:31.150 17:33:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:31.150 17:33:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:31.150 [2024-11-27 17:33:02.243657] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev2_malloc 00:11:31.150 [2024-11-27 17:33:02.243738] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:31.150 [2024-11-27 17:33:02.243773] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007880 00:11:31.150 [2024-11-27 17:33:02.243788] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:31.150 [2024-11-27 17:33:02.247379] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:31.150 [2024-11-27 17:33:02.247424] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:11:31.150 BaseBdev2 00:11:31.150 17:33:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:31.150 17:33:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:11:31.150 17:33:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:11:31.150 17:33:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:31.150 17:33:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:31.150 BaseBdev3_malloc 00:11:31.150 17:33:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:31.150 17:33:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev3_malloc 00:11:31.150 17:33:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:31.150 17:33:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:31.150 true 00:11:31.150 17:33:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:31.150 17:33:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev3_malloc -p BaseBdev3 00:11:31.150 17:33:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:31.150 17:33:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:31.150 [2024-11-27 17:33:02.290042] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev3_malloc 00:11:31.150 [2024-11-27 17:33:02.290084] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:31.150 [2024-11-27 17:33:02.290101] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008780 00:11:31.150 [2024-11-27 17:33:02.290109] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:31.150 [2024-11-27 17:33:02.292445] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:31.151 [2024-11-27 17:33:02.292474] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:11:31.151 BaseBdev3 00:11:31.151 17:33:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:31.151 17:33:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@814 -- # for bdev in "${base_bdevs[@]}" 00:11:31.151 17:33:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@815 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:11:31.151 17:33:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:31.151 17:33:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:31.151 BaseBdev4_malloc 00:11:31.151 17:33:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:31.151 17:33:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@816 -- # rpc_cmd bdev_error_create BaseBdev4_malloc 00:11:31.151 17:33:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:31.151 17:33:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:31.151 true 00:11:31.151 17:33:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:31.151 17:33:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@817 -- # rpc_cmd bdev_passthru_create -b EE_BaseBdev4_malloc -p BaseBdev4 00:11:31.151 17:33:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:31.151 17:33:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:31.151 [2024-11-27 17:33:02.336462] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on EE_BaseBdev4_malloc 00:11:31.151 [2024-11-27 17:33:02.336513] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:31.151 [2024-11-27 17:33:02.336534] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:11:31.151 [2024-11-27 17:33:02.336544] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:31.151 [2024-11-27 17:33:02.338930] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:31.151 [2024-11-27 17:33:02.338961] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:11:31.410 BaseBdev4 00:11:31.410 17:33:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:31.410 17:33:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@821 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 -s 00:11:31.410 17:33:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:31.410 17:33:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:31.410 [2024-11-27 17:33:02.348506] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:11:31.410 [2024-11-27 17:33:02.350588] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:11:31.410 [2024-11-27 17:33:02.350665] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:11:31.410 [2024-11-27 17:33:02.350750] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:11:31.410 [2024-11-27 17:33:02.350951] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000002000 00:11:31.410 [2024-11-27 17:33:02.350970] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:11:31.410 [2024-11-27 17:33:02.351260] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:11:31.410 [2024-11-27 17:33:02.351431] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000002000 00:11:31.410 [2024-11-27 17:33:02.351458] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000002000 00:11:31.410 [2024-11-27 17:33:02.351587] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:31.410 17:33:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:31.410 17:33:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@822 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:11:31.410 17:33:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:31.410 17:33:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:31.410 17:33:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:31.410 17:33:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:31.410 17:33:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:11:31.410 17:33:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:31.410 17:33:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:31.410 17:33:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:31.410 17:33:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:31.410 17:33:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:31.410 17:33:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:31.410 17:33:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:31.410 17:33:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:31.410 17:33:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:31.410 17:33:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:31.410 "name": "raid_bdev1", 00:11:31.410 "uuid": "7afc0612-0a98-4f1a-92e4-02f77600c992", 00:11:31.410 "strip_size_kb": 0, 00:11:31.410 "state": "online", 00:11:31.410 "raid_level": "raid1", 00:11:31.410 "superblock": true, 00:11:31.410 "num_base_bdevs": 4, 00:11:31.410 "num_base_bdevs_discovered": 4, 00:11:31.410 "num_base_bdevs_operational": 4, 00:11:31.410 "base_bdevs_list": [ 00:11:31.410 { 00:11:31.410 "name": "BaseBdev1", 00:11:31.410 "uuid": "28509eae-b73b-5c90-8c54-6d59a2c42208", 00:11:31.410 "is_configured": true, 00:11:31.410 "data_offset": 2048, 00:11:31.410 "data_size": 63488 00:11:31.410 }, 00:11:31.410 { 00:11:31.410 "name": "BaseBdev2", 00:11:31.410 "uuid": "01c1f285-4d57-5144-84a9-d5eaf25cfeca", 00:11:31.410 "is_configured": true, 00:11:31.410 "data_offset": 2048, 00:11:31.410 "data_size": 63488 00:11:31.410 }, 00:11:31.410 { 00:11:31.410 "name": "BaseBdev3", 00:11:31.410 "uuid": "69b8cde0-c454-5529-9712-605d0c4646e1", 00:11:31.410 "is_configured": true, 00:11:31.410 "data_offset": 2048, 00:11:31.410 "data_size": 63488 00:11:31.410 }, 00:11:31.410 { 00:11:31.410 "name": "BaseBdev4", 00:11:31.410 "uuid": "1271b29b-6363-54fc-89da-fcc0abf1ceeb", 00:11:31.410 "is_configured": true, 00:11:31.410 "data_offset": 2048, 00:11:31.410 "data_size": 63488 00:11:31.410 } 00:11:31.410 ] 00:11:31.410 }' 00:11:31.410 17:33:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:31.410 17:33:02 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:31.670 17:33:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@826 -- # sleep 1 00:11:31.670 17:33:02 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@825 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:11:31.929 [2024-11-27 17:33:02.908087] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000026d0 00:11:32.869 17:33:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@829 -- # rpc_cmd bdev_error_inject_error EE_BaseBdev1_malloc write failure 00:11:32.869 17:33:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:32.869 17:33:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:32.869 [2024-11-27 17:33:03.824147] bdev_raid.c:2272:_raid_bdev_fail_base_bdev: *NOTICE*: Failing base bdev in slot 0 ('BaseBdev1') of raid bdev 'raid_bdev1' 00:11:32.869 [2024-11-27 17:33:03.824216] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:11:32.869 [2024-11-27 17:33:03.824466] bdev_raid.c:1970:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 0 raid_ch: 0x60d0000026d0 00:11:32.869 17:33:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:32.869 17:33:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@831 -- # local expected_num_base_bdevs 00:11:32.869 17:33:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ raid1 = \r\a\i\d\1 ]] 00:11:32.869 17:33:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@832 -- # [[ write = \w\r\i\t\e ]] 00:11:32.869 17:33:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@833 -- # expected_num_base_bdevs=3 00:11:32.869 17:33:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@837 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:11:32.869 17:33:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:32.869 17:33:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:32.869 17:33:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:32.869 17:33:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:32.869 17:33:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:11:32.869 17:33:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:32.869 17:33:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:32.869 17:33:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:32.869 17:33:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:32.869 17:33:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:32.869 17:33:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:32.869 17:33:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:32.869 17:33:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:32.869 17:33:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:32.869 17:33:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:32.869 "name": "raid_bdev1", 00:11:32.869 "uuid": "7afc0612-0a98-4f1a-92e4-02f77600c992", 00:11:32.869 "strip_size_kb": 0, 00:11:32.869 "state": "online", 00:11:32.869 "raid_level": "raid1", 00:11:32.869 "superblock": true, 00:11:32.869 "num_base_bdevs": 4, 00:11:32.869 "num_base_bdevs_discovered": 3, 00:11:32.869 "num_base_bdevs_operational": 3, 00:11:32.869 "base_bdevs_list": [ 00:11:32.869 { 00:11:32.869 "name": null, 00:11:32.869 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:32.869 "is_configured": false, 00:11:32.869 "data_offset": 0, 00:11:32.869 "data_size": 63488 00:11:32.869 }, 00:11:32.869 { 00:11:32.869 "name": "BaseBdev2", 00:11:32.869 "uuid": "01c1f285-4d57-5144-84a9-d5eaf25cfeca", 00:11:32.869 "is_configured": true, 00:11:32.869 "data_offset": 2048, 00:11:32.869 "data_size": 63488 00:11:32.869 }, 00:11:32.869 { 00:11:32.869 "name": "BaseBdev3", 00:11:32.869 "uuid": "69b8cde0-c454-5529-9712-605d0c4646e1", 00:11:32.869 "is_configured": true, 00:11:32.869 "data_offset": 2048, 00:11:32.869 "data_size": 63488 00:11:32.869 }, 00:11:32.869 { 00:11:32.869 "name": "BaseBdev4", 00:11:32.869 "uuid": "1271b29b-6363-54fc-89da-fcc0abf1ceeb", 00:11:32.869 "is_configured": true, 00:11:32.869 "data_offset": 2048, 00:11:32.869 "data_size": 63488 00:11:32.869 } 00:11:32.869 ] 00:11:32.869 }' 00:11:32.869 17:33:03 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:32.869 17:33:03 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:33.129 17:33:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@839 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:11:33.129 17:33:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:33.129 17:33:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:33.129 [2024-11-27 17:33:04.226402] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:11:33.129 [2024-11-27 17:33:04.226445] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:11:33.129 [2024-11-27 17:33:04.229084] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:11:33.129 [2024-11-27 17:33:04.229168] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:33.129 [2024-11-27 17:33:04.229276] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:11:33.129 [2024-11-27 17:33:04.229292] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002000 name raid_bdev1, state offline 00:11:33.129 { 00:11:33.129 "results": [ 00:11:33.129 { 00:11:33.129 "job": "raid_bdev1", 00:11:33.129 "core_mask": "0x1", 00:11:33.129 "workload": "randrw", 00:11:33.129 "percentage": 50, 00:11:33.129 "status": "finished", 00:11:33.129 "queue_depth": 1, 00:11:33.129 "io_size": 131072, 00:11:33.129 "runtime": 1.31873, 00:11:33.129 "iops": 9342.321779287648, 00:11:33.129 "mibps": 1167.790222410956, 00:11:33.129 "io_failed": 0, 00:11:33.129 "io_timeout": 0, 00:11:33.129 "avg_latency_us": 104.4913378324732, 00:11:33.129 "min_latency_us": 22.134497816593885, 00:11:33.129 "max_latency_us": 1337.907423580786 00:11:33.129 } 00:11:33.129 ], 00:11:33.129 "core_count": 1 00:11:33.129 } 00:11:33.129 17:33:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:33.129 17:33:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@841 -- # killprocess 85752 00:11:33.129 17:33:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@950 -- # '[' -z 85752 ']' 00:11:33.129 17:33:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@954 -- # kill -0 85752 00:11:33.129 17:33:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@955 -- # uname 00:11:33.129 17:33:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:11:33.129 17:33:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 85752 00:11:33.129 17:33:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:11:33.129 17:33:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:11:33.129 killing process with pid 85752 00:11:33.129 17:33:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 85752' 00:11:33.129 17:33:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@969 -- # kill 85752 00:11:33.129 [2024-11-27 17:33:04.275244] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:11:33.129 17:33:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@974 -- # wait 85752 00:11:33.389 [2024-11-27 17:33:04.341653] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:11:33.703 17:33:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # awk '{print $6}' 00:11:33.703 17:33:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep -v Job /raidtest/tmp.nt2LJ5PjWR 00:11:33.703 17:33:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # grep raid_bdev1 00:11:33.703 17:33:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@845 -- # fail_per_s=0.00 00:11:33.703 17:33:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@846 -- # has_redundancy raid1 00:11:33.703 17:33:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:11:33.703 17:33:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@199 -- # return 0 00:11:33.703 17:33:04 bdev_raid.raid_write_error_test -- bdev/bdev_raid.sh@847 -- # [[ 0.00 = \0\.\0\0 ]] 00:11:33.703 00:11:33.703 real 0m3.516s 00:11:33.703 user 0m4.246s 00:11:33.703 sys 0m0.659s 00:11:33.703 17:33:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:11:33.703 17:33:04 bdev_raid.raid_write_error_test -- common/autotest_common.sh@10 -- # set +x 00:11:33.703 ************************************ 00:11:33.703 END TEST raid_write_error_test 00:11:33.703 ************************************ 00:11:33.703 17:33:04 bdev_raid -- bdev/bdev_raid.sh@976 -- # '[' true = true ']' 00:11:33.703 17:33:04 bdev_raid -- bdev/bdev_raid.sh@977 -- # for n in 2 4 00:11:33.703 17:33:04 bdev_raid -- bdev/bdev_raid.sh@978 -- # run_test raid_rebuild_test raid_rebuild_test raid1 2 false false true 00:11:33.703 17:33:04 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 7 -le 1 ']' 00:11:33.703 17:33:04 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:11:33.703 17:33:04 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:11:33.703 ************************************ 00:11:33.703 START TEST raid_rebuild_test 00:11:33.703 ************************************ 00:11:33.703 17:33:04 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@1125 -- # raid_rebuild_test raid1 2 false false true 00:11:33.703 17:33:04 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:11:33.703 17:33:04 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=2 00:11:33.703 17:33:04 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@571 -- # local superblock=false 00:11:33.703 17:33:04 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:11:33.703 17:33:04 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@573 -- # local verify=true 00:11:33.703 17:33:04 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:11:33.703 17:33:04 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:11:33.703 17:33:04 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:11:33.703 17:33:04 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:11:33.703 17:33:04 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:11:33.703 17:33:04 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:11:33.703 17:33:04 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:11:33.703 17:33:04 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:11:33.703 17:33:04 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:11:33.703 17:33:04 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:11:33.704 17:33:04 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:11:33.704 17:33:04 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@576 -- # local strip_size 00:11:33.704 17:33:04 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@577 -- # local create_arg 00:11:33.704 17:33:04 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:11:33.704 17:33:04 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@579 -- # local data_offset 00:11:33.704 17:33:04 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:11:33.704 17:33:04 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:11:33.704 17:33:04 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@592 -- # '[' false = true ']' 00:11:33.704 17:33:04 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@597 -- # raid_pid=85885 00:11:33.704 17:33:04 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:11:33.704 17:33:04 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@598 -- # waitforlisten 85885 00:11:33.704 17:33:04 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@831 -- # '[' -z 85885 ']' 00:11:33.704 17:33:04 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:11:33.704 17:33:04 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:11:33.704 17:33:04 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:11:33.704 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:11:33.704 17:33:04 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:11:33.704 17:33:04 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:33.964 [2024-11-27 17:33:04.906239] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:11:33.964 [2024-11-27 17:33:04.906453] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid85885 ] 00:11:33.964 I/O size of 3145728 is greater than zero copy threshold (65536). 00:11:33.964 Zero copy mechanism will not be used. 00:11:33.964 [2024-11-27 17:33:05.053468] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:11:33.964 [2024-11-27 17:33:05.121525] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:11:34.223 [2024-11-27 17:33:05.198225] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:34.223 [2024-11-27 17:33:05.198366] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:34.793 17:33:05 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:11:34.793 17:33:05 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@864 -- # return 0 00:11:34.793 17:33:05 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:11:34.793 17:33:05 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:11:34.793 17:33:05 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:34.793 17:33:05 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:34.793 BaseBdev1_malloc 00:11:34.793 17:33:05 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:34.793 17:33:05 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:11:34.793 17:33:05 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:34.793 17:33:05 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:34.793 [2024-11-27 17:33:05.760536] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:11:34.793 [2024-11-27 17:33:05.760679] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:34.793 [2024-11-27 17:33:05.760728] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:11:34.793 [2024-11-27 17:33:05.760797] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:34.793 [2024-11-27 17:33:05.763278] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:34.793 [2024-11-27 17:33:05.763349] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:11:34.793 BaseBdev1 00:11:34.793 17:33:05 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:34.793 17:33:05 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:11:34.793 17:33:05 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:11:34.793 17:33:05 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:34.793 17:33:05 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:34.793 BaseBdev2_malloc 00:11:34.793 17:33:05 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:34.793 17:33:05 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:11:34.793 17:33:05 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:34.793 17:33:05 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:34.793 [2024-11-27 17:33:05.810178] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:11:34.793 [2024-11-27 17:33:05.810364] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:34.793 [2024-11-27 17:33:05.810424] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:11:34.793 [2024-11-27 17:33:05.810447] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:34.793 [2024-11-27 17:33:05.815549] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:34.793 [2024-11-27 17:33:05.815620] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:11:34.793 BaseBdev2 00:11:34.793 17:33:05 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:34.793 17:33:05 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:11:34.793 17:33:05 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:34.793 17:33:05 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:34.793 spare_malloc 00:11:34.793 17:33:05 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:34.793 17:33:05 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:11:34.793 17:33:05 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:34.793 17:33:05 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:34.793 spare_delay 00:11:34.793 17:33:05 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:34.793 17:33:05 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:11:34.793 17:33:05 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:34.793 17:33:05 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:34.793 [2024-11-27 17:33:05.858393] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:11:34.793 [2024-11-27 17:33:05.858478] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:34.793 [2024-11-27 17:33:05.858533] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:11:34.793 [2024-11-27 17:33:05.858560] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:34.793 [2024-11-27 17:33:05.860966] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:34.793 [2024-11-27 17:33:05.861027] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:11:34.793 spare 00:11:34.794 17:33:05 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:34.794 17:33:05 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 00:11:34.794 17:33:05 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:34.794 17:33:05 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:34.794 [2024-11-27 17:33:05.870430] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:11:34.794 [2024-11-27 17:33:05.872599] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:11:34.794 [2024-11-27 17:33:05.872741] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:11:34.794 [2024-11-27 17:33:05.872774] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:11:34.794 [2024-11-27 17:33:05.873100] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:11:34.794 [2024-11-27 17:33:05.873282] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:11:34.794 [2024-11-27 17:33:05.873329] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:11:34.794 [2024-11-27 17:33:05.873493] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:34.794 17:33:05 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:34.794 17:33:05 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:11:34.794 17:33:05 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:34.794 17:33:05 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:34.794 17:33:05 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:34.794 17:33:05 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:34.794 17:33:05 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:11:34.794 17:33:05 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:34.794 17:33:05 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:34.794 17:33:05 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:34.794 17:33:05 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:34.794 17:33:05 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:34.794 17:33:05 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:34.794 17:33:05 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:34.794 17:33:05 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:34.794 17:33:05 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:34.794 17:33:05 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:34.794 "name": "raid_bdev1", 00:11:34.794 "uuid": "1a97666a-4e3c-44fa-99aa-5c7d2e702f16", 00:11:34.794 "strip_size_kb": 0, 00:11:34.794 "state": "online", 00:11:34.794 "raid_level": "raid1", 00:11:34.794 "superblock": false, 00:11:34.794 "num_base_bdevs": 2, 00:11:34.794 "num_base_bdevs_discovered": 2, 00:11:34.794 "num_base_bdevs_operational": 2, 00:11:34.794 "base_bdevs_list": [ 00:11:34.794 { 00:11:34.794 "name": "BaseBdev1", 00:11:34.794 "uuid": "2e873d4d-a0a6-5717-9eb3-6e227cfdf4e3", 00:11:34.794 "is_configured": true, 00:11:34.794 "data_offset": 0, 00:11:34.794 "data_size": 65536 00:11:34.794 }, 00:11:34.794 { 00:11:34.794 "name": "BaseBdev2", 00:11:34.794 "uuid": "2940169b-2fd4-5c3e-b6c3-762b3f89ccb1", 00:11:34.794 "is_configured": true, 00:11:34.794 "data_offset": 0, 00:11:34.794 "data_size": 65536 00:11:34.794 } 00:11:34.794 ] 00:11:34.794 }' 00:11:34.794 17:33:05 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:34.794 17:33:05 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:35.363 17:33:06 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:11:35.363 17:33:06 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:11:35.363 17:33:06 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:35.363 17:33:06 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:35.363 [2024-11-27 17:33:06.329898] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:11:35.363 17:33:06 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:35.363 17:33:06 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=65536 00:11:35.363 17:33:06 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:35.363 17:33:06 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:35.363 17:33:06 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:35.363 17:33:06 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:11:35.363 17:33:06 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:35.363 17:33:06 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@619 -- # data_offset=0 00:11:35.363 17:33:06 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:11:35.363 17:33:06 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:11:35.363 17:33:06 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:11:35.363 17:33:06 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:11:35.363 17:33:06 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:11:35.363 17:33:06 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:11:35.363 17:33:06 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@10 -- # local bdev_list 00:11:35.363 17:33:06 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:11:35.363 17:33:06 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@11 -- # local nbd_list 00:11:35.363 17:33:06 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@12 -- # local i 00:11:35.363 17:33:06 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:11:35.363 17:33:06 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:11:35.363 17:33:06 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:11:35.624 [2024-11-27 17:33:06.593325] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:11:35.624 /dev/nbd0 00:11:35.624 17:33:06 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:11:35.624 17:33:06 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:11:35.624 17:33:06 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:11:35.624 17:33:06 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@869 -- # local i 00:11:35.624 17:33:06 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:11:35.624 17:33:06 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:11:35.624 17:33:06 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:11:35.624 17:33:06 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@873 -- # break 00:11:35.624 17:33:06 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:11:35.624 17:33:06 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:11:35.624 17:33:06 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:11:35.624 1+0 records in 00:11:35.624 1+0 records out 00:11:35.624 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000584494 s, 7.0 MB/s 00:11:35.624 17:33:06 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:35.624 17:33:06 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@886 -- # size=4096 00:11:35.624 17:33:06 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:35.624 17:33:06 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:11:35.624 17:33:06 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@889 -- # return 0 00:11:35.624 17:33:06 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:11:35.624 17:33:06 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:11:35.624 17:33:06 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@629 -- # '[' raid1 = raid5f ']' 00:11:35.624 17:33:06 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@633 -- # write_unit_size=1 00:11:35.624 17:33:06 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=512 count=65536 oflag=direct 00:11:39.823 65536+0 records in 00:11:39.823 65536+0 records out 00:11:39.823 33554432 bytes (34 MB, 32 MiB) copied, 3.57253 s, 9.4 MB/s 00:11:39.823 17:33:10 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:11:39.823 17:33:10 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:11:39.823 17:33:10 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:11:39.823 17:33:10 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@50 -- # local nbd_list 00:11:39.823 17:33:10 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@51 -- # local i 00:11:39.823 17:33:10 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:11:39.823 17:33:10 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:11:39.823 17:33:10 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:11:39.823 [2024-11-27 17:33:10.503597] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:39.823 17:33:10 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:11:39.823 17:33:10 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:11:39.823 17:33:10 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:11:39.823 17:33:10 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:11:39.823 17:33:10 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:11:39.823 17:33:10 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:11:39.823 17:33:10 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:11:39.823 17:33:10 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:11:39.823 17:33:10 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:39.823 17:33:10 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:39.823 [2024-11-27 17:33:10.523649] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:11:39.823 17:33:10 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:39.823 17:33:10 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:11:39.823 17:33:10 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:39.823 17:33:10 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:39.823 17:33:10 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:39.823 17:33:10 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:39.823 17:33:10 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:11:39.823 17:33:10 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:39.823 17:33:10 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:39.823 17:33:10 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:39.823 17:33:10 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:39.823 17:33:10 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:39.823 17:33:10 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:39.823 17:33:10 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:39.823 17:33:10 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:39.823 17:33:10 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:39.823 17:33:10 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:39.823 "name": "raid_bdev1", 00:11:39.823 "uuid": "1a97666a-4e3c-44fa-99aa-5c7d2e702f16", 00:11:39.823 "strip_size_kb": 0, 00:11:39.823 "state": "online", 00:11:39.823 "raid_level": "raid1", 00:11:39.823 "superblock": false, 00:11:39.823 "num_base_bdevs": 2, 00:11:39.823 "num_base_bdevs_discovered": 1, 00:11:39.823 "num_base_bdevs_operational": 1, 00:11:39.823 "base_bdevs_list": [ 00:11:39.823 { 00:11:39.823 "name": null, 00:11:39.823 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:39.823 "is_configured": false, 00:11:39.823 "data_offset": 0, 00:11:39.823 "data_size": 65536 00:11:39.823 }, 00:11:39.823 { 00:11:39.823 "name": "BaseBdev2", 00:11:39.823 "uuid": "2940169b-2fd4-5c3e-b6c3-762b3f89ccb1", 00:11:39.823 "is_configured": true, 00:11:39.823 "data_offset": 0, 00:11:39.823 "data_size": 65536 00:11:39.823 } 00:11:39.823 ] 00:11:39.823 }' 00:11:39.823 17:33:10 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:39.823 17:33:10 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:39.823 17:33:10 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:11:39.823 17:33:10 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:39.823 17:33:10 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:39.823 [2024-11-27 17:33:10.954950] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:11:39.823 [2024-11-27 17:33:10.962277] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000d06220 00:11:39.823 17:33:10 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:39.823 17:33:10 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@647 -- # sleep 1 00:11:39.823 [2024-11-27 17:33:10.964556] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:11:41.205 17:33:11 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:41.205 17:33:11 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:41.205 17:33:11 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:41.205 17:33:11 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:41.205 17:33:11 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:41.205 17:33:11 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:41.205 17:33:11 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:41.205 17:33:11 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:41.205 17:33:11 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:41.205 17:33:11 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:41.205 17:33:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:41.205 "name": "raid_bdev1", 00:11:41.205 "uuid": "1a97666a-4e3c-44fa-99aa-5c7d2e702f16", 00:11:41.205 "strip_size_kb": 0, 00:11:41.205 "state": "online", 00:11:41.205 "raid_level": "raid1", 00:11:41.205 "superblock": false, 00:11:41.205 "num_base_bdevs": 2, 00:11:41.205 "num_base_bdevs_discovered": 2, 00:11:41.205 "num_base_bdevs_operational": 2, 00:11:41.205 "process": { 00:11:41.205 "type": "rebuild", 00:11:41.205 "target": "spare", 00:11:41.205 "progress": { 00:11:41.206 "blocks": 20480, 00:11:41.206 "percent": 31 00:11:41.206 } 00:11:41.206 }, 00:11:41.206 "base_bdevs_list": [ 00:11:41.206 { 00:11:41.206 "name": "spare", 00:11:41.206 "uuid": "1ecf84b4-7847-59f9-a811-7a5c067a5db0", 00:11:41.206 "is_configured": true, 00:11:41.206 "data_offset": 0, 00:11:41.206 "data_size": 65536 00:11:41.206 }, 00:11:41.206 { 00:11:41.206 "name": "BaseBdev2", 00:11:41.206 "uuid": "2940169b-2fd4-5c3e-b6c3-762b3f89ccb1", 00:11:41.206 "is_configured": true, 00:11:41.206 "data_offset": 0, 00:11:41.206 "data_size": 65536 00:11:41.206 } 00:11:41.206 ] 00:11:41.206 }' 00:11:41.206 17:33:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:41.206 17:33:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:41.206 17:33:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:41.206 17:33:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:41.206 17:33:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:11:41.206 17:33:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:41.206 17:33:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:41.206 [2024-11-27 17:33:12.128214] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:11:41.206 [2024-11-27 17:33:12.172886] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:11:41.206 [2024-11-27 17:33:12.172998] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:41.206 [2024-11-27 17:33:12.173022] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:11:41.206 [2024-11-27 17:33:12.173031] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:11:41.206 17:33:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:41.206 17:33:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:11:41.206 17:33:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:41.206 17:33:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:41.206 17:33:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:41.206 17:33:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:41.206 17:33:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:11:41.206 17:33:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:41.206 17:33:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:41.206 17:33:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:41.206 17:33:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:41.206 17:33:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:41.206 17:33:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:41.206 17:33:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:41.206 17:33:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:41.206 17:33:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:41.206 17:33:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:41.206 "name": "raid_bdev1", 00:11:41.206 "uuid": "1a97666a-4e3c-44fa-99aa-5c7d2e702f16", 00:11:41.206 "strip_size_kb": 0, 00:11:41.206 "state": "online", 00:11:41.206 "raid_level": "raid1", 00:11:41.206 "superblock": false, 00:11:41.206 "num_base_bdevs": 2, 00:11:41.206 "num_base_bdevs_discovered": 1, 00:11:41.206 "num_base_bdevs_operational": 1, 00:11:41.206 "base_bdevs_list": [ 00:11:41.206 { 00:11:41.206 "name": null, 00:11:41.206 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:41.206 "is_configured": false, 00:11:41.206 "data_offset": 0, 00:11:41.206 "data_size": 65536 00:11:41.206 }, 00:11:41.206 { 00:11:41.206 "name": "BaseBdev2", 00:11:41.206 "uuid": "2940169b-2fd4-5c3e-b6c3-762b3f89ccb1", 00:11:41.206 "is_configured": true, 00:11:41.206 "data_offset": 0, 00:11:41.206 "data_size": 65536 00:11:41.206 } 00:11:41.206 ] 00:11:41.206 }' 00:11:41.206 17:33:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:41.206 17:33:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:41.466 17:33:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:11:41.466 17:33:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:41.466 17:33:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:11:41.466 17:33:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=none 00:11:41.466 17:33:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:41.466 17:33:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:41.466 17:33:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:41.466 17:33:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:41.466 17:33:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:41.466 17:33:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:41.733 17:33:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:41.733 "name": "raid_bdev1", 00:11:41.733 "uuid": "1a97666a-4e3c-44fa-99aa-5c7d2e702f16", 00:11:41.733 "strip_size_kb": 0, 00:11:41.733 "state": "online", 00:11:41.733 "raid_level": "raid1", 00:11:41.733 "superblock": false, 00:11:41.733 "num_base_bdevs": 2, 00:11:41.733 "num_base_bdevs_discovered": 1, 00:11:41.733 "num_base_bdevs_operational": 1, 00:11:41.733 "base_bdevs_list": [ 00:11:41.733 { 00:11:41.733 "name": null, 00:11:41.733 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:41.733 "is_configured": false, 00:11:41.733 "data_offset": 0, 00:11:41.733 "data_size": 65536 00:11:41.733 }, 00:11:41.733 { 00:11:41.733 "name": "BaseBdev2", 00:11:41.733 "uuid": "2940169b-2fd4-5c3e-b6c3-762b3f89ccb1", 00:11:41.733 "is_configured": true, 00:11:41.733 "data_offset": 0, 00:11:41.733 "data_size": 65536 00:11:41.733 } 00:11:41.733 ] 00:11:41.733 }' 00:11:41.733 17:33:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:41.733 17:33:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:11:41.733 17:33:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:41.733 17:33:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:11:41.733 17:33:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:11:41.733 17:33:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:41.733 17:33:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:41.733 [2024-11-27 17:33:12.763389] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:11:41.733 [2024-11-27 17:33:12.770849] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000d062f0 00:11:41.733 17:33:12 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:41.733 17:33:12 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@663 -- # sleep 1 00:11:41.733 [2024-11-27 17:33:12.773161] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:11:42.684 17:33:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:42.684 17:33:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:42.684 17:33:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:42.684 17:33:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:42.684 17:33:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:42.684 17:33:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:42.684 17:33:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:42.684 17:33:13 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:42.684 17:33:13 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:42.684 17:33:13 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:42.684 17:33:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:42.684 "name": "raid_bdev1", 00:11:42.684 "uuid": "1a97666a-4e3c-44fa-99aa-5c7d2e702f16", 00:11:42.684 "strip_size_kb": 0, 00:11:42.684 "state": "online", 00:11:42.684 "raid_level": "raid1", 00:11:42.684 "superblock": false, 00:11:42.684 "num_base_bdevs": 2, 00:11:42.684 "num_base_bdevs_discovered": 2, 00:11:42.684 "num_base_bdevs_operational": 2, 00:11:42.684 "process": { 00:11:42.684 "type": "rebuild", 00:11:42.684 "target": "spare", 00:11:42.684 "progress": { 00:11:42.684 "blocks": 20480, 00:11:42.684 "percent": 31 00:11:42.684 } 00:11:42.684 }, 00:11:42.684 "base_bdevs_list": [ 00:11:42.684 { 00:11:42.684 "name": "spare", 00:11:42.684 "uuid": "1ecf84b4-7847-59f9-a811-7a5c067a5db0", 00:11:42.684 "is_configured": true, 00:11:42.684 "data_offset": 0, 00:11:42.684 "data_size": 65536 00:11:42.684 }, 00:11:42.684 { 00:11:42.684 "name": "BaseBdev2", 00:11:42.684 "uuid": "2940169b-2fd4-5c3e-b6c3-762b3f89ccb1", 00:11:42.684 "is_configured": true, 00:11:42.684 "data_offset": 0, 00:11:42.684 "data_size": 65536 00:11:42.684 } 00:11:42.684 ] 00:11:42.684 }' 00:11:42.684 17:33:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:42.684 17:33:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:42.944 17:33:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:42.944 17:33:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:42.944 17:33:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@666 -- # '[' false = true ']' 00:11:42.944 17:33:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=2 00:11:42.944 17:33:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:11:42.944 17:33:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@693 -- # '[' 2 -gt 2 ']' 00:11:42.944 17:33:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@706 -- # local timeout=301 00:11:42.944 17:33:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:11:42.944 17:33:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:42.944 17:33:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:42.944 17:33:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:42.944 17:33:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:42.944 17:33:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:42.944 17:33:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:42.944 17:33:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:42.944 17:33:13 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:42.944 17:33:13 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:42.944 17:33:13 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:42.944 17:33:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:42.944 "name": "raid_bdev1", 00:11:42.944 "uuid": "1a97666a-4e3c-44fa-99aa-5c7d2e702f16", 00:11:42.944 "strip_size_kb": 0, 00:11:42.944 "state": "online", 00:11:42.944 "raid_level": "raid1", 00:11:42.944 "superblock": false, 00:11:42.944 "num_base_bdevs": 2, 00:11:42.944 "num_base_bdevs_discovered": 2, 00:11:42.944 "num_base_bdevs_operational": 2, 00:11:42.944 "process": { 00:11:42.944 "type": "rebuild", 00:11:42.944 "target": "spare", 00:11:42.944 "progress": { 00:11:42.944 "blocks": 22528, 00:11:42.944 "percent": 34 00:11:42.944 } 00:11:42.944 }, 00:11:42.944 "base_bdevs_list": [ 00:11:42.944 { 00:11:42.944 "name": "spare", 00:11:42.944 "uuid": "1ecf84b4-7847-59f9-a811-7a5c067a5db0", 00:11:42.944 "is_configured": true, 00:11:42.944 "data_offset": 0, 00:11:42.944 "data_size": 65536 00:11:42.944 }, 00:11:42.944 { 00:11:42.944 "name": "BaseBdev2", 00:11:42.944 "uuid": "2940169b-2fd4-5c3e-b6c3-762b3f89ccb1", 00:11:42.944 "is_configured": true, 00:11:42.944 "data_offset": 0, 00:11:42.944 "data_size": 65536 00:11:42.944 } 00:11:42.944 ] 00:11:42.944 }' 00:11:42.944 17:33:13 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:42.944 17:33:14 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:42.944 17:33:14 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:42.944 17:33:14 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:42.944 17:33:14 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:11:43.883 17:33:15 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:11:43.883 17:33:15 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:43.883 17:33:15 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:43.883 17:33:15 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:43.883 17:33:15 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:43.883 17:33:15 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:43.883 17:33:15 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:43.883 17:33:15 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:43.883 17:33:15 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:43.883 17:33:15 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:43.883 17:33:15 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:44.141 17:33:15 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:44.141 "name": "raid_bdev1", 00:11:44.141 "uuid": "1a97666a-4e3c-44fa-99aa-5c7d2e702f16", 00:11:44.141 "strip_size_kb": 0, 00:11:44.141 "state": "online", 00:11:44.141 "raid_level": "raid1", 00:11:44.141 "superblock": false, 00:11:44.141 "num_base_bdevs": 2, 00:11:44.141 "num_base_bdevs_discovered": 2, 00:11:44.141 "num_base_bdevs_operational": 2, 00:11:44.141 "process": { 00:11:44.141 "type": "rebuild", 00:11:44.141 "target": "spare", 00:11:44.141 "progress": { 00:11:44.141 "blocks": 45056, 00:11:44.141 "percent": 68 00:11:44.141 } 00:11:44.141 }, 00:11:44.141 "base_bdevs_list": [ 00:11:44.141 { 00:11:44.141 "name": "spare", 00:11:44.141 "uuid": "1ecf84b4-7847-59f9-a811-7a5c067a5db0", 00:11:44.141 "is_configured": true, 00:11:44.141 "data_offset": 0, 00:11:44.141 "data_size": 65536 00:11:44.141 }, 00:11:44.141 { 00:11:44.141 "name": "BaseBdev2", 00:11:44.141 "uuid": "2940169b-2fd4-5c3e-b6c3-762b3f89ccb1", 00:11:44.141 "is_configured": true, 00:11:44.141 "data_offset": 0, 00:11:44.141 "data_size": 65536 00:11:44.141 } 00:11:44.141 ] 00:11:44.141 }' 00:11:44.141 17:33:15 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:44.141 17:33:15 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:44.141 17:33:15 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:44.141 17:33:15 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:44.141 17:33:15 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:11:45.080 [2024-11-27 17:33:15.993696] bdev_raid.c:2896:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:11:45.080 [2024-11-27 17:33:15.993800] bdev_raid.c:2558:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:11:45.080 [2024-11-27 17:33:15.993852] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:45.080 17:33:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:11:45.080 17:33:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:45.080 17:33:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:45.080 17:33:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:45.080 17:33:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:45.080 17:33:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:45.080 17:33:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:45.080 17:33:16 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:45.080 17:33:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:45.080 17:33:16 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:45.080 17:33:16 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:45.080 17:33:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:45.080 "name": "raid_bdev1", 00:11:45.080 "uuid": "1a97666a-4e3c-44fa-99aa-5c7d2e702f16", 00:11:45.080 "strip_size_kb": 0, 00:11:45.080 "state": "online", 00:11:45.080 "raid_level": "raid1", 00:11:45.080 "superblock": false, 00:11:45.080 "num_base_bdevs": 2, 00:11:45.080 "num_base_bdevs_discovered": 2, 00:11:45.080 "num_base_bdevs_operational": 2, 00:11:45.080 "base_bdevs_list": [ 00:11:45.080 { 00:11:45.080 "name": "spare", 00:11:45.080 "uuid": "1ecf84b4-7847-59f9-a811-7a5c067a5db0", 00:11:45.080 "is_configured": true, 00:11:45.080 "data_offset": 0, 00:11:45.080 "data_size": 65536 00:11:45.080 }, 00:11:45.080 { 00:11:45.080 "name": "BaseBdev2", 00:11:45.080 "uuid": "2940169b-2fd4-5c3e-b6c3-762b3f89ccb1", 00:11:45.080 "is_configured": true, 00:11:45.080 "data_offset": 0, 00:11:45.080 "data_size": 65536 00:11:45.080 } 00:11:45.080 ] 00:11:45.080 }' 00:11:45.081 17:33:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:45.341 17:33:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:11:45.341 17:33:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:45.341 17:33:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:11:45.341 17:33:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@709 -- # break 00:11:45.341 17:33:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:11:45.341 17:33:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:45.341 17:33:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:11:45.341 17:33:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=none 00:11:45.341 17:33:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:45.341 17:33:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:45.341 17:33:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:45.341 17:33:16 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:45.341 17:33:16 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:45.341 17:33:16 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:45.341 17:33:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:45.341 "name": "raid_bdev1", 00:11:45.341 "uuid": "1a97666a-4e3c-44fa-99aa-5c7d2e702f16", 00:11:45.341 "strip_size_kb": 0, 00:11:45.341 "state": "online", 00:11:45.341 "raid_level": "raid1", 00:11:45.341 "superblock": false, 00:11:45.341 "num_base_bdevs": 2, 00:11:45.341 "num_base_bdevs_discovered": 2, 00:11:45.341 "num_base_bdevs_operational": 2, 00:11:45.341 "base_bdevs_list": [ 00:11:45.341 { 00:11:45.341 "name": "spare", 00:11:45.341 "uuid": "1ecf84b4-7847-59f9-a811-7a5c067a5db0", 00:11:45.341 "is_configured": true, 00:11:45.341 "data_offset": 0, 00:11:45.341 "data_size": 65536 00:11:45.341 }, 00:11:45.341 { 00:11:45.341 "name": "BaseBdev2", 00:11:45.341 "uuid": "2940169b-2fd4-5c3e-b6c3-762b3f89ccb1", 00:11:45.341 "is_configured": true, 00:11:45.341 "data_offset": 0, 00:11:45.341 "data_size": 65536 00:11:45.341 } 00:11:45.341 ] 00:11:45.341 }' 00:11:45.341 17:33:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:45.341 17:33:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:11:45.341 17:33:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:45.341 17:33:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:11:45.341 17:33:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:11:45.341 17:33:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:45.341 17:33:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:45.341 17:33:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:45.341 17:33:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:45.341 17:33:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:11:45.341 17:33:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:45.341 17:33:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:45.341 17:33:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:45.341 17:33:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:45.341 17:33:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:45.341 17:33:16 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:45.341 17:33:16 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:45.341 17:33:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:45.341 17:33:16 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:45.341 17:33:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:45.341 "name": "raid_bdev1", 00:11:45.341 "uuid": "1a97666a-4e3c-44fa-99aa-5c7d2e702f16", 00:11:45.341 "strip_size_kb": 0, 00:11:45.341 "state": "online", 00:11:45.341 "raid_level": "raid1", 00:11:45.341 "superblock": false, 00:11:45.341 "num_base_bdevs": 2, 00:11:45.341 "num_base_bdevs_discovered": 2, 00:11:45.341 "num_base_bdevs_operational": 2, 00:11:45.341 "base_bdevs_list": [ 00:11:45.341 { 00:11:45.341 "name": "spare", 00:11:45.341 "uuid": "1ecf84b4-7847-59f9-a811-7a5c067a5db0", 00:11:45.341 "is_configured": true, 00:11:45.341 "data_offset": 0, 00:11:45.341 "data_size": 65536 00:11:45.341 }, 00:11:45.341 { 00:11:45.341 "name": "BaseBdev2", 00:11:45.342 "uuid": "2940169b-2fd4-5c3e-b6c3-762b3f89ccb1", 00:11:45.342 "is_configured": true, 00:11:45.342 "data_offset": 0, 00:11:45.342 "data_size": 65536 00:11:45.342 } 00:11:45.342 ] 00:11:45.342 }' 00:11:45.342 17:33:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:45.342 17:33:16 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:45.912 17:33:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:11:45.912 17:33:16 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:45.912 17:33:16 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:45.912 [2024-11-27 17:33:16.935792] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:11:45.912 [2024-11-27 17:33:16.935866] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:11:45.912 [2024-11-27 17:33:16.935962] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:11:45.912 [2024-11-27 17:33:16.936038] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:11:45.912 [2024-11-27 17:33:16.936057] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:11:45.912 17:33:16 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:45.912 17:33:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:45.912 17:33:16 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:45.912 17:33:16 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:45.912 17:33:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@720 -- # jq length 00:11:45.912 17:33:16 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:45.912 17:33:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:11:45.912 17:33:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:11:45.912 17:33:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:11:45.912 17:33:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:11:45.912 17:33:16 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:11:45.912 17:33:16 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:11:45.912 17:33:16 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@10 -- # local bdev_list 00:11:45.912 17:33:16 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:11:45.912 17:33:16 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@11 -- # local nbd_list 00:11:45.912 17:33:16 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@12 -- # local i 00:11:45.912 17:33:16 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:11:45.912 17:33:16 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:11:45.912 17:33:16 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:11:46.172 /dev/nbd0 00:11:46.172 17:33:17 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:11:46.172 17:33:17 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:11:46.172 17:33:17 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:11:46.172 17:33:17 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@869 -- # local i 00:11:46.172 17:33:17 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:11:46.172 17:33:17 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:11:46.172 17:33:17 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:11:46.172 17:33:17 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@873 -- # break 00:11:46.172 17:33:17 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:11:46.172 17:33:17 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:11:46.172 17:33:17 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:11:46.172 1+0 records in 00:11:46.172 1+0 records out 00:11:46.172 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000309953 s, 13.2 MB/s 00:11:46.172 17:33:17 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:46.172 17:33:17 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@886 -- # size=4096 00:11:46.172 17:33:17 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:46.172 17:33:17 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:11:46.172 17:33:17 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@889 -- # return 0 00:11:46.172 17:33:17 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:11:46.172 17:33:17 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:11:46.172 17:33:17 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:11:46.433 /dev/nbd1 00:11:46.433 17:33:17 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:11:46.433 17:33:17 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:11:46.433 17:33:17 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@868 -- # local nbd_name=nbd1 00:11:46.433 17:33:17 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@869 -- # local i 00:11:46.433 17:33:17 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:11:46.433 17:33:17 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:11:46.433 17:33:17 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@872 -- # grep -q -w nbd1 /proc/partitions 00:11:46.433 17:33:17 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@873 -- # break 00:11:46.433 17:33:17 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:11:46.433 17:33:17 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:11:46.433 17:33:17 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@885 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:11:46.433 1+0 records in 00:11:46.433 1+0 records out 00:11:46.433 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000371357 s, 11.0 MB/s 00:11:46.433 17:33:17 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:46.433 17:33:17 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@886 -- # size=4096 00:11:46.433 17:33:17 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:46.433 17:33:17 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:11:46.433 17:33:17 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@889 -- # return 0 00:11:46.433 17:33:17 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:11:46.433 17:33:17 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:11:46.433 17:33:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@738 -- # cmp -i 0 /dev/nbd0 /dev/nbd1 00:11:46.433 17:33:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:11:46.433 17:33:17 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:11:46.433 17:33:17 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:11:46.433 17:33:17 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@50 -- # local nbd_list 00:11:46.433 17:33:17 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@51 -- # local i 00:11:46.433 17:33:17 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:11:46.433 17:33:17 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:11:46.694 17:33:17 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:11:46.694 17:33:17 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:11:46.694 17:33:17 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:11:46.694 17:33:17 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:11:46.694 17:33:17 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:11:46.694 17:33:17 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:11:46.694 17:33:17 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:11:46.694 17:33:17 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:11:46.694 17:33:17 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:11:46.694 17:33:17 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:11:46.955 17:33:18 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:11:46.955 17:33:18 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:11:46.955 17:33:18 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:11:46.955 17:33:18 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:11:46.955 17:33:18 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:11:46.955 17:33:18 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:11:46.955 17:33:18 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:11:46.955 17:33:18 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:11:46.955 17:33:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@743 -- # '[' false = true ']' 00:11:46.955 17:33:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@784 -- # killprocess 85885 00:11:46.955 17:33:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@950 -- # '[' -z 85885 ']' 00:11:46.955 17:33:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@954 -- # kill -0 85885 00:11:46.955 17:33:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@955 -- # uname 00:11:46.955 17:33:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:11:46.955 17:33:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 85885 00:11:46.955 17:33:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:11:46.955 17:33:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:11:46.955 killing process with pid 85885 00:11:46.955 17:33:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 85885' 00:11:46.955 17:33:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@969 -- # kill 85885 00:11:46.955 Received shutdown signal, test time was about 60.000000 seconds 00:11:46.955 00:11:46.955 Latency(us) 00:11:46.955 [2024-11-27T17:33:18.147Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:11:46.955 [2024-11-27T17:33:18.147Z] =================================================================================================================== 00:11:46.955 [2024-11-27T17:33:18.147Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:11:46.955 [2024-11-27 17:33:18.084897] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:11:46.955 17:33:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@974 -- # wait 85885 00:11:46.955 [2024-11-27 17:33:18.141379] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:11:47.526 17:33:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@786 -- # return 0 00:11:47.526 00:11:47.526 real 0m13.700s 00:11:47.526 user 0m15.954s 00:11:47.526 sys 0m2.904s 00:11:47.526 17:33:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:11:47.526 17:33:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:11:47.526 ************************************ 00:11:47.526 END TEST raid_rebuild_test 00:11:47.526 ************************************ 00:11:47.526 17:33:18 bdev_raid -- bdev/bdev_raid.sh@979 -- # run_test raid_rebuild_test_sb raid_rebuild_test raid1 2 true false true 00:11:47.526 17:33:18 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 7 -le 1 ']' 00:11:47.526 17:33:18 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:11:47.526 17:33:18 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:11:47.526 ************************************ 00:11:47.526 START TEST raid_rebuild_test_sb 00:11:47.526 ************************************ 00:11:47.526 17:33:18 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@1125 -- # raid_rebuild_test raid1 2 true false true 00:11:47.526 17:33:18 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:11:47.526 17:33:18 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=2 00:11:47.526 17:33:18 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:11:47.526 17:33:18 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:11:47.526 17:33:18 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@573 -- # local verify=true 00:11:47.526 17:33:18 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:11:47.526 17:33:18 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:11:47.526 17:33:18 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:11:47.526 17:33:18 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:11:47.526 17:33:18 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:11:47.527 17:33:18 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:11:47.527 17:33:18 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:11:47.527 17:33:18 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:11:47.527 17:33:18 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:11:47.527 17:33:18 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:11:47.527 17:33:18 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:11:47.527 17:33:18 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # local strip_size 00:11:47.527 17:33:18 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@577 -- # local create_arg 00:11:47.527 17:33:18 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:11:47.527 17:33:18 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@579 -- # local data_offset 00:11:47.527 17:33:18 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:11:47.527 17:33:18 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:11:47.527 17:33:18 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:11:47.527 17:33:18 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:11:47.527 17:33:18 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@597 -- # raid_pid=86285 00:11:47.527 17:33:18 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:11:47.527 17:33:18 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@598 -- # waitforlisten 86285 00:11:47.527 17:33:18 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@831 -- # '[' -z 86285 ']' 00:11:47.527 17:33:18 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:11:47.527 17:33:18 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@836 -- # local max_retries=100 00:11:47.527 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:11:47.527 17:33:18 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:11:47.527 17:33:18 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@840 -- # xtrace_disable 00:11:47.527 17:33:18 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:47.527 I/O size of 3145728 is greater than zero copy threshold (65536). 00:11:47.527 Zero copy mechanism will not be used. 00:11:47.527 [2024-11-27 17:33:18.671591] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:11:47.527 [2024-11-27 17:33:18.671732] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid86285 ] 00:11:47.787 [2024-11-27 17:33:18.816826] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:11:47.787 [2024-11-27 17:33:18.884481] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:11:47.787 [2024-11-27 17:33:18.959677] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:47.787 [2024-11-27 17:33:18.959743] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:11:48.356 17:33:19 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:11:48.356 17:33:19 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@864 -- # return 0 00:11:48.357 17:33:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:11:48.357 17:33:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:11:48.357 17:33:19 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:48.357 17:33:19 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:48.357 BaseBdev1_malloc 00:11:48.357 17:33:19 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:48.357 17:33:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:11:48.357 17:33:19 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:48.357 17:33:19 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:48.357 [2024-11-27 17:33:19.517774] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:11:48.357 [2024-11-27 17:33:19.517855] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:48.357 [2024-11-27 17:33:19.517880] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:11:48.357 [2024-11-27 17:33:19.517896] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:48.357 [2024-11-27 17:33:19.520411] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:48.357 [2024-11-27 17:33:19.520444] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:11:48.357 BaseBdev1 00:11:48.357 17:33:19 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:48.357 17:33:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:11:48.357 17:33:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:11:48.357 17:33:19 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:48.357 17:33:19 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:48.617 BaseBdev2_malloc 00:11:48.617 17:33:19 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:48.617 17:33:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:11:48.617 17:33:19 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:48.617 17:33:19 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:48.617 [2024-11-27 17:33:19.564283] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:11:48.617 [2024-11-27 17:33:19.564358] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:48.617 [2024-11-27 17:33:19.564389] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:11:48.617 [2024-11-27 17:33:19.564404] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:48.617 [2024-11-27 17:33:19.568005] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:48.617 [2024-11-27 17:33:19.568054] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:11:48.617 BaseBdev2 00:11:48.617 17:33:19 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:48.617 17:33:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:11:48.617 17:33:19 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:48.617 17:33:19 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:48.617 spare_malloc 00:11:48.617 17:33:19 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:48.617 17:33:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:11:48.617 17:33:19 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:48.617 17:33:19 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:48.617 spare_delay 00:11:48.617 17:33:19 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:48.617 17:33:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:11:48.617 17:33:19 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:48.617 17:33:19 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:48.617 [2024-11-27 17:33:19.611341] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:11:48.617 [2024-11-27 17:33:19.611394] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:11:48.617 [2024-11-27 17:33:19.611433] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:11:48.617 [2024-11-27 17:33:19.611441] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:11:48.617 [2024-11-27 17:33:19.613823] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:11:48.617 [2024-11-27 17:33:19.613855] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:11:48.617 spare 00:11:48.617 17:33:19 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:48.617 17:33:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 00:11:48.617 17:33:19 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:48.617 17:33:19 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:48.617 [2024-11-27 17:33:19.623379] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:11:48.617 [2024-11-27 17:33:19.625532] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:11:48.617 [2024-11-27 17:33:19.625689] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:11:48.617 [2024-11-27 17:33:19.625701] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:11:48.617 [2024-11-27 17:33:19.625992] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:11:48.617 [2024-11-27 17:33:19.626169] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:11:48.617 [2024-11-27 17:33:19.626192] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:11:48.617 [2024-11-27 17:33:19.626325] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:48.617 17:33:19 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:48.618 17:33:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:11:48.618 17:33:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:48.618 17:33:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:48.618 17:33:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:48.618 17:33:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:48.618 17:33:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:11:48.618 17:33:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:48.618 17:33:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:48.618 17:33:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:48.618 17:33:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:48.618 17:33:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:48.618 17:33:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:48.618 17:33:19 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:48.618 17:33:19 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:48.618 17:33:19 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:48.618 17:33:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:48.618 "name": "raid_bdev1", 00:11:48.618 "uuid": "1e7dc017-2c9f-4f39-a1ba-504c45cdbd67", 00:11:48.618 "strip_size_kb": 0, 00:11:48.618 "state": "online", 00:11:48.618 "raid_level": "raid1", 00:11:48.618 "superblock": true, 00:11:48.618 "num_base_bdevs": 2, 00:11:48.618 "num_base_bdevs_discovered": 2, 00:11:48.618 "num_base_bdevs_operational": 2, 00:11:48.618 "base_bdevs_list": [ 00:11:48.618 { 00:11:48.618 "name": "BaseBdev1", 00:11:48.618 "uuid": "0d12821f-a5d3-57d8-a144-488f799728e1", 00:11:48.618 "is_configured": true, 00:11:48.618 "data_offset": 2048, 00:11:48.618 "data_size": 63488 00:11:48.618 }, 00:11:48.618 { 00:11:48.618 "name": "BaseBdev2", 00:11:48.618 "uuid": "4fe7c978-6106-55d6-8587-c40656640823", 00:11:48.618 "is_configured": true, 00:11:48.618 "data_offset": 2048, 00:11:48.618 "data_size": 63488 00:11:48.618 } 00:11:48.618 ] 00:11:48.618 }' 00:11:48.618 17:33:19 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:48.618 17:33:19 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:48.878 17:33:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:11:48.878 17:33:20 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:48.878 17:33:20 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:48.878 17:33:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:11:48.878 [2024-11-27 17:33:20.030911] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:11:48.878 17:33:20 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:49.138 17:33:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=63488 00:11:49.138 17:33:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:49.138 17:33:20 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:49.138 17:33:20 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:49.138 17:33:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:11:49.138 17:33:20 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:49.138 17:33:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # data_offset=2048 00:11:49.138 17:33:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:11:49.138 17:33:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:11:49.138 17:33:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:11:49.138 17:33:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:11:49.138 17:33:20 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:11:49.138 17:33:20 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:11:49.138 17:33:20 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # local bdev_list 00:11:49.138 17:33:20 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:11:49.138 17:33:20 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # local nbd_list 00:11:49.138 17:33:20 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@12 -- # local i 00:11:49.138 17:33:20 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:11:49.138 17:33:20 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:11:49.138 17:33:20 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:11:49.138 [2024-11-27 17:33:20.318376] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:11:49.399 /dev/nbd0 00:11:49.399 17:33:20 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:11:49.399 17:33:20 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:11:49.399 17:33:20 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:11:49.399 17:33:20 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@869 -- # local i 00:11:49.399 17:33:20 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:11:49.399 17:33:20 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:11:49.399 17:33:20 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:11:49.399 17:33:20 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@873 -- # break 00:11:49.399 17:33:20 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:11:49.399 17:33:20 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:11:49.399 17:33:20 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:11:49.399 1+0 records in 00:11:49.399 1+0 records out 00:11:49.399 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.0004503 s, 9.1 MB/s 00:11:49.399 17:33:20 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:49.399 17:33:20 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@886 -- # size=4096 00:11:49.399 17:33:20 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:11:49.399 17:33:20 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:11:49.399 17:33:20 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@889 -- # return 0 00:11:49.399 17:33:20 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:11:49.399 17:33:20 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:11:49.399 17:33:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@629 -- # '[' raid1 = raid5f ']' 00:11:49.399 17:33:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@633 -- # write_unit_size=1 00:11:49.399 17:33:20 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=512 count=63488 oflag=direct 00:11:53.601 63488+0 records in 00:11:53.601 63488+0 records out 00:11:53.601 32505856 bytes (33 MB, 31 MiB) copied, 3.94587 s, 8.2 MB/s 00:11:53.601 17:33:24 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:11:53.601 17:33:24 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:11:53.601 17:33:24 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:11:53.601 17:33:24 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # local nbd_list 00:11:53.601 17:33:24 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@51 -- # local i 00:11:53.601 17:33:24 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:11:53.601 17:33:24 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:11:53.601 [2024-11-27 17:33:24.517680] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:53.601 17:33:24 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:11:53.601 17:33:24 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:11:53.601 17:33:24 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:11:53.601 17:33:24 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:11:53.601 17:33:24 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:11:53.601 17:33:24 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:11:53.601 17:33:24 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:11:53.601 17:33:24 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:11:53.601 17:33:24 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:11:53.601 17:33:24 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:53.601 17:33:24 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:53.601 [2024-11-27 17:33:24.549693] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:11:53.601 17:33:24 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:53.601 17:33:24 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:11:53.601 17:33:24 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:53.601 17:33:24 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:53.601 17:33:24 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:53.601 17:33:24 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:53.601 17:33:24 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:11:53.601 17:33:24 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:53.601 17:33:24 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:53.601 17:33:24 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:53.601 17:33:24 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:53.601 17:33:24 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:53.601 17:33:24 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:53.601 17:33:24 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:53.601 17:33:24 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:53.601 17:33:24 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:53.601 17:33:24 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:53.601 "name": "raid_bdev1", 00:11:53.601 "uuid": "1e7dc017-2c9f-4f39-a1ba-504c45cdbd67", 00:11:53.601 "strip_size_kb": 0, 00:11:53.601 "state": "online", 00:11:53.601 "raid_level": "raid1", 00:11:53.601 "superblock": true, 00:11:53.601 "num_base_bdevs": 2, 00:11:53.601 "num_base_bdevs_discovered": 1, 00:11:53.601 "num_base_bdevs_operational": 1, 00:11:53.601 "base_bdevs_list": [ 00:11:53.601 { 00:11:53.601 "name": null, 00:11:53.601 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:53.601 "is_configured": false, 00:11:53.601 "data_offset": 0, 00:11:53.601 "data_size": 63488 00:11:53.601 }, 00:11:53.601 { 00:11:53.601 "name": "BaseBdev2", 00:11:53.601 "uuid": "4fe7c978-6106-55d6-8587-c40656640823", 00:11:53.601 "is_configured": true, 00:11:53.601 "data_offset": 2048, 00:11:53.601 "data_size": 63488 00:11:53.601 } 00:11:53.601 ] 00:11:53.601 }' 00:11:53.601 17:33:24 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:53.601 17:33:24 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:53.861 17:33:24 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:11:53.862 17:33:24 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:53.862 17:33:24 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:53.862 [2024-11-27 17:33:25.000958] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:11:53.862 [2024-11-27 17:33:25.008254] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000c3e280 00:11:53.862 17:33:25 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:53.862 17:33:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@647 -- # sleep 1 00:11:53.862 [2024-11-27 17:33:25.010496] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:11:55.245 17:33:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:55.245 17:33:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:55.245 17:33:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:55.245 17:33:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:55.245 17:33:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:55.245 17:33:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:55.245 17:33:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:55.245 17:33:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:55.245 17:33:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:55.245 17:33:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:55.245 17:33:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:55.245 "name": "raid_bdev1", 00:11:55.245 "uuid": "1e7dc017-2c9f-4f39-a1ba-504c45cdbd67", 00:11:55.245 "strip_size_kb": 0, 00:11:55.245 "state": "online", 00:11:55.245 "raid_level": "raid1", 00:11:55.245 "superblock": true, 00:11:55.245 "num_base_bdevs": 2, 00:11:55.245 "num_base_bdevs_discovered": 2, 00:11:55.245 "num_base_bdevs_operational": 2, 00:11:55.245 "process": { 00:11:55.245 "type": "rebuild", 00:11:55.245 "target": "spare", 00:11:55.245 "progress": { 00:11:55.245 "blocks": 20480, 00:11:55.245 "percent": 32 00:11:55.245 } 00:11:55.245 }, 00:11:55.245 "base_bdevs_list": [ 00:11:55.245 { 00:11:55.245 "name": "spare", 00:11:55.245 "uuid": "a4be28a5-7c99-5f92-a551-3de87fd43090", 00:11:55.245 "is_configured": true, 00:11:55.245 "data_offset": 2048, 00:11:55.245 "data_size": 63488 00:11:55.245 }, 00:11:55.245 { 00:11:55.245 "name": "BaseBdev2", 00:11:55.245 "uuid": "4fe7c978-6106-55d6-8587-c40656640823", 00:11:55.245 "is_configured": true, 00:11:55.245 "data_offset": 2048, 00:11:55.245 "data_size": 63488 00:11:55.245 } 00:11:55.245 ] 00:11:55.245 }' 00:11:55.245 17:33:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:55.245 17:33:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:55.245 17:33:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:55.245 17:33:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:55.245 17:33:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:11:55.245 17:33:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:55.245 17:33:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:55.245 [2024-11-27 17:33:26.158886] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:11:55.245 [2024-11-27 17:33:26.218748] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:11:55.245 [2024-11-27 17:33:26.218849] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:55.245 [2024-11-27 17:33:26.218873] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:11:55.245 [2024-11-27 17:33:26.218882] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:11:55.245 17:33:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:55.245 17:33:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:11:55.245 17:33:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:55.245 17:33:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:55.245 17:33:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:55.245 17:33:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:55.245 17:33:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:11:55.245 17:33:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:55.245 17:33:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:55.245 17:33:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:55.245 17:33:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:55.245 17:33:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:55.245 17:33:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:55.245 17:33:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:55.245 17:33:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:55.245 17:33:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:55.245 17:33:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:55.245 "name": "raid_bdev1", 00:11:55.245 "uuid": "1e7dc017-2c9f-4f39-a1ba-504c45cdbd67", 00:11:55.245 "strip_size_kb": 0, 00:11:55.245 "state": "online", 00:11:55.245 "raid_level": "raid1", 00:11:55.245 "superblock": true, 00:11:55.245 "num_base_bdevs": 2, 00:11:55.245 "num_base_bdevs_discovered": 1, 00:11:55.245 "num_base_bdevs_operational": 1, 00:11:55.245 "base_bdevs_list": [ 00:11:55.245 { 00:11:55.245 "name": null, 00:11:55.245 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:55.245 "is_configured": false, 00:11:55.245 "data_offset": 0, 00:11:55.245 "data_size": 63488 00:11:55.245 }, 00:11:55.245 { 00:11:55.245 "name": "BaseBdev2", 00:11:55.245 "uuid": "4fe7c978-6106-55d6-8587-c40656640823", 00:11:55.245 "is_configured": true, 00:11:55.245 "data_offset": 2048, 00:11:55.245 "data_size": 63488 00:11:55.245 } 00:11:55.245 ] 00:11:55.245 }' 00:11:55.245 17:33:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:55.245 17:33:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:55.505 17:33:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:11:55.505 17:33:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:55.505 17:33:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:11:55.505 17:33:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:11:55.505 17:33:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:55.505 17:33:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:55.505 17:33:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:55.505 17:33:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:55.505 17:33:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:55.766 17:33:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:55.766 17:33:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:55.766 "name": "raid_bdev1", 00:11:55.766 "uuid": "1e7dc017-2c9f-4f39-a1ba-504c45cdbd67", 00:11:55.766 "strip_size_kb": 0, 00:11:55.766 "state": "online", 00:11:55.766 "raid_level": "raid1", 00:11:55.766 "superblock": true, 00:11:55.766 "num_base_bdevs": 2, 00:11:55.766 "num_base_bdevs_discovered": 1, 00:11:55.766 "num_base_bdevs_operational": 1, 00:11:55.766 "base_bdevs_list": [ 00:11:55.766 { 00:11:55.766 "name": null, 00:11:55.766 "uuid": "00000000-0000-0000-0000-000000000000", 00:11:55.766 "is_configured": false, 00:11:55.766 "data_offset": 0, 00:11:55.766 "data_size": 63488 00:11:55.766 }, 00:11:55.766 { 00:11:55.766 "name": "BaseBdev2", 00:11:55.766 "uuid": "4fe7c978-6106-55d6-8587-c40656640823", 00:11:55.766 "is_configured": true, 00:11:55.766 "data_offset": 2048, 00:11:55.766 "data_size": 63488 00:11:55.766 } 00:11:55.766 ] 00:11:55.766 }' 00:11:55.766 17:33:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:55.766 17:33:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:11:55.766 17:33:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:55.766 17:33:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:11:55.766 17:33:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:11:55.766 17:33:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:55.766 17:33:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:55.766 [2024-11-27 17:33:26.841374] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:11:55.766 [2024-11-27 17:33:26.848715] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000c3e350 00:11:55.766 17:33:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:55.766 17:33:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@663 -- # sleep 1 00:11:55.766 [2024-11-27 17:33:26.850997] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:11:56.705 17:33:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:56.705 17:33:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:56.705 17:33:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:56.705 17:33:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:56.705 17:33:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:56.705 17:33:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:56.705 17:33:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:56.705 17:33:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:56.705 17:33:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:56.705 17:33:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:56.964 17:33:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:56.964 "name": "raid_bdev1", 00:11:56.964 "uuid": "1e7dc017-2c9f-4f39-a1ba-504c45cdbd67", 00:11:56.964 "strip_size_kb": 0, 00:11:56.964 "state": "online", 00:11:56.964 "raid_level": "raid1", 00:11:56.964 "superblock": true, 00:11:56.964 "num_base_bdevs": 2, 00:11:56.964 "num_base_bdevs_discovered": 2, 00:11:56.964 "num_base_bdevs_operational": 2, 00:11:56.964 "process": { 00:11:56.964 "type": "rebuild", 00:11:56.964 "target": "spare", 00:11:56.964 "progress": { 00:11:56.964 "blocks": 20480, 00:11:56.964 "percent": 32 00:11:56.964 } 00:11:56.964 }, 00:11:56.964 "base_bdevs_list": [ 00:11:56.964 { 00:11:56.964 "name": "spare", 00:11:56.964 "uuid": "a4be28a5-7c99-5f92-a551-3de87fd43090", 00:11:56.964 "is_configured": true, 00:11:56.964 "data_offset": 2048, 00:11:56.964 "data_size": 63488 00:11:56.964 }, 00:11:56.964 { 00:11:56.964 "name": "BaseBdev2", 00:11:56.964 "uuid": "4fe7c978-6106-55d6-8587-c40656640823", 00:11:56.964 "is_configured": true, 00:11:56.964 "data_offset": 2048, 00:11:56.964 "data_size": 63488 00:11:56.964 } 00:11:56.964 ] 00:11:56.964 }' 00:11:56.964 17:33:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:56.964 17:33:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:56.964 17:33:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:56.964 17:33:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:56.964 17:33:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:11:56.964 17:33:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:11:56.964 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:11:56.964 17:33:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=2 00:11:56.964 17:33:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:11:56.964 17:33:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@693 -- # '[' 2 -gt 2 ']' 00:11:56.964 17:33:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@706 -- # local timeout=315 00:11:56.964 17:33:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:11:56.964 17:33:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:56.964 17:33:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:56.964 17:33:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:56.964 17:33:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:56.964 17:33:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:56.964 17:33:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:56.964 17:33:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:56.964 17:33:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:56.964 17:33:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:56.964 17:33:28 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:56.964 17:33:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:56.964 "name": "raid_bdev1", 00:11:56.964 "uuid": "1e7dc017-2c9f-4f39-a1ba-504c45cdbd67", 00:11:56.964 "strip_size_kb": 0, 00:11:56.964 "state": "online", 00:11:56.964 "raid_level": "raid1", 00:11:56.964 "superblock": true, 00:11:56.964 "num_base_bdevs": 2, 00:11:56.964 "num_base_bdevs_discovered": 2, 00:11:56.964 "num_base_bdevs_operational": 2, 00:11:56.964 "process": { 00:11:56.964 "type": "rebuild", 00:11:56.964 "target": "spare", 00:11:56.964 "progress": { 00:11:56.964 "blocks": 22528, 00:11:56.964 "percent": 35 00:11:56.964 } 00:11:56.964 }, 00:11:56.964 "base_bdevs_list": [ 00:11:56.964 { 00:11:56.964 "name": "spare", 00:11:56.964 "uuid": "a4be28a5-7c99-5f92-a551-3de87fd43090", 00:11:56.964 "is_configured": true, 00:11:56.964 "data_offset": 2048, 00:11:56.964 "data_size": 63488 00:11:56.964 }, 00:11:56.964 { 00:11:56.964 "name": "BaseBdev2", 00:11:56.964 "uuid": "4fe7c978-6106-55d6-8587-c40656640823", 00:11:56.964 "is_configured": true, 00:11:56.964 "data_offset": 2048, 00:11:56.964 "data_size": 63488 00:11:56.964 } 00:11:56.964 ] 00:11:56.964 }' 00:11:56.964 17:33:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:56.964 17:33:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:56.964 17:33:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:56.964 17:33:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:56.964 17:33:28 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:11:58.345 17:33:29 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:11:58.345 17:33:29 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:58.345 17:33:29 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:58.345 17:33:29 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:58.345 17:33:29 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:58.345 17:33:29 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:58.345 17:33:29 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:58.345 17:33:29 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:58.345 17:33:29 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:58.345 17:33:29 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:58.345 17:33:29 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:58.345 17:33:29 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:58.345 "name": "raid_bdev1", 00:11:58.345 "uuid": "1e7dc017-2c9f-4f39-a1ba-504c45cdbd67", 00:11:58.345 "strip_size_kb": 0, 00:11:58.345 "state": "online", 00:11:58.345 "raid_level": "raid1", 00:11:58.345 "superblock": true, 00:11:58.345 "num_base_bdevs": 2, 00:11:58.345 "num_base_bdevs_discovered": 2, 00:11:58.345 "num_base_bdevs_operational": 2, 00:11:58.345 "process": { 00:11:58.345 "type": "rebuild", 00:11:58.345 "target": "spare", 00:11:58.345 "progress": { 00:11:58.345 "blocks": 45056, 00:11:58.345 "percent": 70 00:11:58.345 } 00:11:58.345 }, 00:11:58.345 "base_bdevs_list": [ 00:11:58.345 { 00:11:58.345 "name": "spare", 00:11:58.345 "uuid": "a4be28a5-7c99-5f92-a551-3de87fd43090", 00:11:58.345 "is_configured": true, 00:11:58.345 "data_offset": 2048, 00:11:58.345 "data_size": 63488 00:11:58.345 }, 00:11:58.345 { 00:11:58.345 "name": "BaseBdev2", 00:11:58.345 "uuid": "4fe7c978-6106-55d6-8587-c40656640823", 00:11:58.345 "is_configured": true, 00:11:58.345 "data_offset": 2048, 00:11:58.345 "data_size": 63488 00:11:58.345 } 00:11:58.345 ] 00:11:58.345 }' 00:11:58.345 17:33:29 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:58.345 17:33:29 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:11:58.345 17:33:29 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:58.345 17:33:29 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:11:58.345 17:33:29 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:11:58.915 [2024-11-27 17:33:29.970663] bdev_raid.c:2896:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:11:58.915 [2024-11-27 17:33:29.970752] bdev_raid.c:2558:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:11:58.915 [2024-11-27 17:33:29.970914] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:11:59.174 17:33:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:11:59.174 17:33:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:11:59.174 17:33:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:59.174 17:33:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:11:59.174 17:33:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:11:59.174 17:33:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:59.174 17:33:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:59.174 17:33:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:59.174 17:33:30 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:59.174 17:33:30 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:59.174 17:33:30 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:59.174 17:33:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:59.174 "name": "raid_bdev1", 00:11:59.174 "uuid": "1e7dc017-2c9f-4f39-a1ba-504c45cdbd67", 00:11:59.174 "strip_size_kb": 0, 00:11:59.174 "state": "online", 00:11:59.174 "raid_level": "raid1", 00:11:59.174 "superblock": true, 00:11:59.174 "num_base_bdevs": 2, 00:11:59.174 "num_base_bdevs_discovered": 2, 00:11:59.174 "num_base_bdevs_operational": 2, 00:11:59.174 "base_bdevs_list": [ 00:11:59.174 { 00:11:59.174 "name": "spare", 00:11:59.174 "uuid": "a4be28a5-7c99-5f92-a551-3de87fd43090", 00:11:59.174 "is_configured": true, 00:11:59.174 "data_offset": 2048, 00:11:59.174 "data_size": 63488 00:11:59.174 }, 00:11:59.174 { 00:11:59.174 "name": "BaseBdev2", 00:11:59.174 "uuid": "4fe7c978-6106-55d6-8587-c40656640823", 00:11:59.174 "is_configured": true, 00:11:59.174 "data_offset": 2048, 00:11:59.174 "data_size": 63488 00:11:59.174 } 00:11:59.174 ] 00:11:59.174 }' 00:11:59.174 17:33:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:59.174 17:33:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:11:59.435 17:33:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:59.435 17:33:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:11:59.435 17:33:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@709 -- # break 00:11:59.435 17:33:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:11:59.435 17:33:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:11:59.435 17:33:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:11:59.435 17:33:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:11:59.435 17:33:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:11:59.435 17:33:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:59.435 17:33:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:59.435 17:33:30 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:59.435 17:33:30 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:59.435 17:33:30 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:59.435 17:33:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:11:59.435 "name": "raid_bdev1", 00:11:59.435 "uuid": "1e7dc017-2c9f-4f39-a1ba-504c45cdbd67", 00:11:59.435 "strip_size_kb": 0, 00:11:59.435 "state": "online", 00:11:59.435 "raid_level": "raid1", 00:11:59.435 "superblock": true, 00:11:59.435 "num_base_bdevs": 2, 00:11:59.435 "num_base_bdevs_discovered": 2, 00:11:59.435 "num_base_bdevs_operational": 2, 00:11:59.435 "base_bdevs_list": [ 00:11:59.435 { 00:11:59.435 "name": "spare", 00:11:59.435 "uuid": "a4be28a5-7c99-5f92-a551-3de87fd43090", 00:11:59.435 "is_configured": true, 00:11:59.435 "data_offset": 2048, 00:11:59.435 "data_size": 63488 00:11:59.435 }, 00:11:59.435 { 00:11:59.435 "name": "BaseBdev2", 00:11:59.435 "uuid": "4fe7c978-6106-55d6-8587-c40656640823", 00:11:59.435 "is_configured": true, 00:11:59.435 "data_offset": 2048, 00:11:59.435 "data_size": 63488 00:11:59.435 } 00:11:59.435 ] 00:11:59.435 }' 00:11:59.435 17:33:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:11:59.435 17:33:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:11:59.435 17:33:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:11:59.435 17:33:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:11:59.435 17:33:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:11:59.435 17:33:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:11:59.435 17:33:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:11:59.435 17:33:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:11:59.435 17:33:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:11:59.435 17:33:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:11:59.435 17:33:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:11:59.435 17:33:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:11:59.435 17:33:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:11:59.435 17:33:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:11:59.435 17:33:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:11:59.435 17:33:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:11:59.436 17:33:30 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:59.436 17:33:30 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:11:59.436 17:33:30 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:59.436 17:33:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:11:59.436 "name": "raid_bdev1", 00:11:59.436 "uuid": "1e7dc017-2c9f-4f39-a1ba-504c45cdbd67", 00:11:59.436 "strip_size_kb": 0, 00:11:59.436 "state": "online", 00:11:59.436 "raid_level": "raid1", 00:11:59.436 "superblock": true, 00:11:59.436 "num_base_bdevs": 2, 00:11:59.436 "num_base_bdevs_discovered": 2, 00:11:59.436 "num_base_bdevs_operational": 2, 00:11:59.436 "base_bdevs_list": [ 00:11:59.436 { 00:11:59.436 "name": "spare", 00:11:59.436 "uuid": "a4be28a5-7c99-5f92-a551-3de87fd43090", 00:11:59.436 "is_configured": true, 00:11:59.436 "data_offset": 2048, 00:11:59.436 "data_size": 63488 00:11:59.436 }, 00:11:59.436 { 00:11:59.436 "name": "BaseBdev2", 00:11:59.436 "uuid": "4fe7c978-6106-55d6-8587-c40656640823", 00:11:59.436 "is_configured": true, 00:11:59.436 "data_offset": 2048, 00:11:59.436 "data_size": 63488 00:11:59.436 } 00:11:59.436 ] 00:11:59.436 }' 00:11:59.436 17:33:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:11:59.436 17:33:30 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:00.006 17:33:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:12:00.006 17:33:30 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:00.006 17:33:30 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:00.006 [2024-11-27 17:33:30.936064] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:12:00.006 [2024-11-27 17:33:30.936109] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:12:00.006 [2024-11-27 17:33:30.936228] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:12:00.006 [2024-11-27 17:33:30.936308] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:12:00.006 [2024-11-27 17:33:30.936330] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:12:00.006 17:33:30 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:00.006 17:33:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:00.006 17:33:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # jq length 00:12:00.006 17:33:30 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:00.006 17:33:30 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:00.006 17:33:30 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:00.006 17:33:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:12:00.006 17:33:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:12:00.006 17:33:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:12:00.006 17:33:30 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:12:00.006 17:33:30 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:12:00.006 17:33:30 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:12:00.006 17:33:30 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # local bdev_list 00:12:00.006 17:33:30 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:12:00.006 17:33:30 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # local nbd_list 00:12:00.006 17:33:30 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@12 -- # local i 00:12:00.006 17:33:30 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:12:00.006 17:33:30 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:12:00.006 17:33:30 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:12:00.006 /dev/nbd0 00:12:00.267 17:33:31 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:12:00.267 17:33:31 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:12:00.267 17:33:31 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:12:00.267 17:33:31 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@869 -- # local i 00:12:00.267 17:33:31 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:12:00.267 17:33:31 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:12:00.267 17:33:31 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:12:00.268 17:33:31 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@873 -- # break 00:12:00.268 17:33:31 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:12:00.268 17:33:31 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:12:00.268 17:33:31 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:12:00.268 1+0 records in 00:12:00.268 1+0 records out 00:12:00.268 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000326147 s, 12.6 MB/s 00:12:00.268 17:33:31 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:00.268 17:33:31 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@886 -- # size=4096 00:12:00.268 17:33:31 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:00.268 17:33:31 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:12:00.268 17:33:31 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@889 -- # return 0 00:12:00.268 17:33:31 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:12:00.268 17:33:31 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:12:00.268 17:33:31 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:12:00.268 /dev/nbd1 00:12:00.268 17:33:31 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:12:00.268 17:33:31 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:12:00.268 17:33:31 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@868 -- # local nbd_name=nbd1 00:12:00.268 17:33:31 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@869 -- # local i 00:12:00.268 17:33:31 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:12:00.268 17:33:31 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:12:00.268 17:33:31 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@872 -- # grep -q -w nbd1 /proc/partitions 00:12:00.268 17:33:31 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@873 -- # break 00:12:00.268 17:33:31 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:12:00.268 17:33:31 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:12:00.268 17:33:31 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@885 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:12:00.268 1+0 records in 00:12:00.268 1+0 records out 00:12:00.268 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000272923 s, 15.0 MB/s 00:12:00.268 17:33:31 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:00.528 17:33:31 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@886 -- # size=4096 00:12:00.528 17:33:31 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:00.528 17:33:31 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:12:00.528 17:33:31 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@889 -- # return 0 00:12:00.528 17:33:31 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:12:00.528 17:33:31 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:12:00.528 17:33:31 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@738 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:12:00.528 17:33:31 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:12:00.529 17:33:31 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:12:00.529 17:33:31 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:12:00.529 17:33:31 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # local nbd_list 00:12:00.529 17:33:31 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@51 -- # local i 00:12:00.529 17:33:31 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:12:00.529 17:33:31 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:12:00.789 17:33:31 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:12:00.789 17:33:31 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:12:00.789 17:33:31 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:12:00.789 17:33:31 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:12:00.789 17:33:31 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:12:00.789 17:33:31 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:12:00.789 17:33:31 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:12:00.789 17:33:31 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:12:00.789 17:33:31 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:12:00.789 17:33:31 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:12:00.789 17:33:31 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:12:00.789 17:33:31 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:12:00.789 17:33:31 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:12:00.789 17:33:31 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:12:00.789 17:33:31 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:12:00.789 17:33:31 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:12:00.789 17:33:31 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:12:00.789 17:33:31 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:12:00.789 17:33:31 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:12:00.789 17:33:31 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:12:00.789 17:33:31 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:00.789 17:33:31 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:00.789 17:33:31 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:00.789 17:33:31 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:12:00.789 17:33:31 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:00.789 17:33:31 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:00.789 [2024-11-27 17:33:31.972349] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:12:00.789 [2024-11-27 17:33:31.972410] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:00.789 [2024-11-27 17:33:31.972448] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:12:00.789 [2024-11-27 17:33:31.972463] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:00.789 [2024-11-27 17:33:31.974976] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:00.789 [2024-11-27 17:33:31.975017] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:12:00.789 [2024-11-27 17:33:31.975107] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:12:00.789 [2024-11-27 17:33:31.975176] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:00.789 [2024-11-27 17:33:31.975325] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:12:00.789 spare 00:12:00.789 17:33:31 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:00.789 17:33:31 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:12:00.789 17:33:31 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:00.789 17:33:31 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:01.049 [2024-11-27 17:33:32.075247] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001580 00:12:01.050 [2024-11-27 17:33:32.075272] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:12:01.050 [2024-11-27 17:33:32.075569] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000cae960 00:12:01.050 [2024-11-27 17:33:32.075762] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001580 00:12:01.050 [2024-11-27 17:33:32.075788] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001580 00:12:01.050 [2024-11-27 17:33:32.075925] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:01.050 17:33:32 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:01.050 17:33:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:12:01.050 17:33:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:01.050 17:33:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:01.050 17:33:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:01.050 17:33:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:01.050 17:33:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:12:01.050 17:33:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:01.050 17:33:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:01.050 17:33:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:01.050 17:33:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:01.050 17:33:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:01.050 17:33:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:01.050 17:33:32 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:01.050 17:33:32 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:01.050 17:33:32 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:01.050 17:33:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:01.050 "name": "raid_bdev1", 00:12:01.050 "uuid": "1e7dc017-2c9f-4f39-a1ba-504c45cdbd67", 00:12:01.050 "strip_size_kb": 0, 00:12:01.050 "state": "online", 00:12:01.050 "raid_level": "raid1", 00:12:01.050 "superblock": true, 00:12:01.050 "num_base_bdevs": 2, 00:12:01.050 "num_base_bdevs_discovered": 2, 00:12:01.050 "num_base_bdevs_operational": 2, 00:12:01.050 "base_bdevs_list": [ 00:12:01.050 { 00:12:01.050 "name": "spare", 00:12:01.050 "uuid": "a4be28a5-7c99-5f92-a551-3de87fd43090", 00:12:01.050 "is_configured": true, 00:12:01.050 "data_offset": 2048, 00:12:01.050 "data_size": 63488 00:12:01.050 }, 00:12:01.050 { 00:12:01.050 "name": "BaseBdev2", 00:12:01.050 "uuid": "4fe7c978-6106-55d6-8587-c40656640823", 00:12:01.050 "is_configured": true, 00:12:01.050 "data_offset": 2048, 00:12:01.050 "data_size": 63488 00:12:01.050 } 00:12:01.050 ] 00:12:01.050 }' 00:12:01.050 17:33:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:01.050 17:33:32 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:01.620 17:33:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:01.620 17:33:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:01.620 17:33:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:01.620 17:33:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:01.620 17:33:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:01.620 17:33:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:01.620 17:33:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:01.620 17:33:32 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:01.620 17:33:32 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:01.620 17:33:32 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:01.620 17:33:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:01.620 "name": "raid_bdev1", 00:12:01.620 "uuid": "1e7dc017-2c9f-4f39-a1ba-504c45cdbd67", 00:12:01.620 "strip_size_kb": 0, 00:12:01.620 "state": "online", 00:12:01.620 "raid_level": "raid1", 00:12:01.620 "superblock": true, 00:12:01.620 "num_base_bdevs": 2, 00:12:01.620 "num_base_bdevs_discovered": 2, 00:12:01.620 "num_base_bdevs_operational": 2, 00:12:01.620 "base_bdevs_list": [ 00:12:01.620 { 00:12:01.620 "name": "spare", 00:12:01.620 "uuid": "a4be28a5-7c99-5f92-a551-3de87fd43090", 00:12:01.620 "is_configured": true, 00:12:01.620 "data_offset": 2048, 00:12:01.620 "data_size": 63488 00:12:01.620 }, 00:12:01.620 { 00:12:01.620 "name": "BaseBdev2", 00:12:01.620 "uuid": "4fe7c978-6106-55d6-8587-c40656640823", 00:12:01.620 "is_configured": true, 00:12:01.620 "data_offset": 2048, 00:12:01.620 "data_size": 63488 00:12:01.620 } 00:12:01.620 ] 00:12:01.620 }' 00:12:01.620 17:33:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:01.620 17:33:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:01.620 17:33:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:01.620 17:33:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:01.620 17:33:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:01.620 17:33:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:12:01.620 17:33:32 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:01.620 17:33:32 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:01.620 17:33:32 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:01.620 17:33:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:12:01.620 17:33:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:12:01.620 17:33:32 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:01.620 17:33:32 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:01.620 [2024-11-27 17:33:32.703145] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:01.620 17:33:32 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:01.620 17:33:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:12:01.620 17:33:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:01.620 17:33:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:01.620 17:33:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:01.620 17:33:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:01.620 17:33:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:12:01.620 17:33:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:01.620 17:33:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:01.620 17:33:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:01.620 17:33:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:01.620 17:33:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:01.620 17:33:32 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:01.620 17:33:32 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:01.620 17:33:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:01.620 17:33:32 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:01.620 17:33:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:01.620 "name": "raid_bdev1", 00:12:01.620 "uuid": "1e7dc017-2c9f-4f39-a1ba-504c45cdbd67", 00:12:01.620 "strip_size_kb": 0, 00:12:01.620 "state": "online", 00:12:01.620 "raid_level": "raid1", 00:12:01.620 "superblock": true, 00:12:01.620 "num_base_bdevs": 2, 00:12:01.620 "num_base_bdevs_discovered": 1, 00:12:01.620 "num_base_bdevs_operational": 1, 00:12:01.621 "base_bdevs_list": [ 00:12:01.621 { 00:12:01.621 "name": null, 00:12:01.621 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:01.621 "is_configured": false, 00:12:01.621 "data_offset": 0, 00:12:01.621 "data_size": 63488 00:12:01.621 }, 00:12:01.621 { 00:12:01.621 "name": "BaseBdev2", 00:12:01.621 "uuid": "4fe7c978-6106-55d6-8587-c40656640823", 00:12:01.621 "is_configured": true, 00:12:01.621 "data_offset": 2048, 00:12:01.621 "data_size": 63488 00:12:01.621 } 00:12:01.621 ] 00:12:01.621 }' 00:12:01.621 17:33:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:01.621 17:33:32 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:02.191 17:33:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:12:02.191 17:33:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:02.191 17:33:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:02.191 [2024-11-27 17:33:33.158437] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:02.191 [2024-11-27 17:33:33.158659] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:12:02.191 [2024-11-27 17:33:33.158679] bdev_raid.c:3748:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:12:02.191 [2024-11-27 17:33:33.158728] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:02.191 [2024-11-27 17:33:33.165791] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000caea30 00:12:02.191 17:33:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:02.191 17:33:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@757 -- # sleep 1 00:12:02.191 [2024-11-27 17:33:33.167984] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:12:03.131 17:33:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:03.131 17:33:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:03.131 17:33:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:03.131 17:33:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:03.131 17:33:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:03.131 17:33:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:03.131 17:33:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:03.131 17:33:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:03.131 17:33:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:03.131 17:33:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:03.131 17:33:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:03.131 "name": "raid_bdev1", 00:12:03.131 "uuid": "1e7dc017-2c9f-4f39-a1ba-504c45cdbd67", 00:12:03.131 "strip_size_kb": 0, 00:12:03.131 "state": "online", 00:12:03.132 "raid_level": "raid1", 00:12:03.132 "superblock": true, 00:12:03.132 "num_base_bdevs": 2, 00:12:03.132 "num_base_bdevs_discovered": 2, 00:12:03.132 "num_base_bdevs_operational": 2, 00:12:03.132 "process": { 00:12:03.132 "type": "rebuild", 00:12:03.132 "target": "spare", 00:12:03.132 "progress": { 00:12:03.132 "blocks": 20480, 00:12:03.132 "percent": 32 00:12:03.132 } 00:12:03.132 }, 00:12:03.132 "base_bdevs_list": [ 00:12:03.132 { 00:12:03.132 "name": "spare", 00:12:03.132 "uuid": "a4be28a5-7c99-5f92-a551-3de87fd43090", 00:12:03.132 "is_configured": true, 00:12:03.132 "data_offset": 2048, 00:12:03.132 "data_size": 63488 00:12:03.132 }, 00:12:03.132 { 00:12:03.132 "name": "BaseBdev2", 00:12:03.132 "uuid": "4fe7c978-6106-55d6-8587-c40656640823", 00:12:03.132 "is_configured": true, 00:12:03.132 "data_offset": 2048, 00:12:03.132 "data_size": 63488 00:12:03.132 } 00:12:03.132 ] 00:12:03.132 }' 00:12:03.132 17:33:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:03.132 17:33:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:03.132 17:33:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:03.392 17:33:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:03.392 17:33:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:12:03.392 17:33:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:03.392 17:33:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:03.392 [2024-11-27 17:33:34.331815] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:03.392 [2024-11-27 17:33:34.375295] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:12:03.392 [2024-11-27 17:33:34.375354] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:03.392 [2024-11-27 17:33:34.375372] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:03.392 [2024-11-27 17:33:34.375380] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:12:03.392 17:33:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:03.392 17:33:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:12:03.392 17:33:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:03.392 17:33:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:03.392 17:33:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:03.392 17:33:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:03.392 17:33:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:12:03.392 17:33:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:03.392 17:33:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:03.392 17:33:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:03.392 17:33:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:03.392 17:33:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:03.392 17:33:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:03.392 17:33:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:03.392 17:33:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:03.392 17:33:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:03.392 17:33:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:03.392 "name": "raid_bdev1", 00:12:03.392 "uuid": "1e7dc017-2c9f-4f39-a1ba-504c45cdbd67", 00:12:03.392 "strip_size_kb": 0, 00:12:03.392 "state": "online", 00:12:03.392 "raid_level": "raid1", 00:12:03.392 "superblock": true, 00:12:03.392 "num_base_bdevs": 2, 00:12:03.392 "num_base_bdevs_discovered": 1, 00:12:03.393 "num_base_bdevs_operational": 1, 00:12:03.393 "base_bdevs_list": [ 00:12:03.393 { 00:12:03.393 "name": null, 00:12:03.393 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:03.393 "is_configured": false, 00:12:03.393 "data_offset": 0, 00:12:03.393 "data_size": 63488 00:12:03.393 }, 00:12:03.393 { 00:12:03.393 "name": "BaseBdev2", 00:12:03.393 "uuid": "4fe7c978-6106-55d6-8587-c40656640823", 00:12:03.393 "is_configured": true, 00:12:03.393 "data_offset": 2048, 00:12:03.393 "data_size": 63488 00:12:03.393 } 00:12:03.393 ] 00:12:03.393 }' 00:12:03.393 17:33:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:03.393 17:33:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:03.652 17:33:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:12:03.652 17:33:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:03.652 17:33:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:03.652 [2024-11-27 17:33:34.825665] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:12:03.652 [2024-11-27 17:33:34.825732] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:03.652 [2024-11-27 17:33:34.825779] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009f80 00:12:03.652 [2024-11-27 17:33:34.825793] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:03.652 [2024-11-27 17:33:34.826327] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:03.652 [2024-11-27 17:33:34.826352] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:12:03.652 [2024-11-27 17:33:34.826448] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:12:03.652 [2024-11-27 17:33:34.826466] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:12:03.652 [2024-11-27 17:33:34.826486] bdev_raid.c:3748:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:12:03.652 [2024-11-27 17:33:34.826509] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:03.652 [2024-11-27 17:33:34.833238] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000caeb00 00:12:03.652 spare 00:12:03.652 17:33:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:03.652 17:33:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@764 -- # sleep 1 00:12:03.652 [2024-11-27 17:33:34.835436] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:12:05.034 17:33:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:05.034 17:33:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:05.034 17:33:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:05.034 17:33:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:05.034 17:33:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:05.034 17:33:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:05.034 17:33:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:05.034 17:33:35 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:05.034 17:33:35 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:05.034 17:33:35 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:05.034 17:33:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:05.034 "name": "raid_bdev1", 00:12:05.034 "uuid": "1e7dc017-2c9f-4f39-a1ba-504c45cdbd67", 00:12:05.034 "strip_size_kb": 0, 00:12:05.034 "state": "online", 00:12:05.034 "raid_level": "raid1", 00:12:05.034 "superblock": true, 00:12:05.034 "num_base_bdevs": 2, 00:12:05.034 "num_base_bdevs_discovered": 2, 00:12:05.034 "num_base_bdevs_operational": 2, 00:12:05.034 "process": { 00:12:05.034 "type": "rebuild", 00:12:05.034 "target": "spare", 00:12:05.034 "progress": { 00:12:05.034 "blocks": 20480, 00:12:05.034 "percent": 32 00:12:05.034 } 00:12:05.034 }, 00:12:05.034 "base_bdevs_list": [ 00:12:05.034 { 00:12:05.034 "name": "spare", 00:12:05.034 "uuid": "a4be28a5-7c99-5f92-a551-3de87fd43090", 00:12:05.034 "is_configured": true, 00:12:05.034 "data_offset": 2048, 00:12:05.034 "data_size": 63488 00:12:05.034 }, 00:12:05.034 { 00:12:05.034 "name": "BaseBdev2", 00:12:05.034 "uuid": "4fe7c978-6106-55d6-8587-c40656640823", 00:12:05.034 "is_configured": true, 00:12:05.034 "data_offset": 2048, 00:12:05.034 "data_size": 63488 00:12:05.034 } 00:12:05.034 ] 00:12:05.034 }' 00:12:05.034 17:33:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:05.034 17:33:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:05.034 17:33:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:05.034 17:33:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:05.034 17:33:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:12:05.034 17:33:35 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:05.034 17:33:35 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:05.034 [2024-11-27 17:33:35.999349] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:05.034 [2024-11-27 17:33:36.043019] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:12:05.034 [2024-11-27 17:33:36.043096] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:05.034 [2024-11-27 17:33:36.043111] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:05.034 [2024-11-27 17:33:36.043121] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:12:05.034 17:33:36 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:05.034 17:33:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:12:05.034 17:33:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:05.034 17:33:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:05.034 17:33:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:05.034 17:33:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:05.034 17:33:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:12:05.034 17:33:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:05.034 17:33:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:05.034 17:33:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:05.034 17:33:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:05.034 17:33:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:05.034 17:33:36 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:05.034 17:33:36 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:05.034 17:33:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:05.034 17:33:36 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:05.034 17:33:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:05.034 "name": "raid_bdev1", 00:12:05.034 "uuid": "1e7dc017-2c9f-4f39-a1ba-504c45cdbd67", 00:12:05.034 "strip_size_kb": 0, 00:12:05.034 "state": "online", 00:12:05.034 "raid_level": "raid1", 00:12:05.034 "superblock": true, 00:12:05.034 "num_base_bdevs": 2, 00:12:05.034 "num_base_bdevs_discovered": 1, 00:12:05.034 "num_base_bdevs_operational": 1, 00:12:05.035 "base_bdevs_list": [ 00:12:05.035 { 00:12:05.035 "name": null, 00:12:05.035 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:05.035 "is_configured": false, 00:12:05.035 "data_offset": 0, 00:12:05.035 "data_size": 63488 00:12:05.035 }, 00:12:05.035 { 00:12:05.035 "name": "BaseBdev2", 00:12:05.035 "uuid": "4fe7c978-6106-55d6-8587-c40656640823", 00:12:05.035 "is_configured": true, 00:12:05.035 "data_offset": 2048, 00:12:05.035 "data_size": 63488 00:12:05.035 } 00:12:05.035 ] 00:12:05.035 }' 00:12:05.035 17:33:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:05.035 17:33:36 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:05.605 17:33:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:05.605 17:33:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:05.605 17:33:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:05.605 17:33:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:05.605 17:33:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:05.605 17:33:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:05.605 17:33:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:05.605 17:33:36 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:05.605 17:33:36 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:05.605 17:33:36 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:05.605 17:33:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:05.605 "name": "raid_bdev1", 00:12:05.605 "uuid": "1e7dc017-2c9f-4f39-a1ba-504c45cdbd67", 00:12:05.605 "strip_size_kb": 0, 00:12:05.605 "state": "online", 00:12:05.605 "raid_level": "raid1", 00:12:05.605 "superblock": true, 00:12:05.605 "num_base_bdevs": 2, 00:12:05.605 "num_base_bdevs_discovered": 1, 00:12:05.605 "num_base_bdevs_operational": 1, 00:12:05.605 "base_bdevs_list": [ 00:12:05.605 { 00:12:05.605 "name": null, 00:12:05.605 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:05.605 "is_configured": false, 00:12:05.605 "data_offset": 0, 00:12:05.605 "data_size": 63488 00:12:05.605 }, 00:12:05.605 { 00:12:05.605 "name": "BaseBdev2", 00:12:05.605 "uuid": "4fe7c978-6106-55d6-8587-c40656640823", 00:12:05.605 "is_configured": true, 00:12:05.605 "data_offset": 2048, 00:12:05.605 "data_size": 63488 00:12:05.605 } 00:12:05.605 ] 00:12:05.605 }' 00:12:05.605 17:33:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:05.605 17:33:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:05.605 17:33:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:05.605 17:33:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:05.605 17:33:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:12:05.605 17:33:36 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:05.605 17:33:36 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:05.605 17:33:36 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:05.605 17:33:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:12:05.605 17:33:36 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:05.605 17:33:36 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:05.605 [2024-11-27 17:33:36.637532] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:12:05.605 [2024-11-27 17:33:36.637598] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:05.605 [2024-11-27 17:33:36.637621] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a580 00:12:05.605 [2024-11-27 17:33:36.637632] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:05.605 [2024-11-27 17:33:36.638119] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:05.605 [2024-11-27 17:33:36.638161] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:12:05.605 [2024-11-27 17:33:36.638244] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:12:05.605 [2024-11-27 17:33:36.638276] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:12:05.605 [2024-11-27 17:33:36.638285] bdev_raid.c:3709:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:12:05.605 [2024-11-27 17:33:36.638300] bdev_raid.c:3884:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:12:05.605 BaseBdev1 00:12:05.605 17:33:36 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:05.605 17:33:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@775 -- # sleep 1 00:12:06.545 17:33:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:12:06.545 17:33:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:06.545 17:33:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:06.545 17:33:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:06.545 17:33:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:06.545 17:33:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:12:06.545 17:33:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:06.545 17:33:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:06.545 17:33:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:06.545 17:33:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:06.545 17:33:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:06.545 17:33:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:06.545 17:33:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:06.545 17:33:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:06.545 17:33:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:06.545 17:33:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:06.545 "name": "raid_bdev1", 00:12:06.545 "uuid": "1e7dc017-2c9f-4f39-a1ba-504c45cdbd67", 00:12:06.545 "strip_size_kb": 0, 00:12:06.545 "state": "online", 00:12:06.545 "raid_level": "raid1", 00:12:06.545 "superblock": true, 00:12:06.545 "num_base_bdevs": 2, 00:12:06.545 "num_base_bdevs_discovered": 1, 00:12:06.545 "num_base_bdevs_operational": 1, 00:12:06.545 "base_bdevs_list": [ 00:12:06.545 { 00:12:06.545 "name": null, 00:12:06.545 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:06.545 "is_configured": false, 00:12:06.545 "data_offset": 0, 00:12:06.545 "data_size": 63488 00:12:06.545 }, 00:12:06.545 { 00:12:06.545 "name": "BaseBdev2", 00:12:06.545 "uuid": "4fe7c978-6106-55d6-8587-c40656640823", 00:12:06.546 "is_configured": true, 00:12:06.546 "data_offset": 2048, 00:12:06.546 "data_size": 63488 00:12:06.546 } 00:12:06.546 ] 00:12:06.546 }' 00:12:06.546 17:33:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:06.546 17:33:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:07.116 17:33:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:07.116 17:33:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:07.116 17:33:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:07.116 17:33:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:07.116 17:33:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:07.116 17:33:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:07.116 17:33:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:07.116 17:33:38 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:07.116 17:33:38 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:07.116 17:33:38 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:07.116 17:33:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:07.116 "name": "raid_bdev1", 00:12:07.116 "uuid": "1e7dc017-2c9f-4f39-a1ba-504c45cdbd67", 00:12:07.116 "strip_size_kb": 0, 00:12:07.116 "state": "online", 00:12:07.116 "raid_level": "raid1", 00:12:07.116 "superblock": true, 00:12:07.116 "num_base_bdevs": 2, 00:12:07.116 "num_base_bdevs_discovered": 1, 00:12:07.116 "num_base_bdevs_operational": 1, 00:12:07.116 "base_bdevs_list": [ 00:12:07.116 { 00:12:07.116 "name": null, 00:12:07.116 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:07.116 "is_configured": false, 00:12:07.116 "data_offset": 0, 00:12:07.116 "data_size": 63488 00:12:07.116 }, 00:12:07.116 { 00:12:07.116 "name": "BaseBdev2", 00:12:07.116 "uuid": "4fe7c978-6106-55d6-8587-c40656640823", 00:12:07.116 "is_configured": true, 00:12:07.116 "data_offset": 2048, 00:12:07.116 "data_size": 63488 00:12:07.116 } 00:12:07.116 ] 00:12:07.116 }' 00:12:07.116 17:33:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:07.116 17:33:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:07.116 17:33:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:07.116 17:33:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:07.116 17:33:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:12:07.116 17:33:38 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@650 -- # local es=0 00:12:07.116 17:33:38 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:12:07.116 17:33:38 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:12:07.116 17:33:38 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:12:07.116 17:33:38 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:12:07.116 17:33:38 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:12:07.116 17:33:38 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:12:07.116 17:33:38 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:07.116 17:33:38 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:07.116 [2024-11-27 17:33:38.238877] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:12:07.116 [2024-11-27 17:33:38.239078] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:12:07.116 [2024-11-27 17:33:38.239100] bdev_raid.c:3709:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:12:07.116 request: 00:12:07.116 { 00:12:07.116 "base_bdev": "BaseBdev1", 00:12:07.116 "raid_bdev": "raid_bdev1", 00:12:07.116 "method": "bdev_raid_add_base_bdev", 00:12:07.116 "req_id": 1 00:12:07.116 } 00:12:07.116 Got JSON-RPC error response 00:12:07.116 response: 00:12:07.116 { 00:12:07.116 "code": -22, 00:12:07.116 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:12:07.116 } 00:12:07.116 17:33:38 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:12:07.116 17:33:38 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@653 -- # es=1 00:12:07.116 17:33:38 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:12:07.116 17:33:38 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:12:07.116 17:33:38 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:12:07.116 17:33:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@779 -- # sleep 1 00:12:08.496 17:33:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:12:08.496 17:33:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:08.496 17:33:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:08.496 17:33:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:08.496 17:33:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:08.496 17:33:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:12:08.496 17:33:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:08.496 17:33:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:08.496 17:33:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:08.496 17:33:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:08.496 17:33:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:08.496 17:33:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:08.496 17:33:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:08.496 17:33:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:08.496 17:33:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:08.496 17:33:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:08.496 "name": "raid_bdev1", 00:12:08.496 "uuid": "1e7dc017-2c9f-4f39-a1ba-504c45cdbd67", 00:12:08.496 "strip_size_kb": 0, 00:12:08.496 "state": "online", 00:12:08.496 "raid_level": "raid1", 00:12:08.496 "superblock": true, 00:12:08.496 "num_base_bdevs": 2, 00:12:08.496 "num_base_bdevs_discovered": 1, 00:12:08.496 "num_base_bdevs_operational": 1, 00:12:08.496 "base_bdevs_list": [ 00:12:08.496 { 00:12:08.496 "name": null, 00:12:08.496 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:08.496 "is_configured": false, 00:12:08.496 "data_offset": 0, 00:12:08.496 "data_size": 63488 00:12:08.496 }, 00:12:08.496 { 00:12:08.496 "name": "BaseBdev2", 00:12:08.496 "uuid": "4fe7c978-6106-55d6-8587-c40656640823", 00:12:08.496 "is_configured": true, 00:12:08.496 "data_offset": 2048, 00:12:08.496 "data_size": 63488 00:12:08.496 } 00:12:08.496 ] 00:12:08.496 }' 00:12:08.496 17:33:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:08.496 17:33:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:08.496 17:33:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:08.496 17:33:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:08.496 17:33:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:08.497 17:33:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:08.497 17:33:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:08.757 17:33:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:08.757 17:33:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:08.757 17:33:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:08.757 17:33:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:08.757 17:33:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:08.757 17:33:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:08.757 "name": "raid_bdev1", 00:12:08.757 "uuid": "1e7dc017-2c9f-4f39-a1ba-504c45cdbd67", 00:12:08.757 "strip_size_kb": 0, 00:12:08.757 "state": "online", 00:12:08.757 "raid_level": "raid1", 00:12:08.757 "superblock": true, 00:12:08.757 "num_base_bdevs": 2, 00:12:08.757 "num_base_bdevs_discovered": 1, 00:12:08.757 "num_base_bdevs_operational": 1, 00:12:08.757 "base_bdevs_list": [ 00:12:08.757 { 00:12:08.757 "name": null, 00:12:08.757 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:08.757 "is_configured": false, 00:12:08.757 "data_offset": 0, 00:12:08.757 "data_size": 63488 00:12:08.757 }, 00:12:08.757 { 00:12:08.757 "name": "BaseBdev2", 00:12:08.757 "uuid": "4fe7c978-6106-55d6-8587-c40656640823", 00:12:08.757 "is_configured": true, 00:12:08.757 "data_offset": 2048, 00:12:08.757 "data_size": 63488 00:12:08.757 } 00:12:08.757 ] 00:12:08.757 }' 00:12:08.757 17:33:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:08.757 17:33:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:08.757 17:33:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:08.757 17:33:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:08.757 17:33:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@784 -- # killprocess 86285 00:12:08.757 17:33:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@950 -- # '[' -z 86285 ']' 00:12:08.757 17:33:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@954 -- # kill -0 86285 00:12:08.757 17:33:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@955 -- # uname 00:12:08.757 17:33:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:12:08.757 17:33:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 86285 00:12:08.757 17:33:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:12:08.757 17:33:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:12:08.757 killing process with pid 86285 00:12:08.757 17:33:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@968 -- # echo 'killing process with pid 86285' 00:12:08.757 17:33:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@969 -- # kill 86285 00:12:08.757 Received shutdown signal, test time was about 60.000000 seconds 00:12:08.757 00:12:08.757 Latency(us) 00:12:08.757 [2024-11-27T17:33:39.949Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:12:08.757 [2024-11-27T17:33:39.949Z] =================================================================================================================== 00:12:08.757 [2024-11-27T17:33:39.949Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:12:08.757 [2024-11-27 17:33:39.862457] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:12:08.757 [2024-11-27 17:33:39.862614] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:12:08.757 17:33:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@974 -- # wait 86285 00:12:08.757 [2024-11-27 17:33:39.862680] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:12:08.757 [2024-11-27 17:33:39.862690] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state offline 00:12:08.757 [2024-11-27 17:33:39.920537] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:12:09.327 17:33:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@786 -- # return 0 00:12:09.327 00:12:09.327 real 0m21.705s 00:12:09.327 user 0m26.480s 00:12:09.327 sys 0m3.758s 00:12:09.327 17:33:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@1126 -- # xtrace_disable 00:12:09.327 17:33:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:09.327 ************************************ 00:12:09.327 END TEST raid_rebuild_test_sb 00:12:09.327 ************************************ 00:12:09.327 17:33:40 bdev_raid -- bdev/bdev_raid.sh@980 -- # run_test raid_rebuild_test_io raid_rebuild_test raid1 2 false true true 00:12:09.327 17:33:40 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 7 -le 1 ']' 00:12:09.327 17:33:40 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:12:09.327 17:33:40 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:12:09.327 ************************************ 00:12:09.327 START TEST raid_rebuild_test_io 00:12:09.327 ************************************ 00:12:09.327 17:33:40 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@1125 -- # raid_rebuild_test raid1 2 false true true 00:12:09.327 17:33:40 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:12:09.327 17:33:40 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=2 00:12:09.327 17:33:40 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@571 -- # local superblock=false 00:12:09.327 17:33:40 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@572 -- # local background_io=true 00:12:09.327 17:33:40 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@573 -- # local verify=true 00:12:09.327 17:33:40 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:12:09.327 17:33:40 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:09.327 17:33:40 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:12:09.327 17:33:40 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:09.327 17:33:40 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:09.327 17:33:40 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:12:09.327 17:33:40 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:09.327 17:33:40 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:09.327 17:33:40 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:12:09.327 17:33:40 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:12:09.327 17:33:40 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:12:09.327 17:33:40 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@576 -- # local strip_size 00:12:09.327 17:33:40 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@577 -- # local create_arg 00:12:09.327 17:33:40 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:12:09.327 17:33:40 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@579 -- # local data_offset 00:12:09.327 17:33:40 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:12:09.327 17:33:40 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:12:09.327 17:33:40 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@592 -- # '[' false = true ']' 00:12:09.327 17:33:40 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@597 -- # raid_pid=87003 00:12:09.327 17:33:40 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@598 -- # waitforlisten 87003 00:12:09.327 17:33:40 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:12:09.327 17:33:40 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@831 -- # '[' -z 87003 ']' 00:12:09.327 17:33:40 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:12:09.327 17:33:40 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@836 -- # local max_retries=100 00:12:09.327 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:12:09.327 17:33:40 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:12:09.327 17:33:40 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@840 -- # xtrace_disable 00:12:09.327 17:33:40 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:09.327 I/O size of 3145728 is greater than zero copy threshold (65536). 00:12:09.327 Zero copy mechanism will not be used. 00:12:09.327 [2024-11-27 17:33:40.456774] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:12:09.328 [2024-11-27 17:33:40.456884] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid87003 ] 00:12:09.588 [2024-11-27 17:33:40.599647] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:12:09.588 [2024-11-27 17:33:40.666816] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:12:09.588 [2024-11-27 17:33:40.743136] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:12:09.588 [2024-11-27 17:33:40.743183] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:12:10.158 17:33:41 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:12:10.158 17:33:41 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@864 -- # return 0 00:12:10.158 17:33:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:10.158 17:33:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:12:10.158 17:33:41 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:10.158 17:33:41 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:10.158 BaseBdev1_malloc 00:12:10.158 17:33:41 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:10.158 17:33:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:12:10.158 17:33:41 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:10.158 17:33:41 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:10.158 [2024-11-27 17:33:41.301083] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:12:10.158 [2024-11-27 17:33:41.301154] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:10.158 [2024-11-27 17:33:41.301181] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:12:10.158 [2024-11-27 17:33:41.301196] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:10.158 [2024-11-27 17:33:41.303659] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:10.158 [2024-11-27 17:33:41.303694] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:12:10.158 BaseBdev1 00:12:10.158 17:33:41 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:10.158 17:33:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:10.158 17:33:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:12:10.158 17:33:41 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:10.158 17:33:41 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:10.158 BaseBdev2_malloc 00:12:10.158 17:33:41 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:10.158 17:33:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:12:10.158 17:33:41 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:10.158 17:33:41 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:10.418 [2024-11-27 17:33:41.351113] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:12:10.418 [2024-11-27 17:33:41.351215] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:10.418 [2024-11-27 17:33:41.351258] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:12:10.418 [2024-11-27 17:33:41.351277] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:10.418 [2024-11-27 17:33:41.355838] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:10.418 [2024-11-27 17:33:41.355896] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:12:10.418 BaseBdev2 00:12:10.418 17:33:41 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:10.418 17:33:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:12:10.418 17:33:41 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:10.418 17:33:41 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:10.418 spare_malloc 00:12:10.418 17:33:41 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:10.418 17:33:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:12:10.418 17:33:41 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:10.418 17:33:41 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:10.418 spare_delay 00:12:10.418 17:33:41 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:10.418 17:33:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:12:10.418 17:33:41 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:10.418 17:33:41 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:10.418 [2024-11-27 17:33:41.399839] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:12:10.418 [2024-11-27 17:33:41.399904] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:10.418 [2024-11-27 17:33:41.399936] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:12:10.418 [2024-11-27 17:33:41.399945] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:10.418 [2024-11-27 17:33:41.402369] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:10.418 [2024-11-27 17:33:41.402397] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:12:10.418 spare 00:12:10.418 17:33:41 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:10.418 17:33:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 00:12:10.418 17:33:41 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:10.418 17:33:41 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:10.418 [2024-11-27 17:33:41.411875] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:12:10.418 [2024-11-27 17:33:41.413993] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:12:10.419 [2024-11-27 17:33:41.414084] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:12:10.419 [2024-11-27 17:33:41.414096] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:12:10.419 [2024-11-27 17:33:41.414413] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:12:10.419 [2024-11-27 17:33:41.414540] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:12:10.419 [2024-11-27 17:33:41.414561] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:12:10.419 [2024-11-27 17:33:41.414689] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:10.419 17:33:41 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:10.419 17:33:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:12:10.419 17:33:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:10.419 17:33:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:10.419 17:33:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:10.419 17:33:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:10.419 17:33:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:12:10.419 17:33:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:10.419 17:33:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:10.419 17:33:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:10.419 17:33:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:10.419 17:33:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:10.419 17:33:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:10.419 17:33:41 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:10.419 17:33:41 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:10.419 17:33:41 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:10.419 17:33:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:10.419 "name": "raid_bdev1", 00:12:10.419 "uuid": "87009852-bfe6-4e82-8dcd-67fda32884ed", 00:12:10.419 "strip_size_kb": 0, 00:12:10.419 "state": "online", 00:12:10.419 "raid_level": "raid1", 00:12:10.419 "superblock": false, 00:12:10.419 "num_base_bdevs": 2, 00:12:10.419 "num_base_bdevs_discovered": 2, 00:12:10.419 "num_base_bdevs_operational": 2, 00:12:10.419 "base_bdevs_list": [ 00:12:10.419 { 00:12:10.419 "name": "BaseBdev1", 00:12:10.419 "uuid": "70e6034a-578e-5cd2-9b6b-69e993444c88", 00:12:10.419 "is_configured": true, 00:12:10.419 "data_offset": 0, 00:12:10.419 "data_size": 65536 00:12:10.419 }, 00:12:10.419 { 00:12:10.419 "name": "BaseBdev2", 00:12:10.419 "uuid": "4c36ab84-9c35-583f-a6c3-6244915797fd", 00:12:10.419 "is_configured": true, 00:12:10.419 "data_offset": 0, 00:12:10.419 "data_size": 65536 00:12:10.419 } 00:12:10.419 ] 00:12:10.419 }' 00:12:10.419 17:33:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:10.419 17:33:41 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:10.678 17:33:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:12:10.678 17:33:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:12:10.678 17:33:41 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:10.678 17:33:41 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:10.678 [2024-11-27 17:33:41.859344] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:12:10.939 17:33:41 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:10.939 17:33:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=65536 00:12:10.939 17:33:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:10.939 17:33:41 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:10.939 17:33:41 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:10.939 17:33:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:12:10.939 17:33:41 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:10.939 17:33:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@619 -- # data_offset=0 00:12:10.939 17:33:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@621 -- # '[' true = true ']' 00:12:10.939 17:33:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:12:10.939 17:33:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@623 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:12:10.939 17:33:41 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:10.939 17:33:41 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:10.939 [2024-11-27 17:33:41.958893] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:12:10.939 17:33:41 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:10.939 17:33:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:12:10.939 17:33:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:10.939 17:33:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:10.939 17:33:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:10.939 17:33:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:10.939 17:33:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:12:10.939 17:33:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:10.939 17:33:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:10.939 17:33:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:10.939 17:33:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:10.939 17:33:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:10.939 17:33:41 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:10.939 17:33:41 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:10.939 17:33:41 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:10.939 17:33:41 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:10.939 17:33:42 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:10.939 "name": "raid_bdev1", 00:12:10.939 "uuid": "87009852-bfe6-4e82-8dcd-67fda32884ed", 00:12:10.939 "strip_size_kb": 0, 00:12:10.939 "state": "online", 00:12:10.939 "raid_level": "raid1", 00:12:10.939 "superblock": false, 00:12:10.939 "num_base_bdevs": 2, 00:12:10.939 "num_base_bdevs_discovered": 1, 00:12:10.939 "num_base_bdevs_operational": 1, 00:12:10.939 "base_bdevs_list": [ 00:12:10.939 { 00:12:10.939 "name": null, 00:12:10.939 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:10.939 "is_configured": false, 00:12:10.939 "data_offset": 0, 00:12:10.939 "data_size": 65536 00:12:10.939 }, 00:12:10.939 { 00:12:10.939 "name": "BaseBdev2", 00:12:10.939 "uuid": "4c36ab84-9c35-583f-a6c3-6244915797fd", 00:12:10.939 "is_configured": true, 00:12:10.939 "data_offset": 0, 00:12:10.939 "data_size": 65536 00:12:10.939 } 00:12:10.939 ] 00:12:10.939 }' 00:12:10.939 17:33:42 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:10.939 17:33:42 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:10.939 [2024-11-27 17:33:42.038315] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000026d0 00:12:10.939 I/O size of 3145728 is greater than zero copy threshold (65536). 00:12:10.939 Zero copy mechanism will not be used. 00:12:10.939 Running I/O for 60 seconds... 00:12:11.199 17:33:42 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:12:11.199 17:33:42 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:11.199 17:33:42 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:11.199 [2024-11-27 17:33:42.338034] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:11.199 17:33:42 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:11.199 17:33:42 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@647 -- # sleep 1 00:12:11.459 [2024-11-27 17:33:42.394135] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000027a0 00:12:11.459 [2024-11-27 17:33:42.396449] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:12:11.459 [2024-11-27 17:33:42.504776] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:12:11.459 [2024-11-27 17:33:42.505258] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:12:11.718 [2024-11-27 17:33:42.712387] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:12:11.718 [2024-11-27 17:33:42.712692] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:12:11.979 167.00 IOPS, 501.00 MiB/s [2024-11-27T17:33:43.171Z] [2024-11-27 17:33:43.049288] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:12:11.979 [2024-11-27 17:33:43.157881] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:12:11.979 [2024-11-27 17:33:43.158283] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:12:12.239 17:33:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:12.239 17:33:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:12.239 17:33:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:12.239 17:33:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:12.239 17:33:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:12.239 17:33:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:12.239 17:33:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:12.239 17:33:43 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:12.239 17:33:43 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:12.239 17:33:43 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:12.500 17:33:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:12.500 "name": "raid_bdev1", 00:12:12.500 "uuid": "87009852-bfe6-4e82-8dcd-67fda32884ed", 00:12:12.500 "strip_size_kb": 0, 00:12:12.500 "state": "online", 00:12:12.500 "raid_level": "raid1", 00:12:12.500 "superblock": false, 00:12:12.500 "num_base_bdevs": 2, 00:12:12.500 "num_base_bdevs_discovered": 2, 00:12:12.500 "num_base_bdevs_operational": 2, 00:12:12.500 "process": { 00:12:12.500 "type": "rebuild", 00:12:12.500 "target": "spare", 00:12:12.500 "progress": { 00:12:12.500 "blocks": 12288, 00:12:12.500 "percent": 18 00:12:12.500 } 00:12:12.500 }, 00:12:12.500 "base_bdevs_list": [ 00:12:12.500 { 00:12:12.500 "name": "spare", 00:12:12.500 "uuid": "ab1dad7c-8bf0-5526-bf88-45f884961d05", 00:12:12.500 "is_configured": true, 00:12:12.500 "data_offset": 0, 00:12:12.500 "data_size": 65536 00:12:12.500 }, 00:12:12.500 { 00:12:12.500 "name": "BaseBdev2", 00:12:12.500 "uuid": "4c36ab84-9c35-583f-a6c3-6244915797fd", 00:12:12.500 "is_configured": true, 00:12:12.500 "data_offset": 0, 00:12:12.500 "data_size": 65536 00:12:12.500 } 00:12:12.500 ] 00:12:12.500 }' 00:12:12.500 17:33:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:12.500 17:33:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:12.500 17:33:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:12.500 17:33:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:12.500 17:33:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:12:12.500 17:33:43 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:12.500 17:33:43 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:12.500 [2024-11-27 17:33:43.541048] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:12.500 [2024-11-27 17:33:43.583535] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:12:12.500 [2024-11-27 17:33:43.624928] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:12:12.500 [2024-11-27 17:33:43.626912] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:12.500 [2024-11-27 17:33:43.626954] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:12.500 [2024-11-27 17:33:43.626964] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:12:12.500 [2024-11-27 17:33:43.647525] bdev_raid.c:1970:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 0 raid_ch: 0x60d0000026d0 00:12:12.500 17:33:43 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:12.500 17:33:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:12:12.500 17:33:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:12.500 17:33:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:12.500 17:33:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:12.500 17:33:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:12.500 17:33:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:12:12.500 17:33:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:12.500 17:33:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:12.500 17:33:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:12.500 17:33:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:12.500 17:33:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:12.500 17:33:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:12.500 17:33:43 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:12.500 17:33:43 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:12.760 17:33:43 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:12.760 17:33:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:12.760 "name": "raid_bdev1", 00:12:12.760 "uuid": "87009852-bfe6-4e82-8dcd-67fda32884ed", 00:12:12.760 "strip_size_kb": 0, 00:12:12.760 "state": "online", 00:12:12.760 "raid_level": "raid1", 00:12:12.760 "superblock": false, 00:12:12.760 "num_base_bdevs": 2, 00:12:12.760 "num_base_bdevs_discovered": 1, 00:12:12.760 "num_base_bdevs_operational": 1, 00:12:12.760 "base_bdevs_list": [ 00:12:12.760 { 00:12:12.760 "name": null, 00:12:12.760 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:12.760 "is_configured": false, 00:12:12.760 "data_offset": 0, 00:12:12.760 "data_size": 65536 00:12:12.760 }, 00:12:12.760 { 00:12:12.760 "name": "BaseBdev2", 00:12:12.760 "uuid": "4c36ab84-9c35-583f-a6c3-6244915797fd", 00:12:12.760 "is_configured": true, 00:12:12.760 "data_offset": 0, 00:12:12.760 "data_size": 65536 00:12:12.760 } 00:12:12.760 ] 00:12:12.760 }' 00:12:12.760 17:33:43 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:12.760 17:33:43 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:13.020 162.50 IOPS, 487.50 MiB/s [2024-11-27T17:33:44.212Z] 17:33:44 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:13.020 17:33:44 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:13.020 17:33:44 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:13.020 17:33:44 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:13.020 17:33:44 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:13.020 17:33:44 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:13.020 17:33:44 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:13.020 17:33:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:13.020 17:33:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:13.020 17:33:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:13.020 17:33:44 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:13.020 "name": "raid_bdev1", 00:12:13.020 "uuid": "87009852-bfe6-4e82-8dcd-67fda32884ed", 00:12:13.020 "strip_size_kb": 0, 00:12:13.020 "state": "online", 00:12:13.020 "raid_level": "raid1", 00:12:13.020 "superblock": false, 00:12:13.020 "num_base_bdevs": 2, 00:12:13.020 "num_base_bdevs_discovered": 1, 00:12:13.020 "num_base_bdevs_operational": 1, 00:12:13.020 "base_bdevs_list": [ 00:12:13.020 { 00:12:13.020 "name": null, 00:12:13.020 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:13.020 "is_configured": false, 00:12:13.020 "data_offset": 0, 00:12:13.020 "data_size": 65536 00:12:13.020 }, 00:12:13.020 { 00:12:13.020 "name": "BaseBdev2", 00:12:13.020 "uuid": "4c36ab84-9c35-583f-a6c3-6244915797fd", 00:12:13.020 "is_configured": true, 00:12:13.020 "data_offset": 0, 00:12:13.020 "data_size": 65536 00:12:13.020 } 00:12:13.020 ] 00:12:13.020 }' 00:12:13.021 17:33:44 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:13.021 17:33:44 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:13.021 17:33:44 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:13.281 17:33:44 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:13.281 17:33:44 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:12:13.281 17:33:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:13.281 17:33:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:13.281 [2024-11-27 17:33:44.238541] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:13.281 17:33:44 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:13.281 17:33:44 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@663 -- # sleep 1 00:12:13.281 [2024-11-27 17:33:44.292347] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002870 00:12:13.281 [2024-11-27 17:33:44.294713] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:12:13.281 [2024-11-27 17:33:44.421235] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:12:13.281 [2024-11-27 17:33:44.421708] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:12:13.542 [2024-11-27 17:33:44.556092] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:12:13.801 [2024-11-27 17:33:44.900842] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:12:13.802 [2024-11-27 17:33:44.901597] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:12:14.061 171.00 IOPS, 513.00 MiB/s [2024-11-27T17:33:45.253Z] [2024-11-27 17:33:45.115002] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:12:14.061 [2024-11-27 17:33:45.115358] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:12:14.322 17:33:45 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:14.322 17:33:45 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:14.322 17:33:45 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:14.322 17:33:45 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:14.322 17:33:45 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:14.322 17:33:45 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:14.322 17:33:45 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:14.322 17:33:45 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:14.322 17:33:45 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:14.322 17:33:45 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:14.322 17:33:45 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:14.322 "name": "raid_bdev1", 00:12:14.322 "uuid": "87009852-bfe6-4e82-8dcd-67fda32884ed", 00:12:14.322 "strip_size_kb": 0, 00:12:14.322 "state": "online", 00:12:14.322 "raid_level": "raid1", 00:12:14.322 "superblock": false, 00:12:14.322 "num_base_bdevs": 2, 00:12:14.322 "num_base_bdevs_discovered": 2, 00:12:14.322 "num_base_bdevs_operational": 2, 00:12:14.322 "process": { 00:12:14.322 "type": "rebuild", 00:12:14.322 "target": "spare", 00:12:14.322 "progress": { 00:12:14.322 "blocks": 12288, 00:12:14.322 "percent": 18 00:12:14.322 } 00:12:14.322 }, 00:12:14.322 "base_bdevs_list": [ 00:12:14.322 { 00:12:14.322 "name": "spare", 00:12:14.322 "uuid": "ab1dad7c-8bf0-5526-bf88-45f884961d05", 00:12:14.322 "is_configured": true, 00:12:14.322 "data_offset": 0, 00:12:14.322 "data_size": 65536 00:12:14.322 }, 00:12:14.322 { 00:12:14.322 "name": "BaseBdev2", 00:12:14.322 "uuid": "4c36ab84-9c35-583f-a6c3-6244915797fd", 00:12:14.322 "is_configured": true, 00:12:14.322 "data_offset": 0, 00:12:14.322 "data_size": 65536 00:12:14.322 } 00:12:14.322 ] 00:12:14.322 }' 00:12:14.322 17:33:45 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:14.322 17:33:45 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:14.322 17:33:45 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:14.322 17:33:45 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:14.322 17:33:45 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@666 -- # '[' false = true ']' 00:12:14.322 17:33:45 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=2 00:12:14.322 17:33:45 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:12:14.322 17:33:45 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@693 -- # '[' 2 -gt 2 ']' 00:12:14.322 17:33:45 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@706 -- # local timeout=333 00:12:14.322 17:33:45 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:12:14.322 17:33:45 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:14.322 17:33:45 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:14.322 17:33:45 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:14.322 17:33:45 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:14.322 17:33:45 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:14.322 17:33:45 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:14.322 17:33:45 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:14.322 17:33:45 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:14.322 17:33:45 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:14.322 17:33:45 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:14.322 17:33:45 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:14.322 "name": "raid_bdev1", 00:12:14.322 "uuid": "87009852-bfe6-4e82-8dcd-67fda32884ed", 00:12:14.322 "strip_size_kb": 0, 00:12:14.322 "state": "online", 00:12:14.322 "raid_level": "raid1", 00:12:14.322 "superblock": false, 00:12:14.322 "num_base_bdevs": 2, 00:12:14.322 "num_base_bdevs_discovered": 2, 00:12:14.322 "num_base_bdevs_operational": 2, 00:12:14.322 "process": { 00:12:14.322 "type": "rebuild", 00:12:14.322 "target": "spare", 00:12:14.322 "progress": { 00:12:14.322 "blocks": 14336, 00:12:14.322 "percent": 21 00:12:14.322 } 00:12:14.322 }, 00:12:14.322 "base_bdevs_list": [ 00:12:14.322 { 00:12:14.322 "name": "spare", 00:12:14.322 "uuid": "ab1dad7c-8bf0-5526-bf88-45f884961d05", 00:12:14.322 "is_configured": true, 00:12:14.322 "data_offset": 0, 00:12:14.322 "data_size": 65536 00:12:14.322 }, 00:12:14.322 { 00:12:14.322 "name": "BaseBdev2", 00:12:14.322 "uuid": "4c36ab84-9c35-583f-a6c3-6244915797fd", 00:12:14.322 "is_configured": true, 00:12:14.322 "data_offset": 0, 00:12:14.322 "data_size": 65536 00:12:14.322 } 00:12:14.322 ] 00:12:14.322 }' 00:12:14.322 17:33:45 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:14.322 [2024-11-27 17:33:45.436210] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:12:14.322 17:33:45 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:14.322 17:33:45 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:14.585 17:33:45 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:14.585 17:33:45 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:12:14.585 [2024-11-27 17:33:45.666301] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 20480 offset_begin: 18432 offset_end: 24576 00:12:14.860 [2024-11-27 17:33:45.992207] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 26624 offset_begin: 24576 offset_end: 30720 00:12:15.134 158.00 IOPS, 474.00 MiB/s [2024-11-27T17:33:46.326Z] [2024-11-27 17:33:46.105414] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 28672 offset_begin: 24576 offset_end: 30720 00:12:15.393 17:33:46 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:12:15.393 17:33:46 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:15.393 17:33:46 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:15.393 17:33:46 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:15.393 17:33:46 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:15.393 17:33:46 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:15.393 17:33:46 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:15.393 17:33:46 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:15.393 17:33:46 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:15.393 17:33:46 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:15.393 [2024-11-27 17:33:46.543898] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 34816 offset_begin: 30720 offset_end: 36864 00:12:15.393 17:33:46 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:15.393 17:33:46 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:15.393 "name": "raid_bdev1", 00:12:15.393 "uuid": "87009852-bfe6-4e82-8dcd-67fda32884ed", 00:12:15.393 "strip_size_kb": 0, 00:12:15.393 "state": "online", 00:12:15.393 "raid_level": "raid1", 00:12:15.393 "superblock": false, 00:12:15.393 "num_base_bdevs": 2, 00:12:15.393 "num_base_bdevs_discovered": 2, 00:12:15.393 "num_base_bdevs_operational": 2, 00:12:15.393 "process": { 00:12:15.393 "type": "rebuild", 00:12:15.393 "target": "spare", 00:12:15.393 "progress": { 00:12:15.393 "blocks": 32768, 00:12:15.393 "percent": 50 00:12:15.393 } 00:12:15.393 }, 00:12:15.393 "base_bdevs_list": [ 00:12:15.393 { 00:12:15.393 "name": "spare", 00:12:15.393 "uuid": "ab1dad7c-8bf0-5526-bf88-45f884961d05", 00:12:15.393 "is_configured": true, 00:12:15.393 "data_offset": 0, 00:12:15.393 "data_size": 65536 00:12:15.393 }, 00:12:15.393 { 00:12:15.393 "name": "BaseBdev2", 00:12:15.393 "uuid": "4c36ab84-9c35-583f-a6c3-6244915797fd", 00:12:15.393 "is_configured": true, 00:12:15.393 "data_offset": 0, 00:12:15.393 "data_size": 65536 00:12:15.393 } 00:12:15.393 ] 00:12:15.393 }' 00:12:15.393 17:33:46 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:15.652 17:33:46 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:15.652 17:33:46 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:15.652 17:33:46 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:15.652 17:33:46 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:12:15.913 [2024-11-27 17:33:46.995589] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 40960 offset_begin: 36864 offset_end: 43008 00:12:15.913 [2024-11-27 17:33:46.995838] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 40960 offset_begin: 36864 offset_end: 43008 00:12:16.173 135.80 IOPS, 407.40 MiB/s [2024-11-27T17:33:47.365Z] [2024-11-27 17:33:47.221855] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 45056 offset_begin: 43008 offset_end: 49152 00:12:16.750 [2024-11-27 17:33:47.651820] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 51200 offset_begin: 49152 offset_end: 55296 00:12:16.750 17:33:47 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:12:16.750 17:33:47 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:16.750 17:33:47 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:16.751 17:33:47 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:16.751 17:33:47 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:16.751 17:33:47 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:16.751 17:33:47 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:16.751 17:33:47 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:16.751 17:33:47 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:16.751 17:33:47 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:16.751 17:33:47 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:16.751 17:33:47 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:16.751 "name": "raid_bdev1", 00:12:16.751 "uuid": "87009852-bfe6-4e82-8dcd-67fda32884ed", 00:12:16.751 "strip_size_kb": 0, 00:12:16.751 "state": "online", 00:12:16.751 "raid_level": "raid1", 00:12:16.751 "superblock": false, 00:12:16.751 "num_base_bdevs": 2, 00:12:16.751 "num_base_bdevs_discovered": 2, 00:12:16.751 "num_base_bdevs_operational": 2, 00:12:16.751 "process": { 00:12:16.751 "type": "rebuild", 00:12:16.751 "target": "spare", 00:12:16.751 "progress": { 00:12:16.751 "blocks": 51200, 00:12:16.751 "percent": 78 00:12:16.751 } 00:12:16.751 }, 00:12:16.751 "base_bdevs_list": [ 00:12:16.751 { 00:12:16.751 "name": "spare", 00:12:16.751 "uuid": "ab1dad7c-8bf0-5526-bf88-45f884961d05", 00:12:16.751 "is_configured": true, 00:12:16.751 "data_offset": 0, 00:12:16.751 "data_size": 65536 00:12:16.751 }, 00:12:16.751 { 00:12:16.751 "name": "BaseBdev2", 00:12:16.751 "uuid": "4c36ab84-9c35-583f-a6c3-6244915797fd", 00:12:16.751 "is_configured": true, 00:12:16.751 "data_offset": 0, 00:12:16.751 "data_size": 65536 00:12:16.751 } 00:12:16.751 ] 00:12:16.751 }' 00:12:16.751 17:33:47 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:16.751 17:33:47 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:16.751 17:33:47 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:16.751 17:33:47 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:16.751 17:33:47 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:12:17.011 119.00 IOPS, 357.00 MiB/s [2024-11-27T17:33:48.203Z] [2024-11-27 17:33:48.078011] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 57344 offset_begin: 55296 offset_end: 61440 00:12:17.011 [2024-11-27 17:33:48.078528] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 57344 offset_begin: 55296 offset_end: 61440 00:12:17.580 [2024-11-27 17:33:48.615229] bdev_raid.c:2896:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:12:17.580 [2024-11-27 17:33:48.715005] bdev_raid.c:2558:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:12:17.580 [2024-11-27 17:33:48.717159] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:17.841 17:33:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:12:17.841 17:33:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:17.841 17:33:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:17.841 17:33:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:17.841 17:33:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:17.841 17:33:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:17.841 17:33:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:17.841 17:33:48 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:17.841 17:33:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:17.841 17:33:48 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:17.841 17:33:48 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:17.841 17:33:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:17.841 "name": "raid_bdev1", 00:12:17.841 "uuid": "87009852-bfe6-4e82-8dcd-67fda32884ed", 00:12:17.841 "strip_size_kb": 0, 00:12:17.841 "state": "online", 00:12:17.841 "raid_level": "raid1", 00:12:17.841 "superblock": false, 00:12:17.841 "num_base_bdevs": 2, 00:12:17.841 "num_base_bdevs_discovered": 2, 00:12:17.841 "num_base_bdevs_operational": 2, 00:12:17.841 "base_bdevs_list": [ 00:12:17.841 { 00:12:17.841 "name": "spare", 00:12:17.841 "uuid": "ab1dad7c-8bf0-5526-bf88-45f884961d05", 00:12:17.841 "is_configured": true, 00:12:17.841 "data_offset": 0, 00:12:17.841 "data_size": 65536 00:12:17.841 }, 00:12:17.841 { 00:12:17.841 "name": "BaseBdev2", 00:12:17.841 "uuid": "4c36ab84-9c35-583f-a6c3-6244915797fd", 00:12:17.841 "is_configured": true, 00:12:17.841 "data_offset": 0, 00:12:17.841 "data_size": 65536 00:12:17.841 } 00:12:17.841 ] 00:12:17.841 }' 00:12:17.841 17:33:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:17.841 17:33:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:12:17.841 17:33:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:17.841 17:33:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:12:17.841 17:33:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@709 -- # break 00:12:17.841 17:33:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:17.841 17:33:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:17.841 17:33:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:17.841 17:33:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:17.841 17:33:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:17.841 17:33:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:17.841 17:33:48 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:17.841 17:33:48 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:17.841 17:33:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:17.841 17:33:48 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:17.841 17:33:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:17.841 "name": "raid_bdev1", 00:12:17.841 "uuid": "87009852-bfe6-4e82-8dcd-67fda32884ed", 00:12:17.841 "strip_size_kb": 0, 00:12:17.841 "state": "online", 00:12:17.841 "raid_level": "raid1", 00:12:17.841 "superblock": false, 00:12:17.841 "num_base_bdevs": 2, 00:12:17.841 "num_base_bdevs_discovered": 2, 00:12:17.841 "num_base_bdevs_operational": 2, 00:12:17.841 "base_bdevs_list": [ 00:12:17.841 { 00:12:17.841 "name": "spare", 00:12:17.841 "uuid": "ab1dad7c-8bf0-5526-bf88-45f884961d05", 00:12:17.841 "is_configured": true, 00:12:17.841 "data_offset": 0, 00:12:17.841 "data_size": 65536 00:12:17.841 }, 00:12:17.841 { 00:12:17.841 "name": "BaseBdev2", 00:12:17.841 "uuid": "4c36ab84-9c35-583f-a6c3-6244915797fd", 00:12:17.841 "is_configured": true, 00:12:17.841 "data_offset": 0, 00:12:17.841 "data_size": 65536 00:12:17.841 } 00:12:17.841 ] 00:12:17.841 }' 00:12:17.841 17:33:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:18.102 17:33:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:18.102 107.57 IOPS, 322.71 MiB/s [2024-11-27T17:33:49.294Z] 17:33:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:18.102 17:33:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:18.102 17:33:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:12:18.102 17:33:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:18.102 17:33:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:18.102 17:33:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:18.102 17:33:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:18.102 17:33:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:12:18.102 17:33:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:18.102 17:33:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:18.102 17:33:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:18.102 17:33:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:18.102 17:33:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:18.102 17:33:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:18.102 17:33:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:18.102 17:33:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:18.102 17:33:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:18.102 17:33:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:18.102 "name": "raid_bdev1", 00:12:18.102 "uuid": "87009852-bfe6-4e82-8dcd-67fda32884ed", 00:12:18.102 "strip_size_kb": 0, 00:12:18.102 "state": "online", 00:12:18.102 "raid_level": "raid1", 00:12:18.102 "superblock": false, 00:12:18.102 "num_base_bdevs": 2, 00:12:18.102 "num_base_bdevs_discovered": 2, 00:12:18.102 "num_base_bdevs_operational": 2, 00:12:18.102 "base_bdevs_list": [ 00:12:18.102 { 00:12:18.102 "name": "spare", 00:12:18.102 "uuid": "ab1dad7c-8bf0-5526-bf88-45f884961d05", 00:12:18.102 "is_configured": true, 00:12:18.102 "data_offset": 0, 00:12:18.102 "data_size": 65536 00:12:18.102 }, 00:12:18.102 { 00:12:18.102 "name": "BaseBdev2", 00:12:18.102 "uuid": "4c36ab84-9c35-583f-a6c3-6244915797fd", 00:12:18.102 "is_configured": true, 00:12:18.102 "data_offset": 0, 00:12:18.102 "data_size": 65536 00:12:18.102 } 00:12:18.102 ] 00:12:18.102 }' 00:12:18.102 17:33:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:18.102 17:33:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:18.361 17:33:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:12:18.361 17:33:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:18.361 17:33:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:18.361 [2024-11-27 17:33:49.471509] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:12:18.361 [2024-11-27 17:33:49.471549] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:12:18.361 00:12:18.361 Latency(us) 00:12:18.361 [2024-11-27T17:33:49.553Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:12:18.361 Job: raid_bdev1 (Core Mask 0x1, workload: randrw, percentage: 50, depth: 2, IO size: 3145728) 00:12:18.361 raid_bdev1 : 7.46 102.80 308.41 0.00 0.00 12754.63 280.82 115389.15 00:12:18.361 [2024-11-27T17:33:49.553Z] =================================================================================================================== 00:12:18.361 [2024-11-27T17:33:49.553Z] Total : 102.80 308.41 0.00 0.00 12754.63 280.82 115389.15 00:12:18.361 [2024-11-27 17:33:49.490763] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:18.361 [2024-11-27 17:33:49.490808] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:12:18.361 [2024-11-27 17:33:49.490886] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:12:18.361 [2024-11-27 17:33:49.490901] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:12:18.361 { 00:12:18.361 "results": [ 00:12:18.361 { 00:12:18.361 "job": "raid_bdev1", 00:12:18.361 "core_mask": "0x1", 00:12:18.361 "workload": "randrw", 00:12:18.361 "percentage": 50, 00:12:18.361 "status": "finished", 00:12:18.361 "queue_depth": 2, 00:12:18.361 "io_size": 3145728, 00:12:18.361 "runtime": 7.460957, 00:12:18.361 "iops": 102.80182555669467, 00:12:18.361 "mibps": 308.405476670084, 00:12:18.361 "io_failed": 0, 00:12:18.361 "io_timeout": 0, 00:12:18.361 "avg_latency_us": 12754.630298958682, 00:12:18.361 "min_latency_us": 280.8174672489083, 00:12:18.361 "max_latency_us": 115389.14934497817 00:12:18.361 } 00:12:18.361 ], 00:12:18.361 "core_count": 1 00:12:18.361 } 00:12:18.361 17:33:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:18.361 17:33:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@720 -- # jq length 00:12:18.361 17:33:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:18.361 17:33:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:18.361 17:33:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:18.361 17:33:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:18.361 17:33:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:12:18.361 17:33:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:12:18.361 17:33:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@723 -- # '[' true = true ']' 00:12:18.361 17:33:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@725 -- # nbd_start_disks /var/tmp/spdk.sock spare /dev/nbd0 00:12:18.361 17:33:49 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:12:18.361 17:33:49 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # bdev_list=('spare') 00:12:18.361 17:33:49 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:12:18.361 17:33:49 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:12:18.361 17:33:49 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:12:18.361 17:33:49 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@12 -- # local i 00:12:18.361 17:33:49 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:12:18.361 17:33:49 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:12:18.361 17:33:49 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd0 00:12:18.621 /dev/nbd0 00:12:18.621 17:33:49 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:12:18.621 17:33:49 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:12:18.621 17:33:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:12:18.621 17:33:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@869 -- # local i 00:12:18.621 17:33:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:12:18.621 17:33:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:12:18.621 17:33:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:12:18.621 17:33:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@873 -- # break 00:12:18.621 17:33:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:12:18.621 17:33:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:12:18.621 17:33:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:12:18.621 1+0 records in 00:12:18.621 1+0 records out 00:12:18.621 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000356098 s, 11.5 MB/s 00:12:18.621 17:33:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:18.621 17:33:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@886 -- # size=4096 00:12:18.621 17:33:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:18.621 17:33:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:12:18.621 17:33:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@889 -- # return 0 00:12:18.621 17:33:49 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:12:18.621 17:33:49 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:12:18.621 17:33:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@726 -- # for bdev in "${base_bdevs[@]:1}" 00:12:18.621 17:33:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@727 -- # '[' -z BaseBdev2 ']' 00:12:18.621 17:33:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@730 -- # nbd_start_disks /var/tmp/spdk.sock BaseBdev2 /dev/nbd1 00:12:18.621 17:33:49 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:12:18.621 17:33:49 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev2') 00:12:18.621 17:33:49 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:12:18.621 17:33:49 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd1') 00:12:18.621 17:33:49 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:12:18.621 17:33:49 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@12 -- # local i 00:12:18.621 17:33:49 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:12:18.621 17:33:49 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:12:18.621 17:33:49 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev2 /dev/nbd1 00:12:18.882 /dev/nbd1 00:12:18.882 17:33:50 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:12:18.882 17:33:50 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:12:18.882 17:33:50 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@868 -- # local nbd_name=nbd1 00:12:18.882 17:33:50 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@869 -- # local i 00:12:18.882 17:33:50 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:12:18.882 17:33:50 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:12:18.882 17:33:50 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@872 -- # grep -q -w nbd1 /proc/partitions 00:12:18.882 17:33:50 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@873 -- # break 00:12:18.882 17:33:50 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:12:18.882 17:33:50 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:12:18.882 17:33:50 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@885 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:12:18.882 1+0 records in 00:12:18.882 1+0 records out 00:12:18.882 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000436195 s, 9.4 MB/s 00:12:18.882 17:33:50 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:18.882 17:33:50 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@886 -- # size=4096 00:12:18.882 17:33:50 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:18.882 17:33:50 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:12:18.882 17:33:50 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@889 -- # return 0 00:12:18.882 17:33:50 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:12:18.882 17:33:50 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:12:18.882 17:33:50 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@731 -- # cmp -i 0 /dev/nbd0 /dev/nbd1 00:12:19.142 17:33:50 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@732 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd1 00:12:19.142 17:33:50 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:12:19.142 17:33:50 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd1') 00:12:19.142 17:33:50 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:12:19.142 17:33:50 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@51 -- # local i 00:12:19.142 17:33:50 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:12:19.142 17:33:50 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:12:19.402 17:33:50 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:12:19.402 17:33:50 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:12:19.402 17:33:50 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:12:19.402 17:33:50 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:12:19.402 17:33:50 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:12:19.402 17:33:50 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:12:19.402 17:33:50 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@41 -- # break 00:12:19.402 17:33:50 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@45 -- # return 0 00:12:19.402 17:33:50 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@734 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:12:19.403 17:33:50 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:12:19.403 17:33:50 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:12:19.403 17:33:50 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:12:19.403 17:33:50 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@51 -- # local i 00:12:19.403 17:33:50 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:12:19.403 17:33:50 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:12:19.403 17:33:50 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:12:19.403 17:33:50 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:12:19.403 17:33:50 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:12:19.403 17:33:50 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:12:19.403 17:33:50 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:12:19.403 17:33:50 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:12:19.403 17:33:50 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@41 -- # break 00:12:19.403 17:33:50 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@45 -- # return 0 00:12:19.403 17:33:50 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@743 -- # '[' false = true ']' 00:12:19.403 17:33:50 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@784 -- # killprocess 87003 00:12:19.403 17:33:50 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@950 -- # '[' -z 87003 ']' 00:12:19.403 17:33:50 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@954 -- # kill -0 87003 00:12:19.403 17:33:50 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@955 -- # uname 00:12:19.403 17:33:50 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:12:19.403 17:33:50 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 87003 00:12:19.663 17:33:50 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:12:19.663 17:33:50 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:12:19.663 killing process with pid 87003 00:12:19.663 17:33:50 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@968 -- # echo 'killing process with pid 87003' 00:12:19.663 17:33:50 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@969 -- # kill 87003 00:12:19.663 Received shutdown signal, test time was about 8.572409 seconds 00:12:19.663 00:12:19.663 Latency(us) 00:12:19.663 [2024-11-27T17:33:50.855Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:12:19.663 [2024-11-27T17:33:50.855Z] =================================================================================================================== 00:12:19.663 [2024-11-27T17:33:50.855Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:12:19.663 [2024-11-27 17:33:50.596699] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:12:19.663 17:33:50 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@974 -- # wait 87003 00:12:19.663 [2024-11-27 17:33:50.643123] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:12:19.923 17:33:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@786 -- # return 0 00:12:19.923 00:12:19.923 real 0m10.644s 00:12:19.923 user 0m13.445s 00:12:19.923 sys 0m1.531s 00:12:19.923 17:33:51 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@1126 -- # xtrace_disable 00:12:19.923 17:33:51 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:12:19.923 ************************************ 00:12:19.923 END TEST raid_rebuild_test_io 00:12:19.923 ************************************ 00:12:19.923 17:33:51 bdev_raid -- bdev/bdev_raid.sh@981 -- # run_test raid_rebuild_test_sb_io raid_rebuild_test raid1 2 true true true 00:12:19.923 17:33:51 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 7 -le 1 ']' 00:12:19.923 17:33:51 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:12:19.923 17:33:51 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:12:19.923 ************************************ 00:12:19.923 START TEST raid_rebuild_test_sb_io 00:12:19.923 ************************************ 00:12:19.923 17:33:51 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@1125 -- # raid_rebuild_test raid1 2 true true true 00:12:19.923 17:33:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:12:19.924 17:33:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=2 00:12:19.924 17:33:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:12:19.924 17:33:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@572 -- # local background_io=true 00:12:19.924 17:33:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@573 -- # local verify=true 00:12:19.924 17:33:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:12:19.924 17:33:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:19.924 17:33:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:12:19.924 17:33:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:19.924 17:33:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:19.924 17:33:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:12:19.924 17:33:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:19.924 17:33:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:19.924 17:33:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:12:19.924 17:33:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:12:19.924 17:33:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:12:19.924 17:33:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@576 -- # local strip_size 00:12:19.924 17:33:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@577 -- # local create_arg 00:12:19.924 17:33:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:12:19.924 17:33:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@579 -- # local data_offset 00:12:19.924 17:33:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:12:19.924 17:33:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:12:19.924 17:33:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:12:19.924 17:33:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:12:19.924 17:33:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@597 -- # raid_pid=87372 00:12:19.924 17:33:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@598 -- # waitforlisten 87372 00:12:19.924 17:33:51 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:12:19.924 17:33:51 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@831 -- # '[' -z 87372 ']' 00:12:19.924 17:33:51 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:12:19.924 17:33:51 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@836 -- # local max_retries=100 00:12:19.924 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:12:19.924 17:33:51 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:12:19.924 17:33:51 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@840 -- # xtrace_disable 00:12:19.924 17:33:51 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:20.185 I/O size of 3145728 is greater than zero copy threshold (65536). 00:12:20.185 Zero copy mechanism will not be used. 00:12:20.185 [2024-11-27 17:33:51.180413] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:12:20.185 [2024-11-27 17:33:51.180536] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid87372 ] 00:12:20.185 [2024-11-27 17:33:51.326081] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:12:20.446 [2024-11-27 17:33:51.394079] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:12:20.446 [2024-11-27 17:33:51.469373] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:12:20.446 [2024-11-27 17:33:51.469413] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:12:21.015 17:33:51 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:12:21.015 17:33:52 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@864 -- # return 0 00:12:21.015 17:33:52 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:21.015 17:33:52 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:12:21.015 17:33:52 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:21.015 17:33:52 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:21.015 BaseBdev1_malloc 00:12:21.015 17:33:52 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:21.015 17:33:52 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:12:21.015 17:33:52 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:21.015 17:33:52 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:21.015 [2024-11-27 17:33:52.031063] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:12:21.015 [2024-11-27 17:33:52.031126] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:21.015 [2024-11-27 17:33:52.031163] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:12:21.015 [2024-11-27 17:33:52.031179] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:21.015 [2024-11-27 17:33:52.033582] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:21.015 [2024-11-27 17:33:52.033619] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:12:21.015 BaseBdev1 00:12:21.015 17:33:52 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:21.015 17:33:52 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:21.015 17:33:52 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:12:21.015 17:33:52 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:21.015 17:33:52 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:21.015 BaseBdev2_malloc 00:12:21.015 17:33:52 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:21.015 17:33:52 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:12:21.015 17:33:52 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:21.015 17:33:52 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:21.015 [2024-11-27 17:33:52.078491] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:12:21.015 [2024-11-27 17:33:52.078561] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:21.015 [2024-11-27 17:33:52.078595] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:12:21.015 [2024-11-27 17:33:52.078613] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:21.015 [2024-11-27 17:33:52.082389] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:21.015 [2024-11-27 17:33:52.082422] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:12:21.015 BaseBdev2 00:12:21.015 17:33:52 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:21.015 17:33:52 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:12:21.015 17:33:52 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:21.015 17:33:52 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:21.015 spare_malloc 00:12:21.015 17:33:52 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:21.015 17:33:52 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:12:21.015 17:33:52 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:21.015 17:33:52 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:21.015 spare_delay 00:12:21.015 17:33:52 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:21.015 17:33:52 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:12:21.015 17:33:52 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:21.015 17:33:52 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:21.015 [2024-11-27 17:33:52.125399] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:12:21.015 [2024-11-27 17:33:52.125449] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:21.015 [2024-11-27 17:33:52.125473] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:12:21.015 [2024-11-27 17:33:52.125482] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:21.015 [2024-11-27 17:33:52.127919] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:21.015 [2024-11-27 17:33:52.127948] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:12:21.015 spare 00:12:21.015 17:33:52 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:21.015 17:33:52 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 00:12:21.015 17:33:52 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:21.015 17:33:52 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:21.015 [2024-11-27 17:33:52.137441] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:12:21.015 [2024-11-27 17:33:52.139577] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:12:21.015 [2024-11-27 17:33:52.139737] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:12:21.015 [2024-11-27 17:33:52.139755] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:12:21.016 [2024-11-27 17:33:52.140046] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:12:21.016 [2024-11-27 17:33:52.140212] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:12:21.016 [2024-11-27 17:33:52.140231] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:12:21.016 [2024-11-27 17:33:52.140360] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:21.016 17:33:52 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:21.016 17:33:52 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:12:21.016 17:33:52 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:21.016 17:33:52 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:21.016 17:33:52 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:21.016 17:33:52 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:21.016 17:33:52 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:12:21.016 17:33:52 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:21.016 17:33:52 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:21.016 17:33:52 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:21.016 17:33:52 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:21.016 17:33:52 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:21.016 17:33:52 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:21.016 17:33:52 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:21.016 17:33:52 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:21.016 17:33:52 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:21.016 17:33:52 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:21.016 "name": "raid_bdev1", 00:12:21.016 "uuid": "e16f9deb-e8ff-4a2e-abdc-612c84f97b3b", 00:12:21.016 "strip_size_kb": 0, 00:12:21.016 "state": "online", 00:12:21.016 "raid_level": "raid1", 00:12:21.016 "superblock": true, 00:12:21.016 "num_base_bdevs": 2, 00:12:21.016 "num_base_bdevs_discovered": 2, 00:12:21.016 "num_base_bdevs_operational": 2, 00:12:21.016 "base_bdevs_list": [ 00:12:21.016 { 00:12:21.016 "name": "BaseBdev1", 00:12:21.016 "uuid": "a80e574d-ff14-56ee-b437-89abc2a63197", 00:12:21.016 "is_configured": true, 00:12:21.016 "data_offset": 2048, 00:12:21.016 "data_size": 63488 00:12:21.016 }, 00:12:21.016 { 00:12:21.016 "name": "BaseBdev2", 00:12:21.016 "uuid": "b461fd4c-c69e-5eb3-9eb0-b3fe61cb5d81", 00:12:21.016 "is_configured": true, 00:12:21.016 "data_offset": 2048, 00:12:21.016 "data_size": 63488 00:12:21.016 } 00:12:21.016 ] 00:12:21.016 }' 00:12:21.016 17:33:52 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:21.016 17:33:52 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:21.586 17:33:52 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:12:21.586 17:33:52 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:12:21.586 17:33:52 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:21.586 17:33:52 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:21.586 [2024-11-27 17:33:52.572898] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:12:21.586 17:33:52 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:21.586 17:33:52 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=63488 00:12:21.586 17:33:52 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:21.586 17:33:52 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:21.586 17:33:52 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:12:21.586 17:33:52 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:21.586 17:33:52 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:21.586 17:33:52 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@619 -- # data_offset=2048 00:12:21.586 17:33:52 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@621 -- # '[' true = true ']' 00:12:21.586 17:33:52 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:12:21.586 17:33:52 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@623 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:12:21.586 17:33:52 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:21.586 17:33:52 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:21.586 [2024-11-27 17:33:52.644503] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:12:21.586 17:33:52 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:21.586 17:33:52 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:12:21.586 17:33:52 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:21.586 17:33:52 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:21.586 17:33:52 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:21.586 17:33:52 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:21.586 17:33:52 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:12:21.586 17:33:52 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:21.586 17:33:52 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:21.586 17:33:52 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:21.586 17:33:52 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:21.586 17:33:52 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:21.586 17:33:52 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:21.586 17:33:52 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:21.586 17:33:52 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:21.586 17:33:52 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:21.586 17:33:52 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:21.586 "name": "raid_bdev1", 00:12:21.586 "uuid": "e16f9deb-e8ff-4a2e-abdc-612c84f97b3b", 00:12:21.586 "strip_size_kb": 0, 00:12:21.586 "state": "online", 00:12:21.586 "raid_level": "raid1", 00:12:21.586 "superblock": true, 00:12:21.586 "num_base_bdevs": 2, 00:12:21.586 "num_base_bdevs_discovered": 1, 00:12:21.586 "num_base_bdevs_operational": 1, 00:12:21.586 "base_bdevs_list": [ 00:12:21.586 { 00:12:21.586 "name": null, 00:12:21.586 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:21.586 "is_configured": false, 00:12:21.586 "data_offset": 0, 00:12:21.586 "data_size": 63488 00:12:21.586 }, 00:12:21.586 { 00:12:21.586 "name": "BaseBdev2", 00:12:21.586 "uuid": "b461fd4c-c69e-5eb3-9eb0-b3fe61cb5d81", 00:12:21.586 "is_configured": true, 00:12:21.586 "data_offset": 2048, 00:12:21.586 "data_size": 63488 00:12:21.586 } 00:12:21.586 ] 00:12:21.586 }' 00:12:21.586 17:33:52 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:21.586 17:33:52 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:21.586 [2024-11-27 17:33:52.731823] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000026d0 00:12:21.586 I/O size of 3145728 is greater than zero copy threshold (65536). 00:12:21.586 Zero copy mechanism will not be used. 00:12:21.586 Running I/O for 60 seconds... 00:12:21.846 17:33:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:12:21.847 17:33:53 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:21.847 17:33:53 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:21.847 [2024-11-27 17:33:53.033383] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:22.106 17:33:53 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:22.106 17:33:53 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@647 -- # sleep 1 00:12:22.106 [2024-11-27 17:33:53.055013] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000027a0 00:12:22.106 [2024-11-27 17:33:53.057379] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:12:22.106 [2024-11-27 17:33:53.166291] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:12:22.106 [2024-11-27 17:33:53.166975] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:12:22.367 [2024-11-27 17:33:53.381231] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:12:22.367 [2024-11-27 17:33:53.381604] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:12:22.627 [2024-11-27 17:33:53.710822] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:12:22.887 187.00 IOPS, 561.00 MiB/s [2024-11-27T17:33:54.079Z] [2024-11-27 17:33:53.932761] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:12:22.887 [2024-11-27 17:33:53.933010] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:12:22.887 17:33:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:22.887 17:33:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:22.887 17:33:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:22.887 17:33:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:22.887 17:33:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:22.887 17:33:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:22.887 17:33:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:22.887 17:33:54 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:22.887 17:33:54 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:23.147 17:33:54 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:23.147 17:33:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:23.147 "name": "raid_bdev1", 00:12:23.147 "uuid": "e16f9deb-e8ff-4a2e-abdc-612c84f97b3b", 00:12:23.147 "strip_size_kb": 0, 00:12:23.147 "state": "online", 00:12:23.147 "raid_level": "raid1", 00:12:23.147 "superblock": true, 00:12:23.147 "num_base_bdevs": 2, 00:12:23.147 "num_base_bdevs_discovered": 2, 00:12:23.147 "num_base_bdevs_operational": 2, 00:12:23.147 "process": { 00:12:23.147 "type": "rebuild", 00:12:23.147 "target": "spare", 00:12:23.147 "progress": { 00:12:23.147 "blocks": 10240, 00:12:23.147 "percent": 16 00:12:23.147 } 00:12:23.147 }, 00:12:23.147 "base_bdevs_list": [ 00:12:23.147 { 00:12:23.147 "name": "spare", 00:12:23.147 "uuid": "92d0d854-0b10-5ecf-b7c7-8628666a0f48", 00:12:23.147 "is_configured": true, 00:12:23.147 "data_offset": 2048, 00:12:23.147 "data_size": 63488 00:12:23.147 }, 00:12:23.147 { 00:12:23.147 "name": "BaseBdev2", 00:12:23.147 "uuid": "b461fd4c-c69e-5eb3-9eb0-b3fe61cb5d81", 00:12:23.147 "is_configured": true, 00:12:23.147 "data_offset": 2048, 00:12:23.147 "data_size": 63488 00:12:23.147 } 00:12:23.147 ] 00:12:23.147 }' 00:12:23.147 17:33:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:23.147 17:33:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:23.147 17:33:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:23.147 17:33:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:23.147 17:33:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:12:23.147 17:33:54 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:23.147 17:33:54 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:23.147 [2024-11-27 17:33:54.215563] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:23.147 [2024-11-27 17:33:54.280650] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:12:23.147 [2024-11-27 17:33:54.288712] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:23.147 [2024-11-27 17:33:54.288749] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:23.147 [2024-11-27 17:33:54.288764] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:12:23.147 [2024-11-27 17:33:54.319540] bdev_raid.c:1970:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 0 raid_ch: 0x60d0000026d0 00:12:23.147 17:33:54 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:23.147 17:33:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:12:23.147 17:33:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:23.147 17:33:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:23.147 17:33:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:23.147 17:33:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:23.147 17:33:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:12:23.147 17:33:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:23.147 17:33:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:23.147 17:33:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:23.147 17:33:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:23.407 17:33:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:23.407 17:33:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:23.407 17:33:54 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:23.407 17:33:54 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:23.407 17:33:54 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:23.407 17:33:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:23.407 "name": "raid_bdev1", 00:12:23.407 "uuid": "e16f9deb-e8ff-4a2e-abdc-612c84f97b3b", 00:12:23.407 "strip_size_kb": 0, 00:12:23.407 "state": "online", 00:12:23.407 "raid_level": "raid1", 00:12:23.408 "superblock": true, 00:12:23.408 "num_base_bdevs": 2, 00:12:23.408 "num_base_bdevs_discovered": 1, 00:12:23.408 "num_base_bdevs_operational": 1, 00:12:23.408 "base_bdevs_list": [ 00:12:23.408 { 00:12:23.408 "name": null, 00:12:23.408 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:23.408 "is_configured": false, 00:12:23.408 "data_offset": 0, 00:12:23.408 "data_size": 63488 00:12:23.408 }, 00:12:23.408 { 00:12:23.408 "name": "BaseBdev2", 00:12:23.408 "uuid": "b461fd4c-c69e-5eb3-9eb0-b3fe61cb5d81", 00:12:23.408 "is_configured": true, 00:12:23.408 "data_offset": 2048, 00:12:23.408 "data_size": 63488 00:12:23.408 } 00:12:23.408 ] 00:12:23.408 }' 00:12:23.408 17:33:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:23.408 17:33:54 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:23.667 180.00 IOPS, 540.00 MiB/s [2024-11-27T17:33:54.859Z] 17:33:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:23.667 17:33:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:23.667 17:33:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:23.667 17:33:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:23.667 17:33:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:23.667 17:33:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:23.667 17:33:54 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:23.667 17:33:54 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:23.667 17:33:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:23.667 17:33:54 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:23.667 17:33:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:23.667 "name": "raid_bdev1", 00:12:23.667 "uuid": "e16f9deb-e8ff-4a2e-abdc-612c84f97b3b", 00:12:23.667 "strip_size_kb": 0, 00:12:23.667 "state": "online", 00:12:23.667 "raid_level": "raid1", 00:12:23.667 "superblock": true, 00:12:23.667 "num_base_bdevs": 2, 00:12:23.667 "num_base_bdevs_discovered": 1, 00:12:23.667 "num_base_bdevs_operational": 1, 00:12:23.667 "base_bdevs_list": [ 00:12:23.667 { 00:12:23.667 "name": null, 00:12:23.667 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:23.667 "is_configured": false, 00:12:23.667 "data_offset": 0, 00:12:23.667 "data_size": 63488 00:12:23.667 }, 00:12:23.667 { 00:12:23.667 "name": "BaseBdev2", 00:12:23.667 "uuid": "b461fd4c-c69e-5eb3-9eb0-b3fe61cb5d81", 00:12:23.667 "is_configured": true, 00:12:23.667 "data_offset": 2048, 00:12:23.667 "data_size": 63488 00:12:23.667 } 00:12:23.667 ] 00:12:23.667 }' 00:12:23.667 17:33:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:23.667 17:33:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:23.667 17:33:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:23.927 17:33:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:23.927 17:33:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:12:23.927 17:33:54 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:23.927 17:33:54 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:23.927 [2024-11-27 17:33:54.865453] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:23.927 17:33:54 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:23.927 17:33:54 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@663 -- # sleep 1 00:12:23.927 [2024-11-27 17:33:54.906701] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002870 00:12:23.927 [2024-11-27 17:33:54.909072] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:12:23.927 [2024-11-27 17:33:55.022409] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:12:23.927 [2024-11-27 17:33:55.022975] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:12:24.187 [2024-11-27 17:33:55.236452] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:12:24.187 [2024-11-27 17:33:55.236809] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:12:24.447 [2024-11-27 17:33:55.468400] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:12:24.706 [2024-11-27 17:33:55.698464] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:12:24.706 [2024-11-27 17:33:55.698717] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:12:24.967 164.67 IOPS, 494.00 MiB/s [2024-11-27T17:33:56.159Z] 17:33:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:24.967 17:33:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:24.967 17:33:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:24.967 17:33:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:24.967 17:33:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:24.967 17:33:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:24.967 17:33:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:24.967 17:33:55 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:24.967 17:33:55 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:24.967 17:33:55 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:24.967 [2024-11-27 17:33:55.941191] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 14336 offset_begin: 12288 offset_end: 18432 00:12:24.967 [2024-11-27 17:33:55.947819] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 14336 offset_begin: 12288 offset_end: 18432 00:12:24.967 17:33:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:24.967 "name": "raid_bdev1", 00:12:24.967 "uuid": "e16f9deb-e8ff-4a2e-abdc-612c84f97b3b", 00:12:24.967 "strip_size_kb": 0, 00:12:24.967 "state": "online", 00:12:24.967 "raid_level": "raid1", 00:12:24.967 "superblock": true, 00:12:24.967 "num_base_bdevs": 2, 00:12:24.967 "num_base_bdevs_discovered": 2, 00:12:24.967 "num_base_bdevs_operational": 2, 00:12:24.967 "process": { 00:12:24.967 "type": "rebuild", 00:12:24.967 "target": "spare", 00:12:24.967 "progress": { 00:12:24.967 "blocks": 12288, 00:12:24.967 "percent": 19 00:12:24.967 } 00:12:24.967 }, 00:12:24.967 "base_bdevs_list": [ 00:12:24.967 { 00:12:24.967 "name": "spare", 00:12:24.967 "uuid": "92d0d854-0b10-5ecf-b7c7-8628666a0f48", 00:12:24.967 "is_configured": true, 00:12:24.967 "data_offset": 2048, 00:12:24.967 "data_size": 63488 00:12:24.967 }, 00:12:24.967 { 00:12:24.967 "name": "BaseBdev2", 00:12:24.967 "uuid": "b461fd4c-c69e-5eb3-9eb0-b3fe61cb5d81", 00:12:24.967 "is_configured": true, 00:12:24.967 "data_offset": 2048, 00:12:24.967 "data_size": 63488 00:12:24.967 } 00:12:24.967 ] 00:12:24.967 }' 00:12:24.967 17:33:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:24.967 17:33:55 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:24.967 17:33:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:24.967 17:33:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:24.967 17:33:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:12:24.967 17:33:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:12:24.967 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:12:24.967 17:33:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=2 00:12:24.967 17:33:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:12:24.967 17:33:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@693 -- # '[' 2 -gt 2 ']' 00:12:24.967 17:33:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@706 -- # local timeout=344 00:12:24.967 17:33:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:12:24.967 17:33:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:24.967 17:33:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:24.967 17:33:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:24.967 17:33:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:24.967 17:33:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:24.968 17:33:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:24.968 17:33:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:24.968 17:33:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:24.968 17:33:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:24.968 17:33:56 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:24.968 17:33:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:24.968 "name": "raid_bdev1", 00:12:24.968 "uuid": "e16f9deb-e8ff-4a2e-abdc-612c84f97b3b", 00:12:24.968 "strip_size_kb": 0, 00:12:24.968 "state": "online", 00:12:24.968 "raid_level": "raid1", 00:12:24.968 "superblock": true, 00:12:24.968 "num_base_bdevs": 2, 00:12:24.968 "num_base_bdevs_discovered": 2, 00:12:24.968 "num_base_bdevs_operational": 2, 00:12:24.968 "process": { 00:12:24.968 "type": "rebuild", 00:12:24.968 "target": "spare", 00:12:24.968 "progress": { 00:12:24.968 "blocks": 14336, 00:12:24.968 "percent": 22 00:12:24.968 } 00:12:24.968 }, 00:12:24.968 "base_bdevs_list": [ 00:12:24.968 { 00:12:24.968 "name": "spare", 00:12:24.968 "uuid": "92d0d854-0b10-5ecf-b7c7-8628666a0f48", 00:12:24.968 "is_configured": true, 00:12:24.968 "data_offset": 2048, 00:12:24.968 "data_size": 63488 00:12:24.968 }, 00:12:24.968 { 00:12:24.968 "name": "BaseBdev2", 00:12:24.968 "uuid": "b461fd4c-c69e-5eb3-9eb0-b3fe61cb5d81", 00:12:24.968 "is_configured": true, 00:12:24.968 "data_offset": 2048, 00:12:24.968 "data_size": 63488 00:12:24.968 } 00:12:24.968 ] 00:12:24.968 }' 00:12:24.968 17:33:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:24.968 17:33:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:24.968 17:33:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:25.228 [2024-11-27 17:33:56.169947] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:12:25.228 17:33:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:25.228 17:33:56 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:12:25.488 [2024-11-27 17:33:56.497859] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 20480 offset_begin: 18432 offset_end: 24576 00:12:25.488 [2024-11-27 17:33:56.498414] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 20480 offset_begin: 18432 offset_end: 24576 00:12:25.748 [2024-11-27 17:33:56.707016] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 22528 offset_begin: 18432 offset_end: 24576 00:12:25.748 [2024-11-27 17:33:56.707275] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 22528 offset_begin: 18432 offset_end: 24576 00:12:26.009 137.50 IOPS, 412.50 MiB/s [2024-11-27T17:33:57.201Z] 17:33:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:12:26.009 17:33:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:26.009 17:33:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:26.009 17:33:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:26.009 17:33:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:26.009 17:33:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:26.009 17:33:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:26.009 17:33:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:26.009 17:33:57 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:26.009 17:33:57 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:26.270 17:33:57 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:26.270 17:33:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:26.270 "name": "raid_bdev1", 00:12:26.270 "uuid": "e16f9deb-e8ff-4a2e-abdc-612c84f97b3b", 00:12:26.270 "strip_size_kb": 0, 00:12:26.270 "state": "online", 00:12:26.270 "raid_level": "raid1", 00:12:26.270 "superblock": true, 00:12:26.270 "num_base_bdevs": 2, 00:12:26.270 "num_base_bdevs_discovered": 2, 00:12:26.270 "num_base_bdevs_operational": 2, 00:12:26.270 "process": { 00:12:26.270 "type": "rebuild", 00:12:26.270 "target": "spare", 00:12:26.270 "progress": { 00:12:26.270 "blocks": 28672, 00:12:26.270 "percent": 45 00:12:26.270 } 00:12:26.270 }, 00:12:26.270 "base_bdevs_list": [ 00:12:26.270 { 00:12:26.270 "name": "spare", 00:12:26.270 "uuid": "92d0d854-0b10-5ecf-b7c7-8628666a0f48", 00:12:26.270 "is_configured": true, 00:12:26.270 "data_offset": 2048, 00:12:26.270 "data_size": 63488 00:12:26.270 }, 00:12:26.270 { 00:12:26.270 "name": "BaseBdev2", 00:12:26.270 "uuid": "b461fd4c-c69e-5eb3-9eb0-b3fe61cb5d81", 00:12:26.270 "is_configured": true, 00:12:26.270 "data_offset": 2048, 00:12:26.270 "data_size": 63488 00:12:26.270 } 00:12:26.270 ] 00:12:26.270 }' 00:12:26.270 17:33:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:26.270 17:33:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:26.270 17:33:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:26.270 17:33:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:26.270 17:33:57 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:12:26.270 [2024-11-27 17:33:57.400763] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 32768 offset_begin: 30720 offset_end: 36864 00:12:26.270 [2024-11-27 17:33:57.401188] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 32768 offset_begin: 30720 offset_end: 36864 00:12:26.841 119.20 IOPS, 357.60 MiB/s [2024-11-27T17:33:58.033Z] [2024-11-27 17:33:57.859967] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 38912 offset_begin: 36864 offset_end: 43008 00:12:26.841 [2024-11-27 17:33:57.860411] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 38912 offset_begin: 36864 offset_end: 43008 00:12:27.102 [2024-11-27 17:33:58.068792] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 40960 offset_begin: 36864 offset_end: 43008 00:12:27.362 17:33:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:12:27.362 17:33:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:27.362 17:33:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:27.362 17:33:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:27.362 17:33:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:27.362 17:33:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:27.362 17:33:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:27.362 17:33:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:27.362 17:33:58 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:27.362 17:33:58 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:27.362 17:33:58 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:27.362 17:33:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:27.362 "name": "raid_bdev1", 00:12:27.362 "uuid": "e16f9deb-e8ff-4a2e-abdc-612c84f97b3b", 00:12:27.362 "strip_size_kb": 0, 00:12:27.362 "state": "online", 00:12:27.362 "raid_level": "raid1", 00:12:27.362 "superblock": true, 00:12:27.362 "num_base_bdevs": 2, 00:12:27.362 "num_base_bdevs_discovered": 2, 00:12:27.362 "num_base_bdevs_operational": 2, 00:12:27.362 "process": { 00:12:27.362 "type": "rebuild", 00:12:27.362 "target": "spare", 00:12:27.362 "progress": { 00:12:27.362 "blocks": 45056, 00:12:27.362 "percent": 70 00:12:27.362 } 00:12:27.362 }, 00:12:27.362 "base_bdevs_list": [ 00:12:27.362 { 00:12:27.362 "name": "spare", 00:12:27.362 "uuid": "92d0d854-0b10-5ecf-b7c7-8628666a0f48", 00:12:27.362 "is_configured": true, 00:12:27.362 "data_offset": 2048, 00:12:27.362 "data_size": 63488 00:12:27.362 }, 00:12:27.362 { 00:12:27.362 "name": "BaseBdev2", 00:12:27.362 "uuid": "b461fd4c-c69e-5eb3-9eb0-b3fe61cb5d81", 00:12:27.362 "is_configured": true, 00:12:27.362 "data_offset": 2048, 00:12:27.362 "data_size": 63488 00:12:27.362 } 00:12:27.362 ] 00:12:27.362 }' 00:12:27.362 17:33:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:27.362 17:33:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:27.362 17:33:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:27.362 17:33:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:27.362 17:33:58 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:12:27.629 105.17 IOPS, 315.50 MiB/s [2024-11-27T17:33:58.821Z] [2024-11-27 17:33:58.739706] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 53248 offset_begin: 49152 offset_end: 55296 00:12:28.215 [2024-11-27 17:33:59.394440] bdev_raid.c:2896:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:12:28.480 [2024-11-27 17:33:59.494329] bdev_raid.c:2558:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:12:28.480 17:33:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:12:28.480 17:33:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:28.480 17:33:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:28.480 17:33:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:28.480 17:33:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:28.480 [2024-11-27 17:33:59.496092] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:28.480 17:33:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:28.480 17:33:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:28.480 17:33:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:28.480 17:33:59 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:28.480 17:33:59 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:28.480 17:33:59 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:28.480 17:33:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:28.480 "name": "raid_bdev1", 00:12:28.480 "uuid": "e16f9deb-e8ff-4a2e-abdc-612c84f97b3b", 00:12:28.480 "strip_size_kb": 0, 00:12:28.480 "state": "online", 00:12:28.480 "raid_level": "raid1", 00:12:28.480 "superblock": true, 00:12:28.480 "num_base_bdevs": 2, 00:12:28.480 "num_base_bdevs_discovered": 2, 00:12:28.480 "num_base_bdevs_operational": 2, 00:12:28.480 "base_bdevs_list": [ 00:12:28.480 { 00:12:28.480 "name": "spare", 00:12:28.480 "uuid": "92d0d854-0b10-5ecf-b7c7-8628666a0f48", 00:12:28.480 "is_configured": true, 00:12:28.480 "data_offset": 2048, 00:12:28.480 "data_size": 63488 00:12:28.480 }, 00:12:28.480 { 00:12:28.480 "name": "BaseBdev2", 00:12:28.480 "uuid": "b461fd4c-c69e-5eb3-9eb0-b3fe61cb5d81", 00:12:28.480 "is_configured": true, 00:12:28.480 "data_offset": 2048, 00:12:28.480 "data_size": 63488 00:12:28.480 } 00:12:28.480 ] 00:12:28.480 }' 00:12:28.480 17:33:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:28.480 17:33:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:12:28.480 17:33:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:28.480 17:33:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:12:28.480 17:33:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@709 -- # break 00:12:28.480 17:33:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:28.480 17:33:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:28.480 17:33:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:28.480 17:33:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:28.480 17:33:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:28.480 17:33:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:28.480 17:33:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:28.480 17:33:59 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:28.480 17:33:59 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:28.480 17:33:59 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:28.740 17:33:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:28.740 "name": "raid_bdev1", 00:12:28.740 "uuid": "e16f9deb-e8ff-4a2e-abdc-612c84f97b3b", 00:12:28.740 "strip_size_kb": 0, 00:12:28.740 "state": "online", 00:12:28.740 "raid_level": "raid1", 00:12:28.740 "superblock": true, 00:12:28.740 "num_base_bdevs": 2, 00:12:28.740 "num_base_bdevs_discovered": 2, 00:12:28.740 "num_base_bdevs_operational": 2, 00:12:28.740 "base_bdevs_list": [ 00:12:28.740 { 00:12:28.740 "name": "spare", 00:12:28.740 "uuid": "92d0d854-0b10-5ecf-b7c7-8628666a0f48", 00:12:28.740 "is_configured": true, 00:12:28.740 "data_offset": 2048, 00:12:28.740 "data_size": 63488 00:12:28.740 }, 00:12:28.740 { 00:12:28.740 "name": "BaseBdev2", 00:12:28.740 "uuid": "b461fd4c-c69e-5eb3-9eb0-b3fe61cb5d81", 00:12:28.740 "is_configured": true, 00:12:28.740 "data_offset": 2048, 00:12:28.740 "data_size": 63488 00:12:28.740 } 00:12:28.740 ] 00:12:28.740 }' 00:12:28.740 17:33:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:28.740 95.29 IOPS, 285.86 MiB/s [2024-11-27T17:33:59.932Z] 17:33:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:28.740 17:33:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:28.740 17:33:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:28.740 17:33:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:12:28.740 17:33:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:28.740 17:33:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:28.740 17:33:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:28.740 17:33:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:28.740 17:33:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:12:28.740 17:33:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:28.740 17:33:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:28.740 17:33:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:28.740 17:33:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:28.740 17:33:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:28.740 17:33:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:28.740 17:33:59 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:28.740 17:33:59 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:28.740 17:33:59 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:28.740 17:33:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:28.740 "name": "raid_bdev1", 00:12:28.740 "uuid": "e16f9deb-e8ff-4a2e-abdc-612c84f97b3b", 00:12:28.740 "strip_size_kb": 0, 00:12:28.740 "state": "online", 00:12:28.740 "raid_level": "raid1", 00:12:28.740 "superblock": true, 00:12:28.740 "num_base_bdevs": 2, 00:12:28.740 "num_base_bdevs_discovered": 2, 00:12:28.740 "num_base_bdevs_operational": 2, 00:12:28.740 "base_bdevs_list": [ 00:12:28.740 { 00:12:28.740 "name": "spare", 00:12:28.740 "uuid": "92d0d854-0b10-5ecf-b7c7-8628666a0f48", 00:12:28.740 "is_configured": true, 00:12:28.740 "data_offset": 2048, 00:12:28.740 "data_size": 63488 00:12:28.740 }, 00:12:28.740 { 00:12:28.740 "name": "BaseBdev2", 00:12:28.740 "uuid": "b461fd4c-c69e-5eb3-9eb0-b3fe61cb5d81", 00:12:28.740 "is_configured": true, 00:12:28.740 "data_offset": 2048, 00:12:28.740 "data_size": 63488 00:12:28.740 } 00:12:28.740 ] 00:12:28.740 }' 00:12:28.740 17:33:59 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:28.740 17:33:59 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:29.310 17:34:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:12:29.310 17:34:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:29.310 17:34:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:29.310 [2024-11-27 17:34:00.241699] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:12:29.310 [2024-11-27 17:34:00.241796] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:12:29.310 00:12:29.310 Latency(us) 00:12:29.310 [2024-11-27T17:34:00.502Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:12:29.310 Job: raid_bdev1 (Core Mask 0x1, workload: randrw, percentage: 50, depth: 2, IO size: 3145728) 00:12:29.310 raid_bdev1 : 7.56 90.77 272.30 0.00 0.00 14657.46 266.51 112641.79 00:12:29.310 [2024-11-27T17:34:00.502Z] =================================================================================================================== 00:12:29.310 [2024-11-27T17:34:00.502Z] Total : 90.77 272.30 0.00 0.00 14657.46 266.51 112641.79 00:12:29.310 [2024-11-27 17:34:00.282587] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:29.310 [2024-11-27 17:34:00.282675] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:12:29.310 [2024-11-27 17:34:00.282812] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:12:29.310 [2024-11-27 17:34:00.282826] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:12:29.310 { 00:12:29.310 "results": [ 00:12:29.310 { 00:12:29.310 "job": "raid_bdev1", 00:12:29.310 "core_mask": "0x1", 00:12:29.310 "workload": "randrw", 00:12:29.310 "percentage": 50, 00:12:29.310 "status": "finished", 00:12:29.310 "queue_depth": 2, 00:12:29.310 "io_size": 3145728, 00:12:29.310 "runtime": 7.557932, 00:12:29.310 "iops": 90.76556920596798, 00:12:29.310 "mibps": 272.29670761790396, 00:12:29.310 "io_failed": 0, 00:12:29.310 "io_timeout": 0, 00:12:29.310 "avg_latency_us": 14657.457023183571, 00:12:29.310 "min_latency_us": 266.5082969432314, 00:12:29.310 "max_latency_us": 112641.78864628822 00:12:29.310 } 00:12:29.310 ], 00:12:29.310 "core_count": 1 00:12:29.310 } 00:12:29.310 17:34:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:29.310 17:34:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:29.310 17:34:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@720 -- # jq length 00:12:29.310 17:34:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:29.310 17:34:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:29.310 17:34:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:29.310 17:34:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:12:29.310 17:34:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:12:29.310 17:34:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@723 -- # '[' true = true ']' 00:12:29.310 17:34:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@725 -- # nbd_start_disks /var/tmp/spdk.sock spare /dev/nbd0 00:12:29.310 17:34:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:12:29.310 17:34:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # bdev_list=('spare') 00:12:29.310 17:34:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:12:29.310 17:34:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:12:29.310 17:34:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:12:29.310 17:34:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@12 -- # local i 00:12:29.310 17:34:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:12:29.310 17:34:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:12:29.310 17:34:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd0 00:12:29.570 /dev/nbd0 00:12:29.570 17:34:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:12:29.570 17:34:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:12:29.570 17:34:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:12:29.570 17:34:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@869 -- # local i 00:12:29.570 17:34:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:12:29.570 17:34:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:12:29.570 17:34:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:12:29.570 17:34:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@873 -- # break 00:12:29.570 17:34:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:12:29.570 17:34:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:12:29.570 17:34:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:12:29.570 1+0 records in 00:12:29.570 1+0 records out 00:12:29.570 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000545539 s, 7.5 MB/s 00:12:29.570 17:34:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:29.570 17:34:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@886 -- # size=4096 00:12:29.570 17:34:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:29.570 17:34:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:12:29.570 17:34:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@889 -- # return 0 00:12:29.570 17:34:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:12:29.570 17:34:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:12:29.570 17:34:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@726 -- # for bdev in "${base_bdevs[@]:1}" 00:12:29.570 17:34:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@727 -- # '[' -z BaseBdev2 ']' 00:12:29.570 17:34:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@730 -- # nbd_start_disks /var/tmp/spdk.sock BaseBdev2 /dev/nbd1 00:12:29.570 17:34:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:12:29.570 17:34:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev2') 00:12:29.570 17:34:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:12:29.570 17:34:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd1') 00:12:29.571 17:34:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:12:29.571 17:34:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@12 -- # local i 00:12:29.571 17:34:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:12:29.571 17:34:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:12:29.571 17:34:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev2 /dev/nbd1 00:12:29.831 /dev/nbd1 00:12:29.831 17:34:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:12:29.831 17:34:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:12:29.831 17:34:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@868 -- # local nbd_name=nbd1 00:12:29.831 17:34:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@869 -- # local i 00:12:29.831 17:34:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:12:29.831 17:34:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:12:29.831 17:34:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@872 -- # grep -q -w nbd1 /proc/partitions 00:12:29.831 17:34:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@873 -- # break 00:12:29.831 17:34:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:12:29.831 17:34:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:12:29.831 17:34:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@885 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:12:29.831 1+0 records in 00:12:29.831 1+0 records out 00:12:29.831 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.00054289 s, 7.5 MB/s 00:12:29.831 17:34:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:29.831 17:34:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@886 -- # size=4096 00:12:29.831 17:34:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:29.831 17:34:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:12:29.831 17:34:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@889 -- # return 0 00:12:29.831 17:34:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:12:29.831 17:34:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:12:29.831 17:34:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@731 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:12:29.831 17:34:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@732 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd1 00:12:29.831 17:34:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:12:29.831 17:34:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd1') 00:12:29.831 17:34:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:12:29.831 17:34:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@51 -- # local i 00:12:29.831 17:34:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:12:29.831 17:34:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:12:30.092 17:34:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:12:30.092 17:34:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:12:30.092 17:34:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:12:30.092 17:34:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:12:30.092 17:34:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:12:30.092 17:34:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:12:30.092 17:34:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@41 -- # break 00:12:30.092 17:34:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@45 -- # return 0 00:12:30.092 17:34:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@734 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:12:30.092 17:34:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:12:30.092 17:34:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:12:30.092 17:34:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:12:30.092 17:34:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@51 -- # local i 00:12:30.092 17:34:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:12:30.092 17:34:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:12:30.092 17:34:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:12:30.092 17:34:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:12:30.092 17:34:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:12:30.092 17:34:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:12:30.092 17:34:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:12:30.092 17:34:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:12:30.352 17:34:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@41 -- # break 00:12:30.352 17:34:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@45 -- # return 0 00:12:30.352 17:34:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:12:30.352 17:34:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:12:30.352 17:34:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:30.352 17:34:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:30.352 17:34:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:30.352 17:34:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:12:30.352 17:34:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:30.352 17:34:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:30.352 [2024-11-27 17:34:01.298783] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:12:30.352 [2024-11-27 17:34:01.298842] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:30.352 [2024-11-27 17:34:01.298864] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009980 00:12:30.352 [2024-11-27 17:34:01.298873] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:30.352 [2024-11-27 17:34:01.300973] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:30.352 [2024-11-27 17:34:01.301060] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:12:30.352 [2024-11-27 17:34:01.301169] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:12:30.352 [2024-11-27 17:34:01.301206] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:30.352 [2024-11-27 17:34:01.301327] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:12:30.352 spare 00:12:30.352 17:34:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:30.352 17:34:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:12:30.352 17:34:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:30.352 17:34:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:30.352 [2024-11-27 17:34:01.401229] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001580 00:12:30.352 [2024-11-27 17:34:01.401257] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:12:30.352 [2024-11-27 17:34:01.401500] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000027720 00:12:30.352 [2024-11-27 17:34:01.401640] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001580 00:12:30.352 [2024-11-27 17:34:01.401650] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001580 00:12:30.352 [2024-11-27 17:34:01.401765] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:30.352 17:34:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:30.352 17:34:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:12:30.352 17:34:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:30.352 17:34:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:30.352 17:34:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:30.352 17:34:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:30.352 17:34:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:12:30.352 17:34:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:30.352 17:34:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:30.352 17:34:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:30.352 17:34:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:30.352 17:34:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:30.352 17:34:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:30.352 17:34:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:30.352 17:34:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:30.352 17:34:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:30.352 17:34:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:30.352 "name": "raid_bdev1", 00:12:30.352 "uuid": "e16f9deb-e8ff-4a2e-abdc-612c84f97b3b", 00:12:30.352 "strip_size_kb": 0, 00:12:30.352 "state": "online", 00:12:30.352 "raid_level": "raid1", 00:12:30.352 "superblock": true, 00:12:30.352 "num_base_bdevs": 2, 00:12:30.352 "num_base_bdevs_discovered": 2, 00:12:30.352 "num_base_bdevs_operational": 2, 00:12:30.352 "base_bdevs_list": [ 00:12:30.352 { 00:12:30.352 "name": "spare", 00:12:30.352 "uuid": "92d0d854-0b10-5ecf-b7c7-8628666a0f48", 00:12:30.352 "is_configured": true, 00:12:30.352 "data_offset": 2048, 00:12:30.352 "data_size": 63488 00:12:30.352 }, 00:12:30.352 { 00:12:30.352 "name": "BaseBdev2", 00:12:30.352 "uuid": "b461fd4c-c69e-5eb3-9eb0-b3fe61cb5d81", 00:12:30.352 "is_configured": true, 00:12:30.352 "data_offset": 2048, 00:12:30.352 "data_size": 63488 00:12:30.352 } 00:12:30.352 ] 00:12:30.352 }' 00:12:30.352 17:34:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:30.352 17:34:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:30.923 17:34:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:30.923 17:34:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:30.923 17:34:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:30.923 17:34:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:30.923 17:34:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:30.923 17:34:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:30.923 17:34:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:30.923 17:34:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:30.923 17:34:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:30.923 17:34:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:30.923 17:34:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:30.923 "name": "raid_bdev1", 00:12:30.923 "uuid": "e16f9deb-e8ff-4a2e-abdc-612c84f97b3b", 00:12:30.923 "strip_size_kb": 0, 00:12:30.923 "state": "online", 00:12:30.923 "raid_level": "raid1", 00:12:30.923 "superblock": true, 00:12:30.923 "num_base_bdevs": 2, 00:12:30.923 "num_base_bdevs_discovered": 2, 00:12:30.923 "num_base_bdevs_operational": 2, 00:12:30.923 "base_bdevs_list": [ 00:12:30.923 { 00:12:30.923 "name": "spare", 00:12:30.923 "uuid": "92d0d854-0b10-5ecf-b7c7-8628666a0f48", 00:12:30.923 "is_configured": true, 00:12:30.923 "data_offset": 2048, 00:12:30.923 "data_size": 63488 00:12:30.923 }, 00:12:30.923 { 00:12:30.923 "name": "BaseBdev2", 00:12:30.923 "uuid": "b461fd4c-c69e-5eb3-9eb0-b3fe61cb5d81", 00:12:30.923 "is_configured": true, 00:12:30.923 "data_offset": 2048, 00:12:30.923 "data_size": 63488 00:12:30.923 } 00:12:30.923 ] 00:12:30.923 }' 00:12:30.923 17:34:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:30.923 17:34:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:30.923 17:34:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:30.923 17:34:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:30.924 17:34:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:30.924 17:34:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:30.924 17:34:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:12:30.924 17:34:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:30.924 17:34:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:30.924 17:34:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:12:30.924 17:34:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:12:30.924 17:34:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:30.924 17:34:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:30.924 [2024-11-27 17:34:02.073562] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:30.924 17:34:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:30.924 17:34:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:12:30.924 17:34:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:30.924 17:34:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:30.924 17:34:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:30.924 17:34:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:30.924 17:34:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:12:30.924 17:34:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:30.924 17:34:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:30.924 17:34:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:30.924 17:34:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:30.924 17:34:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:30.924 17:34:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:30.924 17:34:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:30.924 17:34:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:30.924 17:34:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:31.184 17:34:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:31.184 "name": "raid_bdev1", 00:12:31.184 "uuid": "e16f9deb-e8ff-4a2e-abdc-612c84f97b3b", 00:12:31.184 "strip_size_kb": 0, 00:12:31.184 "state": "online", 00:12:31.184 "raid_level": "raid1", 00:12:31.184 "superblock": true, 00:12:31.184 "num_base_bdevs": 2, 00:12:31.184 "num_base_bdevs_discovered": 1, 00:12:31.184 "num_base_bdevs_operational": 1, 00:12:31.184 "base_bdevs_list": [ 00:12:31.184 { 00:12:31.184 "name": null, 00:12:31.184 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:31.184 "is_configured": false, 00:12:31.184 "data_offset": 0, 00:12:31.184 "data_size": 63488 00:12:31.184 }, 00:12:31.184 { 00:12:31.184 "name": "BaseBdev2", 00:12:31.184 "uuid": "b461fd4c-c69e-5eb3-9eb0-b3fe61cb5d81", 00:12:31.184 "is_configured": true, 00:12:31.184 "data_offset": 2048, 00:12:31.184 "data_size": 63488 00:12:31.184 } 00:12:31.184 ] 00:12:31.184 }' 00:12:31.184 17:34:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:31.184 17:34:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:31.444 17:34:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:12:31.444 17:34:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:31.444 17:34:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:31.444 [2024-11-27 17:34:02.520904] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:31.444 [2024-11-27 17:34:02.521196] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:12:31.444 [2024-11-27 17:34:02.521262] bdev_raid.c:3748:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:12:31.444 [2024-11-27 17:34:02.521331] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:31.444 [2024-11-27 17:34:02.525877] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000277f0 00:12:31.444 17:34:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:31.444 17:34:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@757 -- # sleep 1 00:12:31.444 [2024-11-27 17:34:02.527769] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:12:32.387 17:34:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:32.387 17:34:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:32.387 17:34:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:32.387 17:34:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:32.387 17:34:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:32.387 17:34:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:32.387 17:34:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:32.387 17:34:03 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:32.387 17:34:03 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:32.387 17:34:03 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:32.648 17:34:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:32.648 "name": "raid_bdev1", 00:12:32.648 "uuid": "e16f9deb-e8ff-4a2e-abdc-612c84f97b3b", 00:12:32.648 "strip_size_kb": 0, 00:12:32.648 "state": "online", 00:12:32.648 "raid_level": "raid1", 00:12:32.648 "superblock": true, 00:12:32.648 "num_base_bdevs": 2, 00:12:32.648 "num_base_bdevs_discovered": 2, 00:12:32.648 "num_base_bdevs_operational": 2, 00:12:32.648 "process": { 00:12:32.648 "type": "rebuild", 00:12:32.648 "target": "spare", 00:12:32.648 "progress": { 00:12:32.648 "blocks": 20480, 00:12:32.648 "percent": 32 00:12:32.648 } 00:12:32.648 }, 00:12:32.648 "base_bdevs_list": [ 00:12:32.648 { 00:12:32.648 "name": "spare", 00:12:32.648 "uuid": "92d0d854-0b10-5ecf-b7c7-8628666a0f48", 00:12:32.648 "is_configured": true, 00:12:32.648 "data_offset": 2048, 00:12:32.648 "data_size": 63488 00:12:32.648 }, 00:12:32.648 { 00:12:32.648 "name": "BaseBdev2", 00:12:32.648 "uuid": "b461fd4c-c69e-5eb3-9eb0-b3fe61cb5d81", 00:12:32.648 "is_configured": true, 00:12:32.648 "data_offset": 2048, 00:12:32.648 "data_size": 63488 00:12:32.648 } 00:12:32.648 ] 00:12:32.648 }' 00:12:32.648 17:34:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:32.648 17:34:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:32.648 17:34:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:32.648 17:34:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:32.648 17:34:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:12:32.648 17:34:03 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:32.648 17:34:03 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:32.648 [2024-11-27 17:34:03.692358] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:32.648 [2024-11-27 17:34:03.732176] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:12:32.648 [2024-11-27 17:34:03.732284] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:32.648 [2024-11-27 17:34:03.732314] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:32.648 [2024-11-27 17:34:03.732335] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:12:32.648 17:34:03 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:32.648 17:34:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:12:32.648 17:34:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:32.648 17:34:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:32.648 17:34:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:32.648 17:34:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:32.648 17:34:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:12:32.648 17:34:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:32.648 17:34:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:32.648 17:34:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:32.648 17:34:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:32.648 17:34:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:32.648 17:34:03 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:32.648 17:34:03 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:32.648 17:34:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:32.648 17:34:03 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:32.648 17:34:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:32.648 "name": "raid_bdev1", 00:12:32.648 "uuid": "e16f9deb-e8ff-4a2e-abdc-612c84f97b3b", 00:12:32.648 "strip_size_kb": 0, 00:12:32.648 "state": "online", 00:12:32.648 "raid_level": "raid1", 00:12:32.648 "superblock": true, 00:12:32.648 "num_base_bdevs": 2, 00:12:32.648 "num_base_bdevs_discovered": 1, 00:12:32.648 "num_base_bdevs_operational": 1, 00:12:32.648 "base_bdevs_list": [ 00:12:32.648 { 00:12:32.648 "name": null, 00:12:32.648 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:32.648 "is_configured": false, 00:12:32.648 "data_offset": 0, 00:12:32.648 "data_size": 63488 00:12:32.648 }, 00:12:32.648 { 00:12:32.648 "name": "BaseBdev2", 00:12:32.648 "uuid": "b461fd4c-c69e-5eb3-9eb0-b3fe61cb5d81", 00:12:32.648 "is_configured": true, 00:12:32.648 "data_offset": 2048, 00:12:32.648 "data_size": 63488 00:12:32.648 } 00:12:32.648 ] 00:12:32.648 }' 00:12:32.648 17:34:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:32.648 17:34:03 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:33.217 17:34:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:12:33.217 17:34:04 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:33.217 17:34:04 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:33.217 [2024-11-27 17:34:04.220159] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:12:33.217 [2024-11-27 17:34:04.220289] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:33.217 [2024-11-27 17:34:04.220331] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:12:33.217 [2024-11-27 17:34:04.220364] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:33.217 [2024-11-27 17:34:04.220816] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:33.217 [2024-11-27 17:34:04.220880] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:12:33.217 [2024-11-27 17:34:04.220994] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:12:33.217 [2024-11-27 17:34:04.221036] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:12:33.217 [2024-11-27 17:34:04.221075] bdev_raid.c:3748:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:12:33.217 [2024-11-27 17:34:04.221165] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:33.217 [2024-11-27 17:34:04.225809] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000278c0 00:12:33.217 spare 00:12:33.217 17:34:04 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:33.217 [2024-11-27 17:34:04.227768] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:12:33.217 17:34:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@764 -- # sleep 1 00:12:34.159 17:34:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:34.159 17:34:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:34.159 17:34:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:34.159 17:34:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:34.159 17:34:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:34.159 17:34:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:34.159 17:34:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:34.159 17:34:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:34.159 17:34:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:34.159 17:34:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:34.159 17:34:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:34.159 "name": "raid_bdev1", 00:12:34.159 "uuid": "e16f9deb-e8ff-4a2e-abdc-612c84f97b3b", 00:12:34.159 "strip_size_kb": 0, 00:12:34.159 "state": "online", 00:12:34.159 "raid_level": "raid1", 00:12:34.159 "superblock": true, 00:12:34.159 "num_base_bdevs": 2, 00:12:34.159 "num_base_bdevs_discovered": 2, 00:12:34.159 "num_base_bdevs_operational": 2, 00:12:34.159 "process": { 00:12:34.159 "type": "rebuild", 00:12:34.159 "target": "spare", 00:12:34.159 "progress": { 00:12:34.159 "blocks": 20480, 00:12:34.159 "percent": 32 00:12:34.159 } 00:12:34.159 }, 00:12:34.159 "base_bdevs_list": [ 00:12:34.159 { 00:12:34.159 "name": "spare", 00:12:34.159 "uuid": "92d0d854-0b10-5ecf-b7c7-8628666a0f48", 00:12:34.159 "is_configured": true, 00:12:34.159 "data_offset": 2048, 00:12:34.159 "data_size": 63488 00:12:34.159 }, 00:12:34.159 { 00:12:34.159 "name": "BaseBdev2", 00:12:34.159 "uuid": "b461fd4c-c69e-5eb3-9eb0-b3fe61cb5d81", 00:12:34.159 "is_configured": true, 00:12:34.159 "data_offset": 2048, 00:12:34.159 "data_size": 63488 00:12:34.159 } 00:12:34.159 ] 00:12:34.159 }' 00:12:34.159 17:34:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:34.159 17:34:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:34.159 17:34:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:34.420 17:34:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:34.420 17:34:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:12:34.420 17:34:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:34.420 17:34:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:34.420 [2024-11-27 17:34:05.396336] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:34.420 [2024-11-27 17:34:05.432196] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:12:34.420 [2024-11-27 17:34:05.432295] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:34.420 [2024-11-27 17:34:05.432312] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:34.420 [2024-11-27 17:34:05.432319] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:12:34.420 17:34:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:34.420 17:34:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:12:34.420 17:34:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:34.420 17:34:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:34.420 17:34:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:34.420 17:34:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:34.420 17:34:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:12:34.420 17:34:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:34.420 17:34:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:34.420 17:34:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:34.420 17:34:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:34.420 17:34:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:34.420 17:34:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:34.420 17:34:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:34.420 17:34:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:34.420 17:34:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:34.420 17:34:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:34.420 "name": "raid_bdev1", 00:12:34.420 "uuid": "e16f9deb-e8ff-4a2e-abdc-612c84f97b3b", 00:12:34.420 "strip_size_kb": 0, 00:12:34.420 "state": "online", 00:12:34.420 "raid_level": "raid1", 00:12:34.420 "superblock": true, 00:12:34.420 "num_base_bdevs": 2, 00:12:34.420 "num_base_bdevs_discovered": 1, 00:12:34.420 "num_base_bdevs_operational": 1, 00:12:34.420 "base_bdevs_list": [ 00:12:34.420 { 00:12:34.420 "name": null, 00:12:34.420 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:34.420 "is_configured": false, 00:12:34.420 "data_offset": 0, 00:12:34.420 "data_size": 63488 00:12:34.420 }, 00:12:34.420 { 00:12:34.420 "name": "BaseBdev2", 00:12:34.420 "uuid": "b461fd4c-c69e-5eb3-9eb0-b3fe61cb5d81", 00:12:34.420 "is_configured": true, 00:12:34.420 "data_offset": 2048, 00:12:34.420 "data_size": 63488 00:12:34.420 } 00:12:34.420 ] 00:12:34.420 }' 00:12:34.420 17:34:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:34.420 17:34:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:34.989 17:34:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:34.989 17:34:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:34.989 17:34:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:34.989 17:34:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:34.989 17:34:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:34.990 17:34:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:34.990 17:34:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:34.990 17:34:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:34.990 17:34:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:34.990 17:34:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:34.990 17:34:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:34.990 "name": "raid_bdev1", 00:12:34.990 "uuid": "e16f9deb-e8ff-4a2e-abdc-612c84f97b3b", 00:12:34.990 "strip_size_kb": 0, 00:12:34.990 "state": "online", 00:12:34.990 "raid_level": "raid1", 00:12:34.990 "superblock": true, 00:12:34.990 "num_base_bdevs": 2, 00:12:34.990 "num_base_bdevs_discovered": 1, 00:12:34.990 "num_base_bdevs_operational": 1, 00:12:34.990 "base_bdevs_list": [ 00:12:34.990 { 00:12:34.990 "name": null, 00:12:34.990 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:34.990 "is_configured": false, 00:12:34.990 "data_offset": 0, 00:12:34.990 "data_size": 63488 00:12:34.990 }, 00:12:34.990 { 00:12:34.990 "name": "BaseBdev2", 00:12:34.990 "uuid": "b461fd4c-c69e-5eb3-9eb0-b3fe61cb5d81", 00:12:34.990 "is_configured": true, 00:12:34.990 "data_offset": 2048, 00:12:34.990 "data_size": 63488 00:12:34.990 } 00:12:34.990 ] 00:12:34.990 }' 00:12:34.990 17:34:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:34.990 17:34:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:34.990 17:34:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:34.990 17:34:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:34.990 17:34:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:12:34.990 17:34:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:34.990 17:34:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:34.990 17:34:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:34.990 17:34:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:12:34.990 17:34:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:34.990 17:34:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:34.990 [2024-11-27 17:34:06.047887] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:12:34.990 [2024-11-27 17:34:06.047944] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:34.990 [2024-11-27 17:34:06.047966] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a880 00:12:34.990 [2024-11-27 17:34:06.047975] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:34.990 [2024-11-27 17:34:06.048403] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:34.990 [2024-11-27 17:34:06.048421] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:12:34.990 [2024-11-27 17:34:06.048494] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:12:34.990 [2024-11-27 17:34:06.048516] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:12:34.990 [2024-11-27 17:34:06.048527] bdev_raid.c:3709:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:12:34.990 [2024-11-27 17:34:06.048537] bdev_raid.c:3884:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:12:34.990 BaseBdev1 00:12:34.990 17:34:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:34.990 17:34:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@775 -- # sleep 1 00:12:35.928 17:34:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:12:35.928 17:34:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:35.928 17:34:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:35.928 17:34:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:35.928 17:34:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:35.928 17:34:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:12:35.928 17:34:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:35.928 17:34:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:35.928 17:34:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:35.928 17:34:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:35.928 17:34:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:35.928 17:34:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:35.928 17:34:07 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:35.928 17:34:07 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:35.928 17:34:07 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:35.928 17:34:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:35.928 "name": "raid_bdev1", 00:12:35.928 "uuid": "e16f9deb-e8ff-4a2e-abdc-612c84f97b3b", 00:12:35.928 "strip_size_kb": 0, 00:12:35.928 "state": "online", 00:12:35.928 "raid_level": "raid1", 00:12:35.928 "superblock": true, 00:12:35.928 "num_base_bdevs": 2, 00:12:35.928 "num_base_bdevs_discovered": 1, 00:12:35.928 "num_base_bdevs_operational": 1, 00:12:35.928 "base_bdevs_list": [ 00:12:35.928 { 00:12:35.928 "name": null, 00:12:35.928 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:35.928 "is_configured": false, 00:12:35.928 "data_offset": 0, 00:12:35.928 "data_size": 63488 00:12:35.928 }, 00:12:35.928 { 00:12:35.928 "name": "BaseBdev2", 00:12:35.928 "uuid": "b461fd4c-c69e-5eb3-9eb0-b3fe61cb5d81", 00:12:35.928 "is_configured": true, 00:12:35.928 "data_offset": 2048, 00:12:35.928 "data_size": 63488 00:12:35.928 } 00:12:35.928 ] 00:12:35.928 }' 00:12:35.928 17:34:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:35.928 17:34:07 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:36.496 17:34:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:36.496 17:34:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:36.496 17:34:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:36.496 17:34:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:36.496 17:34:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:36.496 17:34:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:36.496 17:34:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:36.496 17:34:07 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:36.496 17:34:07 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:36.496 17:34:07 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:36.496 17:34:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:36.496 "name": "raid_bdev1", 00:12:36.496 "uuid": "e16f9deb-e8ff-4a2e-abdc-612c84f97b3b", 00:12:36.496 "strip_size_kb": 0, 00:12:36.496 "state": "online", 00:12:36.496 "raid_level": "raid1", 00:12:36.496 "superblock": true, 00:12:36.496 "num_base_bdevs": 2, 00:12:36.496 "num_base_bdevs_discovered": 1, 00:12:36.496 "num_base_bdevs_operational": 1, 00:12:36.496 "base_bdevs_list": [ 00:12:36.496 { 00:12:36.496 "name": null, 00:12:36.496 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:36.496 "is_configured": false, 00:12:36.496 "data_offset": 0, 00:12:36.496 "data_size": 63488 00:12:36.496 }, 00:12:36.496 { 00:12:36.496 "name": "BaseBdev2", 00:12:36.496 "uuid": "b461fd4c-c69e-5eb3-9eb0-b3fe61cb5d81", 00:12:36.496 "is_configured": true, 00:12:36.496 "data_offset": 2048, 00:12:36.496 "data_size": 63488 00:12:36.496 } 00:12:36.496 ] 00:12:36.496 }' 00:12:36.496 17:34:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:36.496 17:34:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:36.496 17:34:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:36.496 17:34:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:36.496 17:34:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:12:36.496 17:34:07 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@650 -- # local es=0 00:12:36.496 17:34:07 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:12:36.496 17:34:07 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:12:36.496 17:34:07 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:12:36.496 17:34:07 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:12:36.755 17:34:07 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:12:36.755 17:34:07 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:12:36.755 17:34:07 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:36.755 17:34:07 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:36.755 [2024-11-27 17:34:07.693586] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:12:36.755 [2024-11-27 17:34:07.693854] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:12:36.755 [2024-11-27 17:34:07.693922] bdev_raid.c:3709:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:12:36.755 request: 00:12:36.755 { 00:12:36.755 "base_bdev": "BaseBdev1", 00:12:36.755 "raid_bdev": "raid_bdev1", 00:12:36.755 "method": "bdev_raid_add_base_bdev", 00:12:36.755 "req_id": 1 00:12:36.755 } 00:12:36.755 Got JSON-RPC error response 00:12:36.755 response: 00:12:36.755 { 00:12:36.755 "code": -22, 00:12:36.755 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:12:36.755 } 00:12:36.755 17:34:07 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:12:36.755 17:34:07 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@653 -- # es=1 00:12:36.755 17:34:07 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:12:36.755 17:34:07 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:12:36.755 17:34:07 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:12:36.755 17:34:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@779 -- # sleep 1 00:12:37.693 17:34:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:12:37.693 17:34:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:37.693 17:34:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:37.693 17:34:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:37.693 17:34:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:37.693 17:34:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:12:37.693 17:34:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:37.693 17:34:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:37.693 17:34:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:37.694 17:34:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:37.694 17:34:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:37.694 17:34:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:37.694 17:34:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:37.694 17:34:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:37.694 17:34:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:37.694 17:34:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:37.694 "name": "raid_bdev1", 00:12:37.694 "uuid": "e16f9deb-e8ff-4a2e-abdc-612c84f97b3b", 00:12:37.694 "strip_size_kb": 0, 00:12:37.694 "state": "online", 00:12:37.694 "raid_level": "raid1", 00:12:37.694 "superblock": true, 00:12:37.694 "num_base_bdevs": 2, 00:12:37.694 "num_base_bdevs_discovered": 1, 00:12:37.694 "num_base_bdevs_operational": 1, 00:12:37.694 "base_bdevs_list": [ 00:12:37.694 { 00:12:37.694 "name": null, 00:12:37.694 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:37.694 "is_configured": false, 00:12:37.694 "data_offset": 0, 00:12:37.694 "data_size": 63488 00:12:37.694 }, 00:12:37.694 { 00:12:37.694 "name": "BaseBdev2", 00:12:37.694 "uuid": "b461fd4c-c69e-5eb3-9eb0-b3fe61cb5d81", 00:12:37.694 "is_configured": true, 00:12:37.694 "data_offset": 2048, 00:12:37.694 "data_size": 63488 00:12:37.694 } 00:12:37.694 ] 00:12:37.694 }' 00:12:37.694 17:34:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:37.694 17:34:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:38.261 17:34:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:38.261 17:34:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:38.261 17:34:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:38.261 17:34:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:38.261 17:34:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:38.261 17:34:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:38.261 17:34:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:38.261 17:34:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:38.261 17:34:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:38.261 17:34:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:38.261 17:34:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:38.261 "name": "raid_bdev1", 00:12:38.261 "uuid": "e16f9deb-e8ff-4a2e-abdc-612c84f97b3b", 00:12:38.261 "strip_size_kb": 0, 00:12:38.261 "state": "online", 00:12:38.261 "raid_level": "raid1", 00:12:38.261 "superblock": true, 00:12:38.261 "num_base_bdevs": 2, 00:12:38.261 "num_base_bdevs_discovered": 1, 00:12:38.261 "num_base_bdevs_operational": 1, 00:12:38.261 "base_bdevs_list": [ 00:12:38.261 { 00:12:38.261 "name": null, 00:12:38.261 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:38.261 "is_configured": false, 00:12:38.261 "data_offset": 0, 00:12:38.261 "data_size": 63488 00:12:38.261 }, 00:12:38.261 { 00:12:38.261 "name": "BaseBdev2", 00:12:38.261 "uuid": "b461fd4c-c69e-5eb3-9eb0-b3fe61cb5d81", 00:12:38.261 "is_configured": true, 00:12:38.261 "data_offset": 2048, 00:12:38.261 "data_size": 63488 00:12:38.261 } 00:12:38.261 ] 00:12:38.261 }' 00:12:38.261 17:34:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:38.261 17:34:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:38.261 17:34:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:38.261 17:34:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:38.261 17:34:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@784 -- # killprocess 87372 00:12:38.261 17:34:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@950 -- # '[' -z 87372 ']' 00:12:38.261 17:34:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@954 -- # kill -0 87372 00:12:38.261 17:34:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@955 -- # uname 00:12:38.261 17:34:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:12:38.261 17:34:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 87372 00:12:38.261 17:34:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:12:38.261 17:34:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:12:38.262 17:34:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@968 -- # echo 'killing process with pid 87372' 00:12:38.262 killing process with pid 87372 00:12:38.262 Received shutdown signal, test time was about 16.674486 seconds 00:12:38.262 00:12:38.262 Latency(us) 00:12:38.262 [2024-11-27T17:34:09.454Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:12:38.262 [2024-11-27T17:34:09.454Z] =================================================================================================================== 00:12:38.262 [2024-11-27T17:34:09.454Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:12:38.262 17:34:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@969 -- # kill 87372 00:12:38.262 [2024-11-27 17:34:09.376536] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:12:38.262 [2024-11-27 17:34:09.376695] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:12:38.262 17:34:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@974 -- # wait 87372 00:12:38.262 [2024-11-27 17:34:09.376755] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:12:38.262 [2024-11-27 17:34:09.376767] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state offline 00:12:38.262 [2024-11-27 17:34:09.425988] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:12:38.832 17:34:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@786 -- # return 0 00:12:38.832 00:12:38.832 real 0m18.711s 00:12:38.832 user 0m24.715s 00:12:38.832 sys 0m2.338s 00:12:38.832 ************************************ 00:12:38.832 END TEST raid_rebuild_test_sb_io 00:12:38.832 ************************************ 00:12:38.832 17:34:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@1126 -- # xtrace_disable 00:12:38.832 17:34:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:12:38.832 17:34:09 bdev_raid -- bdev/bdev_raid.sh@977 -- # for n in 2 4 00:12:38.832 17:34:09 bdev_raid -- bdev/bdev_raid.sh@978 -- # run_test raid_rebuild_test raid_rebuild_test raid1 4 false false true 00:12:38.832 17:34:09 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 7 -le 1 ']' 00:12:38.832 17:34:09 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:12:38.832 17:34:09 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:12:38.832 ************************************ 00:12:38.832 START TEST raid_rebuild_test 00:12:38.832 ************************************ 00:12:38.832 17:34:09 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@1125 -- # raid_rebuild_test raid1 4 false false true 00:12:38.832 17:34:09 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:12:38.832 17:34:09 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=4 00:12:38.832 17:34:09 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@571 -- # local superblock=false 00:12:38.832 17:34:09 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:12:38.832 17:34:09 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@573 -- # local verify=true 00:12:38.832 17:34:09 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:12:38.832 17:34:09 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:38.832 17:34:09 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:12:38.832 17:34:09 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:38.832 17:34:09 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:38.832 17:34:09 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:12:38.832 17:34:09 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:38.832 17:34:09 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:38.832 17:34:09 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev3 00:12:38.832 17:34:09 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:38.832 17:34:09 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:38.832 17:34:09 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev4 00:12:38.832 17:34:09 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:38.832 17:34:09 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:38.833 17:34:09 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:12:38.833 17:34:09 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:12:38.833 17:34:09 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:12:38.833 17:34:09 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@576 -- # local strip_size 00:12:38.833 17:34:09 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@577 -- # local create_arg 00:12:38.833 17:34:09 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:12:38.833 17:34:09 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@579 -- # local data_offset 00:12:38.833 17:34:09 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:12:38.833 17:34:09 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:12:38.833 17:34:09 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@592 -- # '[' false = true ']' 00:12:38.833 17:34:09 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@597 -- # raid_pid=88044 00:12:38.833 17:34:09 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:12:38.833 17:34:09 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@598 -- # waitforlisten 88044 00:12:38.833 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:12:38.833 17:34:09 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@831 -- # '[' -z 88044 ']' 00:12:38.833 17:34:09 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:12:38.833 17:34:09 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:12:38.833 17:34:09 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:12:38.833 17:34:09 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:12:38.833 17:34:09 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:38.833 [2024-11-27 17:34:09.984782] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:12:38.833 I/O size of 3145728 is greater than zero copy threshold (65536). 00:12:38.833 Zero copy mechanism will not be used. 00:12:38.833 [2024-11-27 17:34:09.984982] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid88044 ] 00:12:39.093 [2024-11-27 17:34:10.131740] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:12:39.093 [2024-11-27 17:34:10.203736] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:12:39.093 [2024-11-27 17:34:10.281847] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:12:39.093 [2024-11-27 17:34:10.281885] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:12:39.663 17:34:10 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:12:39.663 17:34:10 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@864 -- # return 0 00:12:39.663 17:34:10 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:39.663 17:34:10 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:12:39.663 17:34:10 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:39.663 17:34:10 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:39.663 BaseBdev1_malloc 00:12:39.663 17:34:10 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:39.663 17:34:10 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:12:39.663 17:34:10 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:39.663 17:34:10 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:39.663 [2024-11-27 17:34:10.821805] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:12:39.663 [2024-11-27 17:34:10.821875] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:39.663 [2024-11-27 17:34:10.821903] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:12:39.663 [2024-11-27 17:34:10.821918] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:39.663 [2024-11-27 17:34:10.824438] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:39.663 [2024-11-27 17:34:10.824477] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:12:39.663 BaseBdev1 00:12:39.663 17:34:10 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:39.663 17:34:10 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:39.663 17:34:10 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:12:39.663 17:34:10 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:39.663 17:34:10 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:39.924 BaseBdev2_malloc 00:12:39.924 17:34:10 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:39.924 17:34:10 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:12:39.924 17:34:10 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:39.924 17:34:10 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:39.924 [2024-11-27 17:34:10.866640] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:12:39.924 [2024-11-27 17:34:10.866690] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:39.924 [2024-11-27 17:34:10.866713] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:12:39.924 [2024-11-27 17:34:10.866721] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:39.924 [2024-11-27 17:34:10.869220] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:39.924 [2024-11-27 17:34:10.869249] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:12:39.924 BaseBdev2 00:12:39.924 17:34:10 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:39.924 17:34:10 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:39.924 17:34:10 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:12:39.924 17:34:10 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:39.924 17:34:10 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:39.924 BaseBdev3_malloc 00:12:39.924 17:34:10 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:39.924 17:34:10 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev3_malloc -p BaseBdev3 00:12:39.924 17:34:10 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:39.924 17:34:10 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:39.924 [2024-11-27 17:34:10.901734] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev3_malloc 00:12:39.924 [2024-11-27 17:34:10.901848] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:39.924 [2024-11-27 17:34:10.901885] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:12:39.924 [2024-11-27 17:34:10.901894] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:39.924 [2024-11-27 17:34:10.904388] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:39.924 [2024-11-27 17:34:10.904422] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:12:39.924 BaseBdev3 00:12:39.924 17:34:10 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:39.924 17:34:10 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:39.924 17:34:10 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:12:39.924 17:34:10 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:39.924 17:34:10 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:39.924 BaseBdev4_malloc 00:12:39.924 17:34:10 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:39.924 17:34:10 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev4_malloc -p BaseBdev4 00:12:39.924 17:34:10 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:39.924 17:34:10 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:39.924 [2024-11-27 17:34:10.936916] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev4_malloc 00:12:39.924 [2024-11-27 17:34:10.937008] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:39.924 [2024-11-27 17:34:10.937038] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:12:39.924 [2024-11-27 17:34:10.937047] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:39.924 [2024-11-27 17:34:10.939530] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:39.924 [2024-11-27 17:34:10.939566] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:12:39.924 BaseBdev4 00:12:39.924 17:34:10 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:39.924 17:34:10 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:12:39.924 17:34:10 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:39.924 17:34:10 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:39.924 spare_malloc 00:12:39.924 17:34:10 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:39.924 17:34:10 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:12:39.924 17:34:10 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:39.924 17:34:10 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:39.924 spare_delay 00:12:39.924 17:34:10 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:39.924 17:34:10 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:12:39.924 17:34:10 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:39.924 17:34:10 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:39.924 [2024-11-27 17:34:10.984059] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:12:39.924 [2024-11-27 17:34:10.984109] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:39.924 [2024-11-27 17:34:10.984132] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009c80 00:12:39.924 [2024-11-27 17:34:10.984154] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:39.924 [2024-11-27 17:34:10.986560] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:39.924 [2024-11-27 17:34:10.986647] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:12:39.924 spare 00:12:39.924 17:34:10 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:39.924 17:34:10 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 00:12:39.924 17:34:10 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:39.924 17:34:10 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:39.924 [2024-11-27 17:34:10.996113] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:12:39.924 [2024-11-27 17:34:10.998258] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:12:39.924 [2024-11-27 17:34:10.998321] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:12:39.924 [2024-11-27 17:34:10.998367] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:12:39.924 [2024-11-27 17:34:10.998440] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:12:39.924 [2024-11-27 17:34:10.998449] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:12:39.924 [2024-11-27 17:34:10.998706] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:12:39.924 [2024-11-27 17:34:10.998874] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:12:39.924 [2024-11-27 17:34:10.998890] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:12:39.924 [2024-11-27 17:34:10.999021] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:39.924 17:34:10 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:39.924 17:34:11 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:12:39.924 17:34:11 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:39.924 17:34:11 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:39.924 17:34:11 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:39.924 17:34:11 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:39.924 17:34:11 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:12:39.924 17:34:11 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:39.924 17:34:11 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:39.924 17:34:11 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:39.924 17:34:11 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:39.924 17:34:11 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:39.924 17:34:11 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:39.924 17:34:11 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:39.924 17:34:11 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:39.924 17:34:11 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:39.924 17:34:11 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:39.924 "name": "raid_bdev1", 00:12:39.924 "uuid": "87f070f4-a5ae-45c2-a45e-016942a63bbf", 00:12:39.924 "strip_size_kb": 0, 00:12:39.924 "state": "online", 00:12:39.924 "raid_level": "raid1", 00:12:39.924 "superblock": false, 00:12:39.924 "num_base_bdevs": 4, 00:12:39.924 "num_base_bdevs_discovered": 4, 00:12:39.924 "num_base_bdevs_operational": 4, 00:12:39.924 "base_bdevs_list": [ 00:12:39.924 { 00:12:39.925 "name": "BaseBdev1", 00:12:39.925 "uuid": "55cd9bf9-4f05-5230-90b1-44fcd23c93fb", 00:12:39.925 "is_configured": true, 00:12:39.925 "data_offset": 0, 00:12:39.925 "data_size": 65536 00:12:39.925 }, 00:12:39.925 { 00:12:39.925 "name": "BaseBdev2", 00:12:39.925 "uuid": "535a19f7-162a-52a2-8e4b-cf95052a2424", 00:12:39.925 "is_configured": true, 00:12:39.925 "data_offset": 0, 00:12:39.925 "data_size": 65536 00:12:39.925 }, 00:12:39.925 { 00:12:39.925 "name": "BaseBdev3", 00:12:39.925 "uuid": "f8a406fe-44e2-5afd-99f6-1fda262aaeae", 00:12:39.925 "is_configured": true, 00:12:39.925 "data_offset": 0, 00:12:39.925 "data_size": 65536 00:12:39.925 }, 00:12:39.925 { 00:12:39.925 "name": "BaseBdev4", 00:12:39.925 "uuid": "39cc8ead-6b09-5b57-8d89-5e8afe49b8a9", 00:12:39.925 "is_configured": true, 00:12:39.925 "data_offset": 0, 00:12:39.925 "data_size": 65536 00:12:39.925 } 00:12:39.925 ] 00:12:39.925 }' 00:12:39.925 17:34:11 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:39.925 17:34:11 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:40.496 17:34:11 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:12:40.496 17:34:11 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:12:40.496 17:34:11 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:40.496 17:34:11 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:40.496 [2024-11-27 17:34:11.459689] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:12:40.496 17:34:11 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:40.496 17:34:11 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=65536 00:12:40.496 17:34:11 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:40.496 17:34:11 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:40.496 17:34:11 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:40.496 17:34:11 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:12:40.496 17:34:11 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:40.496 17:34:11 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@619 -- # data_offset=0 00:12:40.496 17:34:11 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:12:40.496 17:34:11 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:12:40.496 17:34:11 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:12:40.496 17:34:11 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:12:40.496 17:34:11 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:12:40.496 17:34:11 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:12:40.496 17:34:11 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@10 -- # local bdev_list 00:12:40.496 17:34:11 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:12:40.496 17:34:11 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@11 -- # local nbd_list 00:12:40.496 17:34:11 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@12 -- # local i 00:12:40.496 17:34:11 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:12:40.496 17:34:11 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:12:40.496 17:34:11 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:12:40.766 [2024-11-27 17:34:11.722998] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000027a0 00:12:40.766 /dev/nbd0 00:12:40.766 17:34:11 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:12:40.766 17:34:11 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:12:40.766 17:34:11 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:12:40.766 17:34:11 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@869 -- # local i 00:12:40.766 17:34:11 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:12:40.766 17:34:11 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:12:40.766 17:34:11 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:12:40.766 17:34:11 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@873 -- # break 00:12:40.766 17:34:11 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:12:40.766 17:34:11 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:12:40.766 17:34:11 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:12:40.766 1+0 records in 00:12:40.766 1+0 records out 00:12:40.766 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000426726 s, 9.6 MB/s 00:12:40.766 17:34:11 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:40.766 17:34:11 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@886 -- # size=4096 00:12:40.766 17:34:11 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:40.766 17:34:11 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:12:40.766 17:34:11 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@889 -- # return 0 00:12:40.766 17:34:11 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:12:40.766 17:34:11 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:12:40.766 17:34:11 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@629 -- # '[' raid1 = raid5f ']' 00:12:40.766 17:34:11 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@633 -- # write_unit_size=1 00:12:40.766 17:34:11 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=512 count=65536 oflag=direct 00:12:46.055 65536+0 records in 00:12:46.055 65536+0 records out 00:12:46.055 33554432 bytes (34 MB, 32 MiB) copied, 5.02563 s, 6.7 MB/s 00:12:46.055 17:34:16 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:12:46.055 17:34:16 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:12:46.055 17:34:16 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:12:46.055 17:34:16 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@50 -- # local nbd_list 00:12:46.055 17:34:16 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@51 -- # local i 00:12:46.055 17:34:16 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:12:46.055 17:34:16 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:12:46.055 [2024-11-27 17:34:17.015229] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:46.055 17:34:17 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:12:46.055 17:34:17 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:12:46.055 17:34:17 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:12:46.055 17:34:17 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:12:46.055 17:34:17 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:12:46.055 17:34:17 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:12:46.055 17:34:17 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:12:46.055 17:34:17 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:12:46.055 17:34:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:12:46.055 17:34:17 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:46.055 17:34:17 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:46.055 [2024-11-27 17:34:17.047239] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:12:46.055 17:34:17 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:46.055 17:34:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:12:46.055 17:34:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:46.055 17:34:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:46.055 17:34:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:46.055 17:34:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:46.055 17:34:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:46.055 17:34:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:46.055 17:34:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:46.055 17:34:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:46.055 17:34:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:46.055 17:34:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:46.055 17:34:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:46.055 17:34:17 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:46.055 17:34:17 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:46.055 17:34:17 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:46.055 17:34:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:46.055 "name": "raid_bdev1", 00:12:46.055 "uuid": "87f070f4-a5ae-45c2-a45e-016942a63bbf", 00:12:46.055 "strip_size_kb": 0, 00:12:46.055 "state": "online", 00:12:46.055 "raid_level": "raid1", 00:12:46.055 "superblock": false, 00:12:46.055 "num_base_bdevs": 4, 00:12:46.055 "num_base_bdevs_discovered": 3, 00:12:46.055 "num_base_bdevs_operational": 3, 00:12:46.055 "base_bdevs_list": [ 00:12:46.055 { 00:12:46.055 "name": null, 00:12:46.055 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:46.055 "is_configured": false, 00:12:46.055 "data_offset": 0, 00:12:46.055 "data_size": 65536 00:12:46.055 }, 00:12:46.055 { 00:12:46.055 "name": "BaseBdev2", 00:12:46.055 "uuid": "535a19f7-162a-52a2-8e4b-cf95052a2424", 00:12:46.055 "is_configured": true, 00:12:46.055 "data_offset": 0, 00:12:46.055 "data_size": 65536 00:12:46.055 }, 00:12:46.055 { 00:12:46.056 "name": "BaseBdev3", 00:12:46.056 "uuid": "f8a406fe-44e2-5afd-99f6-1fda262aaeae", 00:12:46.056 "is_configured": true, 00:12:46.056 "data_offset": 0, 00:12:46.056 "data_size": 65536 00:12:46.056 }, 00:12:46.056 { 00:12:46.056 "name": "BaseBdev4", 00:12:46.056 "uuid": "39cc8ead-6b09-5b57-8d89-5e8afe49b8a9", 00:12:46.056 "is_configured": true, 00:12:46.056 "data_offset": 0, 00:12:46.056 "data_size": 65536 00:12:46.056 } 00:12:46.056 ] 00:12:46.056 }' 00:12:46.056 17:34:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:46.056 17:34:17 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:46.314 17:34:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:12:46.314 17:34:17 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:46.314 17:34:17 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:46.314 [2024-11-27 17:34:17.486473] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:46.314 [2024-11-27 17:34:17.489914] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000d063c0 00:12:46.314 [2024-11-27 17:34:17.491819] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:12:46.314 17:34:17 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:46.314 17:34:17 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@647 -- # sleep 1 00:12:47.695 17:34:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:47.695 17:34:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:47.695 17:34:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:47.695 17:34:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:47.695 17:34:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:47.695 17:34:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:47.695 17:34:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:47.695 17:34:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:47.695 17:34:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:47.695 17:34:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:47.695 17:34:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:47.695 "name": "raid_bdev1", 00:12:47.695 "uuid": "87f070f4-a5ae-45c2-a45e-016942a63bbf", 00:12:47.695 "strip_size_kb": 0, 00:12:47.695 "state": "online", 00:12:47.695 "raid_level": "raid1", 00:12:47.695 "superblock": false, 00:12:47.695 "num_base_bdevs": 4, 00:12:47.695 "num_base_bdevs_discovered": 4, 00:12:47.695 "num_base_bdevs_operational": 4, 00:12:47.695 "process": { 00:12:47.695 "type": "rebuild", 00:12:47.695 "target": "spare", 00:12:47.695 "progress": { 00:12:47.695 "blocks": 20480, 00:12:47.695 "percent": 31 00:12:47.695 } 00:12:47.695 }, 00:12:47.695 "base_bdevs_list": [ 00:12:47.695 { 00:12:47.695 "name": "spare", 00:12:47.695 "uuid": "bb70b8e8-6ab4-50f9-9c09-8477b339b3b8", 00:12:47.695 "is_configured": true, 00:12:47.695 "data_offset": 0, 00:12:47.695 "data_size": 65536 00:12:47.695 }, 00:12:47.695 { 00:12:47.695 "name": "BaseBdev2", 00:12:47.695 "uuid": "535a19f7-162a-52a2-8e4b-cf95052a2424", 00:12:47.695 "is_configured": true, 00:12:47.695 "data_offset": 0, 00:12:47.695 "data_size": 65536 00:12:47.695 }, 00:12:47.695 { 00:12:47.695 "name": "BaseBdev3", 00:12:47.695 "uuid": "f8a406fe-44e2-5afd-99f6-1fda262aaeae", 00:12:47.695 "is_configured": true, 00:12:47.695 "data_offset": 0, 00:12:47.695 "data_size": 65536 00:12:47.695 }, 00:12:47.695 { 00:12:47.695 "name": "BaseBdev4", 00:12:47.695 "uuid": "39cc8ead-6b09-5b57-8d89-5e8afe49b8a9", 00:12:47.695 "is_configured": true, 00:12:47.695 "data_offset": 0, 00:12:47.695 "data_size": 65536 00:12:47.695 } 00:12:47.695 ] 00:12:47.695 }' 00:12:47.695 17:34:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:47.695 17:34:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:47.695 17:34:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:47.695 17:34:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:47.695 17:34:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:12:47.696 17:34:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:47.696 17:34:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:47.696 [2024-11-27 17:34:18.646943] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:47.696 [2024-11-27 17:34:18.696262] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:12:47.696 [2024-11-27 17:34:18.696319] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:47.696 [2024-11-27 17:34:18.696337] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:12:47.696 [2024-11-27 17:34:18.696344] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:12:47.696 17:34:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:47.696 17:34:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:12:47.696 17:34:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:47.696 17:34:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:47.696 17:34:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:47.696 17:34:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:47.696 17:34:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:47.696 17:34:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:47.696 17:34:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:47.696 17:34:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:47.696 17:34:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:47.696 17:34:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:47.696 17:34:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:47.696 17:34:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:47.696 17:34:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:47.696 17:34:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:47.696 17:34:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:47.696 "name": "raid_bdev1", 00:12:47.696 "uuid": "87f070f4-a5ae-45c2-a45e-016942a63bbf", 00:12:47.696 "strip_size_kb": 0, 00:12:47.696 "state": "online", 00:12:47.696 "raid_level": "raid1", 00:12:47.696 "superblock": false, 00:12:47.696 "num_base_bdevs": 4, 00:12:47.696 "num_base_bdevs_discovered": 3, 00:12:47.696 "num_base_bdevs_operational": 3, 00:12:47.696 "base_bdevs_list": [ 00:12:47.696 { 00:12:47.696 "name": null, 00:12:47.696 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:47.696 "is_configured": false, 00:12:47.696 "data_offset": 0, 00:12:47.696 "data_size": 65536 00:12:47.696 }, 00:12:47.696 { 00:12:47.696 "name": "BaseBdev2", 00:12:47.696 "uuid": "535a19f7-162a-52a2-8e4b-cf95052a2424", 00:12:47.696 "is_configured": true, 00:12:47.696 "data_offset": 0, 00:12:47.696 "data_size": 65536 00:12:47.696 }, 00:12:47.696 { 00:12:47.696 "name": "BaseBdev3", 00:12:47.696 "uuid": "f8a406fe-44e2-5afd-99f6-1fda262aaeae", 00:12:47.696 "is_configured": true, 00:12:47.696 "data_offset": 0, 00:12:47.696 "data_size": 65536 00:12:47.696 }, 00:12:47.696 { 00:12:47.696 "name": "BaseBdev4", 00:12:47.696 "uuid": "39cc8ead-6b09-5b57-8d89-5e8afe49b8a9", 00:12:47.696 "is_configured": true, 00:12:47.696 "data_offset": 0, 00:12:47.696 "data_size": 65536 00:12:47.696 } 00:12:47.696 ] 00:12:47.696 }' 00:12:47.696 17:34:18 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:47.696 17:34:18 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:47.958 17:34:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:47.958 17:34:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:47.958 17:34:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:47.958 17:34:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:47.958 17:34:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:47.958 17:34:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:47.958 17:34:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:47.958 17:34:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:47.958 17:34:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:47.958 17:34:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:48.218 17:34:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:48.218 "name": "raid_bdev1", 00:12:48.218 "uuid": "87f070f4-a5ae-45c2-a45e-016942a63bbf", 00:12:48.218 "strip_size_kb": 0, 00:12:48.218 "state": "online", 00:12:48.218 "raid_level": "raid1", 00:12:48.218 "superblock": false, 00:12:48.218 "num_base_bdevs": 4, 00:12:48.218 "num_base_bdevs_discovered": 3, 00:12:48.218 "num_base_bdevs_operational": 3, 00:12:48.218 "base_bdevs_list": [ 00:12:48.218 { 00:12:48.218 "name": null, 00:12:48.218 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:48.218 "is_configured": false, 00:12:48.218 "data_offset": 0, 00:12:48.218 "data_size": 65536 00:12:48.218 }, 00:12:48.218 { 00:12:48.218 "name": "BaseBdev2", 00:12:48.218 "uuid": "535a19f7-162a-52a2-8e4b-cf95052a2424", 00:12:48.218 "is_configured": true, 00:12:48.218 "data_offset": 0, 00:12:48.218 "data_size": 65536 00:12:48.218 }, 00:12:48.218 { 00:12:48.218 "name": "BaseBdev3", 00:12:48.218 "uuid": "f8a406fe-44e2-5afd-99f6-1fda262aaeae", 00:12:48.218 "is_configured": true, 00:12:48.218 "data_offset": 0, 00:12:48.218 "data_size": 65536 00:12:48.218 }, 00:12:48.218 { 00:12:48.218 "name": "BaseBdev4", 00:12:48.218 "uuid": "39cc8ead-6b09-5b57-8d89-5e8afe49b8a9", 00:12:48.218 "is_configured": true, 00:12:48.218 "data_offset": 0, 00:12:48.218 "data_size": 65536 00:12:48.218 } 00:12:48.218 ] 00:12:48.218 }' 00:12:48.218 17:34:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:48.218 17:34:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:48.218 17:34:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:48.218 17:34:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:48.218 17:34:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:12:48.218 17:34:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:48.218 17:34:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:48.218 [2024-11-27 17:34:19.271082] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:12:48.218 [2024-11-27 17:34:19.273883] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000d06490 00:12:48.218 [2024-11-27 17:34:19.275740] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:12:48.218 17:34:19 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:48.218 17:34:19 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@663 -- # sleep 1 00:12:49.159 17:34:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:49.159 17:34:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:49.159 17:34:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:49.159 17:34:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:49.159 17:34:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:49.159 17:34:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:49.159 17:34:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:49.159 17:34:20 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:49.159 17:34:20 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:49.159 17:34:20 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:49.159 17:34:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:49.159 "name": "raid_bdev1", 00:12:49.159 "uuid": "87f070f4-a5ae-45c2-a45e-016942a63bbf", 00:12:49.159 "strip_size_kb": 0, 00:12:49.159 "state": "online", 00:12:49.159 "raid_level": "raid1", 00:12:49.159 "superblock": false, 00:12:49.159 "num_base_bdevs": 4, 00:12:49.159 "num_base_bdevs_discovered": 4, 00:12:49.159 "num_base_bdevs_operational": 4, 00:12:49.159 "process": { 00:12:49.159 "type": "rebuild", 00:12:49.159 "target": "spare", 00:12:49.159 "progress": { 00:12:49.159 "blocks": 20480, 00:12:49.159 "percent": 31 00:12:49.159 } 00:12:49.159 }, 00:12:49.159 "base_bdevs_list": [ 00:12:49.159 { 00:12:49.159 "name": "spare", 00:12:49.159 "uuid": "bb70b8e8-6ab4-50f9-9c09-8477b339b3b8", 00:12:49.159 "is_configured": true, 00:12:49.159 "data_offset": 0, 00:12:49.159 "data_size": 65536 00:12:49.159 }, 00:12:49.159 { 00:12:49.159 "name": "BaseBdev2", 00:12:49.159 "uuid": "535a19f7-162a-52a2-8e4b-cf95052a2424", 00:12:49.159 "is_configured": true, 00:12:49.159 "data_offset": 0, 00:12:49.159 "data_size": 65536 00:12:49.159 }, 00:12:49.159 { 00:12:49.159 "name": "BaseBdev3", 00:12:49.159 "uuid": "f8a406fe-44e2-5afd-99f6-1fda262aaeae", 00:12:49.159 "is_configured": true, 00:12:49.159 "data_offset": 0, 00:12:49.159 "data_size": 65536 00:12:49.159 }, 00:12:49.159 { 00:12:49.159 "name": "BaseBdev4", 00:12:49.159 "uuid": "39cc8ead-6b09-5b57-8d89-5e8afe49b8a9", 00:12:49.159 "is_configured": true, 00:12:49.159 "data_offset": 0, 00:12:49.159 "data_size": 65536 00:12:49.159 } 00:12:49.159 ] 00:12:49.159 }' 00:12:49.159 17:34:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:49.419 17:34:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:49.419 17:34:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:49.419 17:34:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:49.419 17:34:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@666 -- # '[' false = true ']' 00:12:49.419 17:34:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=4 00:12:49.419 17:34:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:12:49.419 17:34:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@693 -- # '[' 4 -gt 2 ']' 00:12:49.419 17:34:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@695 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:12:49.419 17:34:20 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:49.419 17:34:20 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:49.419 [2024-11-27 17:34:20.422374] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:12:49.419 [2024-11-27 17:34:20.479605] bdev_raid.c:1970:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 1 raid_ch: 0x60d000d06490 00:12:49.419 17:34:20 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:49.419 17:34:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@698 -- # base_bdevs[1]= 00:12:49.419 17:34:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@699 -- # (( num_base_bdevs_operational-- )) 00:12:49.419 17:34:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@702 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:49.419 17:34:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:49.419 17:34:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:49.419 17:34:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:49.419 17:34:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:49.419 17:34:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:49.419 17:34:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:49.419 17:34:20 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:49.419 17:34:20 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:49.419 17:34:20 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:49.419 17:34:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:49.419 "name": "raid_bdev1", 00:12:49.419 "uuid": "87f070f4-a5ae-45c2-a45e-016942a63bbf", 00:12:49.419 "strip_size_kb": 0, 00:12:49.419 "state": "online", 00:12:49.419 "raid_level": "raid1", 00:12:49.419 "superblock": false, 00:12:49.419 "num_base_bdevs": 4, 00:12:49.419 "num_base_bdevs_discovered": 3, 00:12:49.419 "num_base_bdevs_operational": 3, 00:12:49.419 "process": { 00:12:49.419 "type": "rebuild", 00:12:49.419 "target": "spare", 00:12:49.419 "progress": { 00:12:49.419 "blocks": 24576, 00:12:49.419 "percent": 37 00:12:49.419 } 00:12:49.419 }, 00:12:49.419 "base_bdevs_list": [ 00:12:49.419 { 00:12:49.419 "name": "spare", 00:12:49.419 "uuid": "bb70b8e8-6ab4-50f9-9c09-8477b339b3b8", 00:12:49.419 "is_configured": true, 00:12:49.419 "data_offset": 0, 00:12:49.419 "data_size": 65536 00:12:49.419 }, 00:12:49.419 { 00:12:49.419 "name": null, 00:12:49.419 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:49.419 "is_configured": false, 00:12:49.419 "data_offset": 0, 00:12:49.419 "data_size": 65536 00:12:49.419 }, 00:12:49.419 { 00:12:49.419 "name": "BaseBdev3", 00:12:49.419 "uuid": "f8a406fe-44e2-5afd-99f6-1fda262aaeae", 00:12:49.419 "is_configured": true, 00:12:49.419 "data_offset": 0, 00:12:49.419 "data_size": 65536 00:12:49.419 }, 00:12:49.419 { 00:12:49.419 "name": "BaseBdev4", 00:12:49.419 "uuid": "39cc8ead-6b09-5b57-8d89-5e8afe49b8a9", 00:12:49.419 "is_configured": true, 00:12:49.419 "data_offset": 0, 00:12:49.419 "data_size": 65536 00:12:49.419 } 00:12:49.419 ] 00:12:49.419 }' 00:12:49.419 17:34:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:49.419 17:34:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:49.419 17:34:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:49.678 17:34:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:49.678 17:34:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@706 -- # local timeout=368 00:12:49.678 17:34:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:12:49.678 17:34:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:49.678 17:34:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:49.678 17:34:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:49.678 17:34:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:49.678 17:34:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:49.678 17:34:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:49.678 17:34:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:49.678 17:34:20 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:49.678 17:34:20 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:49.678 17:34:20 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:49.678 17:34:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:49.678 "name": "raid_bdev1", 00:12:49.678 "uuid": "87f070f4-a5ae-45c2-a45e-016942a63bbf", 00:12:49.678 "strip_size_kb": 0, 00:12:49.678 "state": "online", 00:12:49.678 "raid_level": "raid1", 00:12:49.678 "superblock": false, 00:12:49.678 "num_base_bdevs": 4, 00:12:49.678 "num_base_bdevs_discovered": 3, 00:12:49.678 "num_base_bdevs_operational": 3, 00:12:49.678 "process": { 00:12:49.678 "type": "rebuild", 00:12:49.678 "target": "spare", 00:12:49.678 "progress": { 00:12:49.678 "blocks": 26624, 00:12:49.678 "percent": 40 00:12:49.678 } 00:12:49.678 }, 00:12:49.678 "base_bdevs_list": [ 00:12:49.678 { 00:12:49.678 "name": "spare", 00:12:49.678 "uuid": "bb70b8e8-6ab4-50f9-9c09-8477b339b3b8", 00:12:49.678 "is_configured": true, 00:12:49.678 "data_offset": 0, 00:12:49.678 "data_size": 65536 00:12:49.678 }, 00:12:49.678 { 00:12:49.678 "name": null, 00:12:49.678 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:49.678 "is_configured": false, 00:12:49.678 "data_offset": 0, 00:12:49.678 "data_size": 65536 00:12:49.678 }, 00:12:49.678 { 00:12:49.678 "name": "BaseBdev3", 00:12:49.678 "uuid": "f8a406fe-44e2-5afd-99f6-1fda262aaeae", 00:12:49.678 "is_configured": true, 00:12:49.678 "data_offset": 0, 00:12:49.678 "data_size": 65536 00:12:49.678 }, 00:12:49.678 { 00:12:49.678 "name": "BaseBdev4", 00:12:49.678 "uuid": "39cc8ead-6b09-5b57-8d89-5e8afe49b8a9", 00:12:49.678 "is_configured": true, 00:12:49.678 "data_offset": 0, 00:12:49.678 "data_size": 65536 00:12:49.678 } 00:12:49.678 ] 00:12:49.678 }' 00:12:49.678 17:34:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:49.678 17:34:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:49.678 17:34:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:49.678 17:34:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:49.678 17:34:20 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:12:50.617 17:34:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:12:50.617 17:34:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:50.617 17:34:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:50.617 17:34:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:50.617 17:34:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:50.617 17:34:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:50.617 17:34:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:50.617 17:34:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:50.617 17:34:21 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:50.617 17:34:21 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:50.617 17:34:21 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:50.878 17:34:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:50.878 "name": "raid_bdev1", 00:12:50.878 "uuid": "87f070f4-a5ae-45c2-a45e-016942a63bbf", 00:12:50.878 "strip_size_kb": 0, 00:12:50.878 "state": "online", 00:12:50.878 "raid_level": "raid1", 00:12:50.878 "superblock": false, 00:12:50.878 "num_base_bdevs": 4, 00:12:50.878 "num_base_bdevs_discovered": 3, 00:12:50.878 "num_base_bdevs_operational": 3, 00:12:50.878 "process": { 00:12:50.878 "type": "rebuild", 00:12:50.878 "target": "spare", 00:12:50.878 "progress": { 00:12:50.878 "blocks": 49152, 00:12:50.878 "percent": 75 00:12:50.878 } 00:12:50.878 }, 00:12:50.878 "base_bdevs_list": [ 00:12:50.878 { 00:12:50.878 "name": "spare", 00:12:50.878 "uuid": "bb70b8e8-6ab4-50f9-9c09-8477b339b3b8", 00:12:50.878 "is_configured": true, 00:12:50.878 "data_offset": 0, 00:12:50.878 "data_size": 65536 00:12:50.878 }, 00:12:50.878 { 00:12:50.878 "name": null, 00:12:50.878 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:50.878 "is_configured": false, 00:12:50.878 "data_offset": 0, 00:12:50.878 "data_size": 65536 00:12:50.878 }, 00:12:50.878 { 00:12:50.878 "name": "BaseBdev3", 00:12:50.878 "uuid": "f8a406fe-44e2-5afd-99f6-1fda262aaeae", 00:12:50.878 "is_configured": true, 00:12:50.878 "data_offset": 0, 00:12:50.878 "data_size": 65536 00:12:50.878 }, 00:12:50.878 { 00:12:50.878 "name": "BaseBdev4", 00:12:50.878 "uuid": "39cc8ead-6b09-5b57-8d89-5e8afe49b8a9", 00:12:50.878 "is_configured": true, 00:12:50.878 "data_offset": 0, 00:12:50.878 "data_size": 65536 00:12:50.878 } 00:12:50.878 ] 00:12:50.878 }' 00:12:50.878 17:34:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:50.878 17:34:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:12:50.878 17:34:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:50.878 17:34:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:12:50.878 17:34:21 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:12:51.448 [2024-11-27 17:34:22.486389] bdev_raid.c:2896:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:12:51.448 [2024-11-27 17:34:22.486524] bdev_raid.c:2558:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:12:51.448 [2024-11-27 17:34:22.486592] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:52.019 17:34:22 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:12:52.019 17:34:22 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:12:52.019 17:34:22 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:52.019 17:34:22 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:12:52.019 17:34:22 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:12:52.019 17:34:22 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:52.019 17:34:22 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:52.019 17:34:22 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:52.019 17:34:22 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:52.019 17:34:22 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:52.019 17:34:22 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:52.019 17:34:22 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:52.019 "name": "raid_bdev1", 00:12:52.019 "uuid": "87f070f4-a5ae-45c2-a45e-016942a63bbf", 00:12:52.019 "strip_size_kb": 0, 00:12:52.019 "state": "online", 00:12:52.019 "raid_level": "raid1", 00:12:52.019 "superblock": false, 00:12:52.019 "num_base_bdevs": 4, 00:12:52.019 "num_base_bdevs_discovered": 3, 00:12:52.019 "num_base_bdevs_operational": 3, 00:12:52.019 "base_bdevs_list": [ 00:12:52.019 { 00:12:52.019 "name": "spare", 00:12:52.019 "uuid": "bb70b8e8-6ab4-50f9-9c09-8477b339b3b8", 00:12:52.019 "is_configured": true, 00:12:52.019 "data_offset": 0, 00:12:52.019 "data_size": 65536 00:12:52.019 }, 00:12:52.019 { 00:12:52.019 "name": null, 00:12:52.019 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:52.019 "is_configured": false, 00:12:52.019 "data_offset": 0, 00:12:52.019 "data_size": 65536 00:12:52.019 }, 00:12:52.019 { 00:12:52.019 "name": "BaseBdev3", 00:12:52.019 "uuid": "f8a406fe-44e2-5afd-99f6-1fda262aaeae", 00:12:52.019 "is_configured": true, 00:12:52.019 "data_offset": 0, 00:12:52.019 "data_size": 65536 00:12:52.019 }, 00:12:52.019 { 00:12:52.019 "name": "BaseBdev4", 00:12:52.019 "uuid": "39cc8ead-6b09-5b57-8d89-5e8afe49b8a9", 00:12:52.019 "is_configured": true, 00:12:52.019 "data_offset": 0, 00:12:52.019 "data_size": 65536 00:12:52.019 } 00:12:52.019 ] 00:12:52.019 }' 00:12:52.019 17:34:22 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:52.019 17:34:23 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:12:52.019 17:34:23 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:52.019 17:34:23 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:12:52.019 17:34:23 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@709 -- # break 00:12:52.019 17:34:23 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:12:52.019 17:34:23 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:12:52.019 17:34:23 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:12:52.019 17:34:23 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=none 00:12:52.019 17:34:23 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:12:52.019 17:34:23 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:52.019 17:34:23 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:52.019 17:34:23 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:52.019 17:34:23 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:52.019 17:34:23 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:52.019 17:34:23 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:12:52.019 "name": "raid_bdev1", 00:12:52.019 "uuid": "87f070f4-a5ae-45c2-a45e-016942a63bbf", 00:12:52.019 "strip_size_kb": 0, 00:12:52.019 "state": "online", 00:12:52.019 "raid_level": "raid1", 00:12:52.019 "superblock": false, 00:12:52.019 "num_base_bdevs": 4, 00:12:52.019 "num_base_bdevs_discovered": 3, 00:12:52.019 "num_base_bdevs_operational": 3, 00:12:52.019 "base_bdevs_list": [ 00:12:52.019 { 00:12:52.019 "name": "spare", 00:12:52.020 "uuid": "bb70b8e8-6ab4-50f9-9c09-8477b339b3b8", 00:12:52.020 "is_configured": true, 00:12:52.020 "data_offset": 0, 00:12:52.020 "data_size": 65536 00:12:52.020 }, 00:12:52.020 { 00:12:52.020 "name": null, 00:12:52.020 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:52.020 "is_configured": false, 00:12:52.020 "data_offset": 0, 00:12:52.020 "data_size": 65536 00:12:52.020 }, 00:12:52.020 { 00:12:52.020 "name": "BaseBdev3", 00:12:52.020 "uuid": "f8a406fe-44e2-5afd-99f6-1fda262aaeae", 00:12:52.020 "is_configured": true, 00:12:52.020 "data_offset": 0, 00:12:52.020 "data_size": 65536 00:12:52.020 }, 00:12:52.020 { 00:12:52.020 "name": "BaseBdev4", 00:12:52.020 "uuid": "39cc8ead-6b09-5b57-8d89-5e8afe49b8a9", 00:12:52.020 "is_configured": true, 00:12:52.020 "data_offset": 0, 00:12:52.020 "data_size": 65536 00:12:52.020 } 00:12:52.020 ] 00:12:52.020 }' 00:12:52.020 17:34:23 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:12:52.020 17:34:23 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:12:52.020 17:34:23 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:12:52.279 17:34:23 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:12:52.279 17:34:23 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:12:52.279 17:34:23 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:52.279 17:34:23 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:52.279 17:34:23 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:52.279 17:34:23 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:52.279 17:34:23 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:12:52.279 17:34:23 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:52.279 17:34:23 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:52.279 17:34:23 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:52.279 17:34:23 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:52.279 17:34:23 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:52.279 17:34:23 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:52.279 17:34:23 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:52.279 17:34:23 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:52.279 17:34:23 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:52.279 17:34:23 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:52.279 "name": "raid_bdev1", 00:12:52.279 "uuid": "87f070f4-a5ae-45c2-a45e-016942a63bbf", 00:12:52.279 "strip_size_kb": 0, 00:12:52.279 "state": "online", 00:12:52.279 "raid_level": "raid1", 00:12:52.279 "superblock": false, 00:12:52.279 "num_base_bdevs": 4, 00:12:52.279 "num_base_bdevs_discovered": 3, 00:12:52.279 "num_base_bdevs_operational": 3, 00:12:52.279 "base_bdevs_list": [ 00:12:52.279 { 00:12:52.279 "name": "spare", 00:12:52.279 "uuid": "bb70b8e8-6ab4-50f9-9c09-8477b339b3b8", 00:12:52.279 "is_configured": true, 00:12:52.279 "data_offset": 0, 00:12:52.279 "data_size": 65536 00:12:52.279 }, 00:12:52.279 { 00:12:52.279 "name": null, 00:12:52.279 "uuid": "00000000-0000-0000-0000-000000000000", 00:12:52.279 "is_configured": false, 00:12:52.279 "data_offset": 0, 00:12:52.279 "data_size": 65536 00:12:52.279 }, 00:12:52.279 { 00:12:52.279 "name": "BaseBdev3", 00:12:52.279 "uuid": "f8a406fe-44e2-5afd-99f6-1fda262aaeae", 00:12:52.279 "is_configured": true, 00:12:52.279 "data_offset": 0, 00:12:52.279 "data_size": 65536 00:12:52.279 }, 00:12:52.279 { 00:12:52.279 "name": "BaseBdev4", 00:12:52.279 "uuid": "39cc8ead-6b09-5b57-8d89-5e8afe49b8a9", 00:12:52.279 "is_configured": true, 00:12:52.279 "data_offset": 0, 00:12:52.279 "data_size": 65536 00:12:52.279 } 00:12:52.279 ] 00:12:52.279 }' 00:12:52.279 17:34:23 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:52.279 17:34:23 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:52.539 17:34:23 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:12:52.539 17:34:23 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:52.539 17:34:23 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:52.539 [2024-11-27 17:34:23.699922] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:12:52.539 [2024-11-27 17:34:23.700002] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:12:52.539 [2024-11-27 17:34:23.700079] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:12:52.539 [2024-11-27 17:34:23.700170] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:12:52.539 [2024-11-27 17:34:23.700185] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:12:52.539 17:34:23 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:52.539 17:34:23 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:52.540 17:34:23 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@720 -- # jq length 00:12:52.540 17:34:23 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:52.540 17:34:23 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:52.540 17:34:23 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:52.799 17:34:23 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:12:52.799 17:34:23 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:12:52.799 17:34:23 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:12:52.799 17:34:23 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:12:52.799 17:34:23 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:12:52.799 17:34:23 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:12:52.799 17:34:23 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@10 -- # local bdev_list 00:12:52.799 17:34:23 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:12:52.799 17:34:23 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@11 -- # local nbd_list 00:12:52.799 17:34:23 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@12 -- # local i 00:12:52.799 17:34:23 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:12:52.799 17:34:23 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:12:52.799 17:34:23 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:12:52.799 /dev/nbd0 00:12:52.799 17:34:23 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:12:53.059 17:34:23 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:12:53.059 17:34:23 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:12:53.059 17:34:23 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@869 -- # local i 00:12:53.059 17:34:23 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:12:53.059 17:34:23 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:12:53.059 17:34:23 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:12:53.059 17:34:23 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@873 -- # break 00:12:53.059 17:34:23 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:12:53.059 17:34:23 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:12:53.059 17:34:24 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:12:53.059 1+0 records in 00:12:53.060 1+0 records out 00:12:53.060 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000313132 s, 13.1 MB/s 00:12:53.060 17:34:24 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:53.060 17:34:24 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@886 -- # size=4096 00:12:53.060 17:34:24 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:53.060 17:34:24 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:12:53.060 17:34:24 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@889 -- # return 0 00:12:53.060 17:34:24 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:12:53.060 17:34:24 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:12:53.060 17:34:24 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:12:53.060 /dev/nbd1 00:12:53.060 17:34:24 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:12:53.320 17:34:24 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:12:53.320 17:34:24 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@868 -- # local nbd_name=nbd1 00:12:53.320 17:34:24 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@869 -- # local i 00:12:53.320 17:34:24 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:12:53.320 17:34:24 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:12:53.320 17:34:24 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@872 -- # grep -q -w nbd1 /proc/partitions 00:12:53.320 17:34:24 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@873 -- # break 00:12:53.320 17:34:24 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:12:53.320 17:34:24 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:12:53.320 17:34:24 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@885 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:12:53.320 1+0 records in 00:12:53.320 1+0 records out 00:12:53.320 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000389077 s, 10.5 MB/s 00:12:53.320 17:34:24 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:53.320 17:34:24 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@886 -- # size=4096 00:12:53.320 17:34:24 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:53.320 17:34:24 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:12:53.320 17:34:24 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@889 -- # return 0 00:12:53.320 17:34:24 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:12:53.320 17:34:24 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:12:53.320 17:34:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@738 -- # cmp -i 0 /dev/nbd0 /dev/nbd1 00:12:53.320 17:34:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:12:53.320 17:34:24 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:12:53.320 17:34:24 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:12:53.320 17:34:24 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@50 -- # local nbd_list 00:12:53.320 17:34:24 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@51 -- # local i 00:12:53.320 17:34:24 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:12:53.320 17:34:24 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:12:53.580 17:34:24 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:12:53.580 17:34:24 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:12:53.580 17:34:24 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:12:53.580 17:34:24 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:12:53.580 17:34:24 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:12:53.580 17:34:24 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:12:53.580 17:34:24 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:12:53.580 17:34:24 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:12:53.580 17:34:24 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:12:53.580 17:34:24 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:12:53.580 17:34:24 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:12:53.580 17:34:24 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:12:53.580 17:34:24 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:12:53.580 17:34:24 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:12:53.580 17:34:24 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:12:53.580 17:34:24 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:12:53.840 17:34:24 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:12:53.840 17:34:24 bdev_raid.raid_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:12:53.840 17:34:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@743 -- # '[' false = true ']' 00:12:53.840 17:34:24 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@784 -- # killprocess 88044 00:12:53.840 17:34:24 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@950 -- # '[' -z 88044 ']' 00:12:53.840 17:34:24 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@954 -- # kill -0 88044 00:12:53.840 17:34:24 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@955 -- # uname 00:12:53.840 17:34:24 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:12:53.840 17:34:24 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 88044 00:12:53.840 17:34:24 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:12:53.840 17:34:24 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:12:53.840 17:34:24 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 88044' 00:12:53.840 killing process with pid 88044 00:12:53.841 17:34:24 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@969 -- # kill 88044 00:12:53.841 Received shutdown signal, test time was about 60.000000 seconds 00:12:53.841 00:12:53.841 Latency(us) 00:12:53.841 [2024-11-27T17:34:25.033Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:12:53.841 [2024-11-27T17:34:25.033Z] =================================================================================================================== 00:12:53.841 [2024-11-27T17:34:25.033Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:12:53.841 [2024-11-27 17:34:24.818701] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:12:53.841 17:34:24 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@974 -- # wait 88044 00:12:53.841 [2024-11-27 17:34:24.869515] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:12:54.100 17:34:25 bdev_raid.raid_rebuild_test -- bdev/bdev_raid.sh@786 -- # return 0 00:12:54.100 00:12:54.100 real 0m15.225s 00:12:54.100 user 0m17.173s 00:12:54.100 sys 0m3.044s 00:12:54.100 17:34:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:12:54.100 ************************************ 00:12:54.100 END TEST raid_rebuild_test 00:12:54.100 ************************************ 00:12:54.100 17:34:25 bdev_raid.raid_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:12:54.100 17:34:25 bdev_raid -- bdev/bdev_raid.sh@979 -- # run_test raid_rebuild_test_sb raid_rebuild_test raid1 4 true false true 00:12:54.100 17:34:25 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 7 -le 1 ']' 00:12:54.100 17:34:25 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:12:54.100 17:34:25 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:12:54.100 ************************************ 00:12:54.100 START TEST raid_rebuild_test_sb 00:12:54.100 ************************************ 00:12:54.100 17:34:25 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@1125 -- # raid_rebuild_test raid1 4 true false true 00:12:54.100 17:34:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:12:54.100 17:34:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=4 00:12:54.100 17:34:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:12:54.100 17:34:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:12:54.100 17:34:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@573 -- # local verify=true 00:12:54.100 17:34:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:12:54.100 17:34:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:54.100 17:34:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:12:54.100 17:34:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:54.100 17:34:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:54.100 17:34:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:12:54.100 17:34:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:54.100 17:34:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:54.100 17:34:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev3 00:12:54.100 17:34:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:54.100 17:34:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:54.100 17:34:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev4 00:12:54.100 17:34:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:12:54.100 17:34:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:12:54.100 17:34:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:12:54.100 17:34:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:12:54.100 17:34:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:12:54.100 17:34:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # local strip_size 00:12:54.100 17:34:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@577 -- # local create_arg 00:12:54.100 17:34:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:12:54.100 17:34:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@579 -- # local data_offset 00:12:54.100 17:34:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:12:54.100 17:34:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:12:54.100 17:34:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:12:54.100 17:34:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:12:54.100 17:34:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@597 -- # raid_pid=88470 00:12:54.100 17:34:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:12:54.100 17:34:25 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@598 -- # waitforlisten 88470 00:12:54.100 17:34:25 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@831 -- # '[' -z 88470 ']' 00:12:54.100 17:34:25 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:12:54.100 17:34:25 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@836 -- # local max_retries=100 00:12:54.100 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:12:54.100 17:34:25 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:12:54.100 17:34:25 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@840 -- # xtrace_disable 00:12:54.100 17:34:25 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:54.100 [2024-11-27 17:34:25.282573] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:12:54.100 [2024-11-27 17:34:25.282763] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.ealI/O size of 3145728 is greater than zero copy threshold (65536). 00:12:54.100 Zero copy mechanism will not be used. 00:12:54.100 :6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid88470 ] 00:12:54.360 [2024-11-27 17:34:25.427960] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:12:54.360 [2024-11-27 17:34:25.473340] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:12:54.360 [2024-11-27 17:34:25.515921] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:12:54.360 [2024-11-27 17:34:25.516037] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:12:54.930 17:34:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:12:55.190 17:34:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@864 -- # return 0 00:12:55.190 17:34:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:55.190 17:34:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:12:55.190 17:34:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:55.190 17:34:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:55.190 BaseBdev1_malloc 00:12:55.190 17:34:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:55.190 17:34:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:12:55.190 17:34:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:55.190 17:34:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:55.190 [2024-11-27 17:34:26.142188] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:12:55.191 [2024-11-27 17:34:26.142258] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:55.191 [2024-11-27 17:34:26.142282] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:12:55.191 [2024-11-27 17:34:26.142295] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:55.191 [2024-11-27 17:34:26.144239] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:55.191 [2024-11-27 17:34:26.144277] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:12:55.191 BaseBdev1 00:12:55.191 17:34:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:55.191 17:34:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:55.191 17:34:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:12:55.191 17:34:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:55.191 17:34:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:55.191 BaseBdev2_malloc 00:12:55.191 17:34:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:55.191 17:34:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:12:55.191 17:34:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:55.191 17:34:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:55.191 [2024-11-27 17:34:26.188005] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:12:55.191 [2024-11-27 17:34:26.188278] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:55.191 [2024-11-27 17:34:26.188381] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:12:55.191 [2024-11-27 17:34:26.188457] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:55.191 [2024-11-27 17:34:26.193294] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:55.191 [2024-11-27 17:34:26.193441] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:12:55.191 BaseBdev2 00:12:55.191 17:34:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:55.191 17:34:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:55.191 17:34:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:12:55.191 17:34:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:55.191 17:34:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:55.191 BaseBdev3_malloc 00:12:55.191 17:34:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:55.191 17:34:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev3_malloc -p BaseBdev3 00:12:55.191 17:34:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:55.191 17:34:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:55.191 [2024-11-27 17:34:26.219570] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev3_malloc 00:12:55.191 [2024-11-27 17:34:26.219627] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:55.191 [2024-11-27 17:34:26.219652] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:12:55.191 [2024-11-27 17:34:26.219661] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:55.191 [2024-11-27 17:34:26.221594] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:55.191 [2024-11-27 17:34:26.221630] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:12:55.191 BaseBdev3 00:12:55.191 17:34:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:55.191 17:34:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:12:55.191 17:34:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:12:55.191 17:34:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:55.191 17:34:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:55.191 BaseBdev4_malloc 00:12:55.191 17:34:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:55.191 17:34:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev4_malloc -p BaseBdev4 00:12:55.191 17:34:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:55.191 17:34:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:55.191 [2024-11-27 17:34:26.248218] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev4_malloc 00:12:55.191 [2024-11-27 17:34:26.248264] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:55.191 [2024-11-27 17:34:26.248283] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:12:55.191 [2024-11-27 17:34:26.248291] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:55.191 [2024-11-27 17:34:26.250150] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:55.191 [2024-11-27 17:34:26.250183] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:12:55.191 BaseBdev4 00:12:55.191 17:34:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:55.191 17:34:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:12:55.191 17:34:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:55.191 17:34:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:55.191 spare_malloc 00:12:55.191 17:34:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:55.191 17:34:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:12:55.191 17:34:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:55.191 17:34:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:55.191 spare_delay 00:12:55.191 17:34:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:55.191 17:34:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:12:55.191 17:34:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:55.191 17:34:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:55.191 [2024-11-27 17:34:26.288596] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:12:55.191 [2024-11-27 17:34:26.288644] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:12:55.191 [2024-11-27 17:34:26.288661] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009c80 00:12:55.191 [2024-11-27 17:34:26.288669] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:12:55.191 [2024-11-27 17:34:26.290556] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:12:55.191 [2024-11-27 17:34:26.290592] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:12:55.191 spare 00:12:55.191 17:34:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:55.191 17:34:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 00:12:55.191 17:34:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:55.191 17:34:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:55.191 [2024-11-27 17:34:26.300645] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:12:55.191 [2024-11-27 17:34:26.302288] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:12:55.191 [2024-11-27 17:34:26.302341] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:12:55.191 [2024-11-27 17:34:26.302385] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:12:55.191 [2024-11-27 17:34:26.302532] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:12:55.191 [2024-11-27 17:34:26.302543] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:12:55.191 [2024-11-27 17:34:26.302779] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:12:55.191 [2024-11-27 17:34:26.302907] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:12:55.191 [2024-11-27 17:34:26.302921] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:12:55.191 [2024-11-27 17:34:26.303035] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:12:55.191 17:34:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:55.191 17:34:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:12:55.191 17:34:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:12:55.191 17:34:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:12:55.191 17:34:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:12:55.191 17:34:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:12:55.191 17:34:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:12:55.191 17:34:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:12:55.191 17:34:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:12:55.191 17:34:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:12:55.191 17:34:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:12:55.191 17:34:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:55.191 17:34:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:12:55.191 17:34:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:55.191 17:34:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:55.191 17:34:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:55.191 17:34:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:12:55.191 "name": "raid_bdev1", 00:12:55.191 "uuid": "1bbbe449-6fda-4272-8c5b-fafa6297ec22", 00:12:55.191 "strip_size_kb": 0, 00:12:55.191 "state": "online", 00:12:55.192 "raid_level": "raid1", 00:12:55.192 "superblock": true, 00:12:55.192 "num_base_bdevs": 4, 00:12:55.192 "num_base_bdevs_discovered": 4, 00:12:55.192 "num_base_bdevs_operational": 4, 00:12:55.192 "base_bdevs_list": [ 00:12:55.192 { 00:12:55.192 "name": "BaseBdev1", 00:12:55.192 "uuid": "c86727eb-319f-54e2-8cf1-e0499d0331d3", 00:12:55.192 "is_configured": true, 00:12:55.192 "data_offset": 2048, 00:12:55.192 "data_size": 63488 00:12:55.192 }, 00:12:55.192 { 00:12:55.192 "name": "BaseBdev2", 00:12:55.192 "uuid": "025285a6-54b2-5fa7-986f-a78716cadec7", 00:12:55.192 "is_configured": true, 00:12:55.192 "data_offset": 2048, 00:12:55.192 "data_size": 63488 00:12:55.192 }, 00:12:55.192 { 00:12:55.192 "name": "BaseBdev3", 00:12:55.192 "uuid": "9b6eb3fc-dd72-5973-afbe-9ecd0fafecab", 00:12:55.192 "is_configured": true, 00:12:55.192 "data_offset": 2048, 00:12:55.192 "data_size": 63488 00:12:55.192 }, 00:12:55.192 { 00:12:55.192 "name": "BaseBdev4", 00:12:55.192 "uuid": "7dd75dc9-9415-55cc-bbe5-e0d8fdd0e6de", 00:12:55.192 "is_configured": true, 00:12:55.192 "data_offset": 2048, 00:12:55.192 "data_size": 63488 00:12:55.192 } 00:12:55.192 ] 00:12:55.192 }' 00:12:55.192 17:34:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:12:55.192 17:34:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:55.761 17:34:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:12:55.761 17:34:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:55.761 17:34:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:55.761 17:34:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:12:55.761 [2024-11-27 17:34:26.736160] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:12:55.761 17:34:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:55.761 17:34:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=63488 00:12:55.761 17:34:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:12:55.761 17:34:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:12:55.761 17:34:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:55.761 17:34:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:12:55.761 17:34:26 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:55.761 17:34:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # data_offset=2048 00:12:55.761 17:34:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:12:55.761 17:34:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:12:55.761 17:34:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:12:55.761 17:34:26 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:12:55.761 17:34:26 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:12:55.761 17:34:26 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:12:55.761 17:34:26 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # local bdev_list 00:12:55.761 17:34:26 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:12:55.761 17:34:26 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # local nbd_list 00:12:55.761 17:34:26 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@12 -- # local i 00:12:55.761 17:34:26 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:12:55.761 17:34:26 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:12:55.761 17:34:26 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:12:56.020 [2024-11-27 17:34:26.987455] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000027a0 00:12:56.020 /dev/nbd0 00:12:56.020 17:34:27 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:12:56.020 17:34:27 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:12:56.020 17:34:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:12:56.020 17:34:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@869 -- # local i 00:12:56.020 17:34:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:12:56.020 17:34:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:12:56.020 17:34:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:12:56.020 17:34:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@873 -- # break 00:12:56.020 17:34:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:12:56.020 17:34:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:12:56.020 17:34:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:12:56.020 1+0 records in 00:12:56.020 1+0 records out 00:12:56.020 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000542993 s, 7.5 MB/s 00:12:56.020 17:34:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:56.020 17:34:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@886 -- # size=4096 00:12:56.020 17:34:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:12:56.020 17:34:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:12:56.020 17:34:27 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@889 -- # return 0 00:12:56.020 17:34:27 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:12:56.020 17:34:27 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:12:56.020 17:34:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@629 -- # '[' raid1 = raid5f ']' 00:12:56.020 17:34:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@633 -- # write_unit_size=1 00:12:56.020 17:34:27 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=512 count=63488 oflag=direct 00:13:01.294 63488+0 records in 00:13:01.294 63488+0 records out 00:13:01.294 32505856 bytes (33 MB, 31 MiB) copied, 5.26079 s, 6.2 MB/s 00:13:01.294 17:34:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:13:01.294 17:34:32 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:13:01.294 17:34:32 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:13:01.294 17:34:32 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # local nbd_list 00:13:01.294 17:34:32 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@51 -- # local i 00:13:01.294 17:34:32 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:13:01.294 17:34:32 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:13:01.555 [2024-11-27 17:34:32.544331] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:01.555 17:34:32 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:13:01.555 17:34:32 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:13:01.555 17:34:32 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:13:01.555 17:34:32 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:13:01.555 17:34:32 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:13:01.555 17:34:32 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:13:01.555 17:34:32 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:13:01.555 17:34:32 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:13:01.555 17:34:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:13:01.555 17:34:32 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:01.555 17:34:32 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:01.555 [2024-11-27 17:34:32.572338] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:13:01.555 17:34:32 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:01.555 17:34:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:13:01.555 17:34:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:01.555 17:34:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:01.555 17:34:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:01.555 17:34:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:01.555 17:34:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:01.555 17:34:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:01.555 17:34:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:01.555 17:34:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:01.555 17:34:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:01.555 17:34:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:01.555 17:34:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:01.555 17:34:32 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:01.555 17:34:32 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:01.555 17:34:32 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:01.555 17:34:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:01.555 "name": "raid_bdev1", 00:13:01.555 "uuid": "1bbbe449-6fda-4272-8c5b-fafa6297ec22", 00:13:01.555 "strip_size_kb": 0, 00:13:01.555 "state": "online", 00:13:01.555 "raid_level": "raid1", 00:13:01.555 "superblock": true, 00:13:01.555 "num_base_bdevs": 4, 00:13:01.555 "num_base_bdevs_discovered": 3, 00:13:01.555 "num_base_bdevs_operational": 3, 00:13:01.555 "base_bdevs_list": [ 00:13:01.555 { 00:13:01.555 "name": null, 00:13:01.555 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:01.555 "is_configured": false, 00:13:01.555 "data_offset": 0, 00:13:01.555 "data_size": 63488 00:13:01.555 }, 00:13:01.555 { 00:13:01.555 "name": "BaseBdev2", 00:13:01.555 "uuid": "025285a6-54b2-5fa7-986f-a78716cadec7", 00:13:01.555 "is_configured": true, 00:13:01.555 "data_offset": 2048, 00:13:01.555 "data_size": 63488 00:13:01.555 }, 00:13:01.555 { 00:13:01.555 "name": "BaseBdev3", 00:13:01.555 "uuid": "9b6eb3fc-dd72-5973-afbe-9ecd0fafecab", 00:13:01.555 "is_configured": true, 00:13:01.555 "data_offset": 2048, 00:13:01.555 "data_size": 63488 00:13:01.555 }, 00:13:01.555 { 00:13:01.555 "name": "BaseBdev4", 00:13:01.555 "uuid": "7dd75dc9-9415-55cc-bbe5-e0d8fdd0e6de", 00:13:01.555 "is_configured": true, 00:13:01.555 "data_offset": 2048, 00:13:01.555 "data_size": 63488 00:13:01.555 } 00:13:01.555 ] 00:13:01.555 }' 00:13:01.555 17:34:32 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:01.555 17:34:32 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:02.126 17:34:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:13:02.126 17:34:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:02.126 17:34:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:02.126 [2024-11-27 17:34:33.035531] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:13:02.126 [2024-11-27 17:34:33.038909] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000c3e420 00:13:02.126 [2024-11-27 17:34:33.040726] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:13:02.126 17:34:33 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:02.126 17:34:33 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@647 -- # sleep 1 00:13:03.067 17:34:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:03.067 17:34:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:03.067 17:34:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:03.067 17:34:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:03.067 17:34:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:03.067 17:34:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:03.067 17:34:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:03.067 17:34:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:03.067 17:34:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:03.067 17:34:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:03.067 17:34:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:03.067 "name": "raid_bdev1", 00:13:03.067 "uuid": "1bbbe449-6fda-4272-8c5b-fafa6297ec22", 00:13:03.067 "strip_size_kb": 0, 00:13:03.067 "state": "online", 00:13:03.067 "raid_level": "raid1", 00:13:03.067 "superblock": true, 00:13:03.067 "num_base_bdevs": 4, 00:13:03.067 "num_base_bdevs_discovered": 4, 00:13:03.067 "num_base_bdevs_operational": 4, 00:13:03.067 "process": { 00:13:03.067 "type": "rebuild", 00:13:03.067 "target": "spare", 00:13:03.067 "progress": { 00:13:03.067 "blocks": 20480, 00:13:03.067 "percent": 32 00:13:03.067 } 00:13:03.067 }, 00:13:03.067 "base_bdevs_list": [ 00:13:03.067 { 00:13:03.067 "name": "spare", 00:13:03.067 "uuid": "4a1b1c6c-4e8e-5972-b73e-719c9e88c88f", 00:13:03.067 "is_configured": true, 00:13:03.067 "data_offset": 2048, 00:13:03.067 "data_size": 63488 00:13:03.067 }, 00:13:03.067 { 00:13:03.067 "name": "BaseBdev2", 00:13:03.067 "uuid": "025285a6-54b2-5fa7-986f-a78716cadec7", 00:13:03.067 "is_configured": true, 00:13:03.067 "data_offset": 2048, 00:13:03.067 "data_size": 63488 00:13:03.067 }, 00:13:03.067 { 00:13:03.067 "name": "BaseBdev3", 00:13:03.067 "uuid": "9b6eb3fc-dd72-5973-afbe-9ecd0fafecab", 00:13:03.067 "is_configured": true, 00:13:03.067 "data_offset": 2048, 00:13:03.067 "data_size": 63488 00:13:03.067 }, 00:13:03.067 { 00:13:03.067 "name": "BaseBdev4", 00:13:03.067 "uuid": "7dd75dc9-9415-55cc-bbe5-e0d8fdd0e6de", 00:13:03.067 "is_configured": true, 00:13:03.067 "data_offset": 2048, 00:13:03.067 "data_size": 63488 00:13:03.067 } 00:13:03.067 ] 00:13:03.067 }' 00:13:03.067 17:34:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:03.067 17:34:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:03.067 17:34:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:03.067 17:34:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:03.067 17:34:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:13:03.067 17:34:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:03.067 17:34:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:03.067 [2024-11-27 17:34:34.179283] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:03.067 [2024-11-27 17:34:34.245159] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:13:03.067 [2024-11-27 17:34:34.245225] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:03.067 [2024-11-27 17:34:34.245246] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:03.067 [2024-11-27 17:34:34.245253] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:13:03.328 17:34:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:03.328 17:34:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:13:03.328 17:34:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:03.328 17:34:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:03.328 17:34:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:03.328 17:34:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:03.328 17:34:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:03.328 17:34:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:03.328 17:34:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:03.328 17:34:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:03.328 17:34:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:03.328 17:34:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:03.328 17:34:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:03.328 17:34:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:03.328 17:34:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:03.328 17:34:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:03.328 17:34:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:03.328 "name": "raid_bdev1", 00:13:03.328 "uuid": "1bbbe449-6fda-4272-8c5b-fafa6297ec22", 00:13:03.328 "strip_size_kb": 0, 00:13:03.328 "state": "online", 00:13:03.328 "raid_level": "raid1", 00:13:03.328 "superblock": true, 00:13:03.328 "num_base_bdevs": 4, 00:13:03.328 "num_base_bdevs_discovered": 3, 00:13:03.328 "num_base_bdevs_operational": 3, 00:13:03.328 "base_bdevs_list": [ 00:13:03.328 { 00:13:03.328 "name": null, 00:13:03.328 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:03.328 "is_configured": false, 00:13:03.328 "data_offset": 0, 00:13:03.328 "data_size": 63488 00:13:03.328 }, 00:13:03.328 { 00:13:03.328 "name": "BaseBdev2", 00:13:03.328 "uuid": "025285a6-54b2-5fa7-986f-a78716cadec7", 00:13:03.328 "is_configured": true, 00:13:03.328 "data_offset": 2048, 00:13:03.328 "data_size": 63488 00:13:03.328 }, 00:13:03.328 { 00:13:03.328 "name": "BaseBdev3", 00:13:03.328 "uuid": "9b6eb3fc-dd72-5973-afbe-9ecd0fafecab", 00:13:03.328 "is_configured": true, 00:13:03.328 "data_offset": 2048, 00:13:03.328 "data_size": 63488 00:13:03.328 }, 00:13:03.328 { 00:13:03.328 "name": "BaseBdev4", 00:13:03.328 "uuid": "7dd75dc9-9415-55cc-bbe5-e0d8fdd0e6de", 00:13:03.328 "is_configured": true, 00:13:03.328 "data_offset": 2048, 00:13:03.328 "data_size": 63488 00:13:03.328 } 00:13:03.328 ] 00:13:03.328 }' 00:13:03.328 17:34:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:03.328 17:34:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:03.588 17:34:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:13:03.588 17:34:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:03.588 17:34:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:13:03.588 17:34:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:13:03.588 17:34:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:03.588 17:34:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:03.588 17:34:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:03.588 17:34:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:03.588 17:34:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:03.588 17:34:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:03.588 17:34:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:03.588 "name": "raid_bdev1", 00:13:03.588 "uuid": "1bbbe449-6fda-4272-8c5b-fafa6297ec22", 00:13:03.588 "strip_size_kb": 0, 00:13:03.588 "state": "online", 00:13:03.588 "raid_level": "raid1", 00:13:03.588 "superblock": true, 00:13:03.588 "num_base_bdevs": 4, 00:13:03.588 "num_base_bdevs_discovered": 3, 00:13:03.588 "num_base_bdevs_operational": 3, 00:13:03.588 "base_bdevs_list": [ 00:13:03.588 { 00:13:03.588 "name": null, 00:13:03.588 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:03.588 "is_configured": false, 00:13:03.588 "data_offset": 0, 00:13:03.588 "data_size": 63488 00:13:03.588 }, 00:13:03.588 { 00:13:03.588 "name": "BaseBdev2", 00:13:03.588 "uuid": "025285a6-54b2-5fa7-986f-a78716cadec7", 00:13:03.588 "is_configured": true, 00:13:03.588 "data_offset": 2048, 00:13:03.588 "data_size": 63488 00:13:03.588 }, 00:13:03.588 { 00:13:03.588 "name": "BaseBdev3", 00:13:03.588 "uuid": "9b6eb3fc-dd72-5973-afbe-9ecd0fafecab", 00:13:03.588 "is_configured": true, 00:13:03.588 "data_offset": 2048, 00:13:03.588 "data_size": 63488 00:13:03.588 }, 00:13:03.588 { 00:13:03.588 "name": "BaseBdev4", 00:13:03.588 "uuid": "7dd75dc9-9415-55cc-bbe5-e0d8fdd0e6de", 00:13:03.588 "is_configured": true, 00:13:03.588 "data_offset": 2048, 00:13:03.588 "data_size": 63488 00:13:03.588 } 00:13:03.588 ] 00:13:03.588 }' 00:13:03.588 17:34:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:03.848 17:34:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:13:03.848 17:34:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:03.848 17:34:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:13:03.848 17:34:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:13:03.848 17:34:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:03.848 17:34:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:03.848 [2024-11-27 17:34:34.851972] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:13:03.848 [2024-11-27 17:34:34.855153] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000c3e4f0 00:13:03.848 [2024-11-27 17:34:34.856955] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:13:03.848 17:34:34 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:03.848 17:34:34 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@663 -- # sleep 1 00:13:04.788 17:34:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:04.788 17:34:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:04.788 17:34:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:04.788 17:34:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:04.788 17:34:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:04.788 17:34:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:04.788 17:34:35 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:04.788 17:34:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:04.788 17:34:35 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:04.788 17:34:35 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:04.788 17:34:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:04.788 "name": "raid_bdev1", 00:13:04.788 "uuid": "1bbbe449-6fda-4272-8c5b-fafa6297ec22", 00:13:04.789 "strip_size_kb": 0, 00:13:04.789 "state": "online", 00:13:04.789 "raid_level": "raid1", 00:13:04.789 "superblock": true, 00:13:04.789 "num_base_bdevs": 4, 00:13:04.789 "num_base_bdevs_discovered": 4, 00:13:04.789 "num_base_bdevs_operational": 4, 00:13:04.789 "process": { 00:13:04.789 "type": "rebuild", 00:13:04.789 "target": "spare", 00:13:04.789 "progress": { 00:13:04.789 "blocks": 20480, 00:13:04.789 "percent": 32 00:13:04.789 } 00:13:04.789 }, 00:13:04.789 "base_bdevs_list": [ 00:13:04.789 { 00:13:04.789 "name": "spare", 00:13:04.789 "uuid": "4a1b1c6c-4e8e-5972-b73e-719c9e88c88f", 00:13:04.789 "is_configured": true, 00:13:04.789 "data_offset": 2048, 00:13:04.789 "data_size": 63488 00:13:04.789 }, 00:13:04.789 { 00:13:04.789 "name": "BaseBdev2", 00:13:04.789 "uuid": "025285a6-54b2-5fa7-986f-a78716cadec7", 00:13:04.789 "is_configured": true, 00:13:04.789 "data_offset": 2048, 00:13:04.789 "data_size": 63488 00:13:04.789 }, 00:13:04.789 { 00:13:04.789 "name": "BaseBdev3", 00:13:04.789 "uuid": "9b6eb3fc-dd72-5973-afbe-9ecd0fafecab", 00:13:04.789 "is_configured": true, 00:13:04.789 "data_offset": 2048, 00:13:04.789 "data_size": 63488 00:13:04.789 }, 00:13:04.789 { 00:13:04.789 "name": "BaseBdev4", 00:13:04.789 "uuid": "7dd75dc9-9415-55cc-bbe5-e0d8fdd0e6de", 00:13:04.789 "is_configured": true, 00:13:04.789 "data_offset": 2048, 00:13:04.789 "data_size": 63488 00:13:04.789 } 00:13:04.789 ] 00:13:04.789 }' 00:13:04.789 17:34:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:04.789 17:34:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:04.789 17:34:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:05.049 17:34:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:05.049 17:34:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:13:05.049 17:34:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:13:05.049 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:13:05.049 17:34:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=4 00:13:05.049 17:34:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:13:05.049 17:34:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@693 -- # '[' 4 -gt 2 ']' 00:13:05.049 17:34:35 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@695 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:13:05.049 17:34:35 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:05.049 17:34:35 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:05.049 [2024-11-27 17:34:35.996089] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:13:05.049 [2024-11-27 17:34:36.160700] bdev_raid.c:1970:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 1 raid_ch: 0x60d000c3e4f0 00:13:05.049 17:34:36 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:05.049 17:34:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@698 -- # base_bdevs[1]= 00:13:05.049 17:34:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@699 -- # (( num_base_bdevs_operational-- )) 00:13:05.049 17:34:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@702 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:05.049 17:34:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:05.049 17:34:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:05.049 17:34:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:05.049 17:34:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:05.049 17:34:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:05.049 17:34:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:05.049 17:34:36 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:05.049 17:34:36 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:05.049 17:34:36 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:05.049 17:34:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:05.049 "name": "raid_bdev1", 00:13:05.049 "uuid": "1bbbe449-6fda-4272-8c5b-fafa6297ec22", 00:13:05.049 "strip_size_kb": 0, 00:13:05.049 "state": "online", 00:13:05.049 "raid_level": "raid1", 00:13:05.049 "superblock": true, 00:13:05.049 "num_base_bdevs": 4, 00:13:05.049 "num_base_bdevs_discovered": 3, 00:13:05.049 "num_base_bdevs_operational": 3, 00:13:05.049 "process": { 00:13:05.049 "type": "rebuild", 00:13:05.049 "target": "spare", 00:13:05.049 "progress": { 00:13:05.049 "blocks": 24576, 00:13:05.049 "percent": 38 00:13:05.049 } 00:13:05.049 }, 00:13:05.049 "base_bdevs_list": [ 00:13:05.049 { 00:13:05.049 "name": "spare", 00:13:05.049 "uuid": "4a1b1c6c-4e8e-5972-b73e-719c9e88c88f", 00:13:05.049 "is_configured": true, 00:13:05.049 "data_offset": 2048, 00:13:05.049 "data_size": 63488 00:13:05.049 }, 00:13:05.049 { 00:13:05.049 "name": null, 00:13:05.049 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:05.049 "is_configured": false, 00:13:05.049 "data_offset": 0, 00:13:05.049 "data_size": 63488 00:13:05.049 }, 00:13:05.049 { 00:13:05.049 "name": "BaseBdev3", 00:13:05.049 "uuid": "9b6eb3fc-dd72-5973-afbe-9ecd0fafecab", 00:13:05.049 "is_configured": true, 00:13:05.049 "data_offset": 2048, 00:13:05.049 "data_size": 63488 00:13:05.049 }, 00:13:05.049 { 00:13:05.049 "name": "BaseBdev4", 00:13:05.049 "uuid": "7dd75dc9-9415-55cc-bbe5-e0d8fdd0e6de", 00:13:05.049 "is_configured": true, 00:13:05.049 "data_offset": 2048, 00:13:05.049 "data_size": 63488 00:13:05.049 } 00:13:05.049 ] 00:13:05.049 }' 00:13:05.049 17:34:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:05.310 17:34:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:05.310 17:34:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:05.310 17:34:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:05.310 17:34:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@706 -- # local timeout=384 00:13:05.310 17:34:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:05.310 17:34:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:05.310 17:34:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:05.310 17:34:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:05.310 17:34:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:05.310 17:34:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:05.310 17:34:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:05.310 17:34:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:05.310 17:34:36 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:05.310 17:34:36 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:05.310 17:34:36 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:05.310 17:34:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:05.310 "name": "raid_bdev1", 00:13:05.310 "uuid": "1bbbe449-6fda-4272-8c5b-fafa6297ec22", 00:13:05.310 "strip_size_kb": 0, 00:13:05.310 "state": "online", 00:13:05.310 "raid_level": "raid1", 00:13:05.310 "superblock": true, 00:13:05.310 "num_base_bdevs": 4, 00:13:05.310 "num_base_bdevs_discovered": 3, 00:13:05.310 "num_base_bdevs_operational": 3, 00:13:05.310 "process": { 00:13:05.310 "type": "rebuild", 00:13:05.310 "target": "spare", 00:13:05.310 "progress": { 00:13:05.310 "blocks": 26624, 00:13:05.310 "percent": 41 00:13:05.310 } 00:13:05.310 }, 00:13:05.310 "base_bdevs_list": [ 00:13:05.310 { 00:13:05.310 "name": "spare", 00:13:05.310 "uuid": "4a1b1c6c-4e8e-5972-b73e-719c9e88c88f", 00:13:05.310 "is_configured": true, 00:13:05.310 "data_offset": 2048, 00:13:05.310 "data_size": 63488 00:13:05.310 }, 00:13:05.310 { 00:13:05.310 "name": null, 00:13:05.310 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:05.310 "is_configured": false, 00:13:05.310 "data_offset": 0, 00:13:05.310 "data_size": 63488 00:13:05.310 }, 00:13:05.310 { 00:13:05.310 "name": "BaseBdev3", 00:13:05.310 "uuid": "9b6eb3fc-dd72-5973-afbe-9ecd0fafecab", 00:13:05.310 "is_configured": true, 00:13:05.310 "data_offset": 2048, 00:13:05.310 "data_size": 63488 00:13:05.310 }, 00:13:05.310 { 00:13:05.310 "name": "BaseBdev4", 00:13:05.310 "uuid": "7dd75dc9-9415-55cc-bbe5-e0d8fdd0e6de", 00:13:05.310 "is_configured": true, 00:13:05.310 "data_offset": 2048, 00:13:05.310 "data_size": 63488 00:13:05.310 } 00:13:05.310 ] 00:13:05.310 }' 00:13:05.310 17:34:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:05.310 17:34:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:05.310 17:34:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:05.310 17:34:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:05.310 17:34:36 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:13:06.692 17:34:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:06.692 17:34:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:06.692 17:34:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:06.692 17:34:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:06.692 17:34:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:06.692 17:34:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:06.692 17:34:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:06.692 17:34:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:06.692 17:34:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:06.692 17:34:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:06.692 17:34:37 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:06.692 17:34:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:06.692 "name": "raid_bdev1", 00:13:06.692 "uuid": "1bbbe449-6fda-4272-8c5b-fafa6297ec22", 00:13:06.692 "strip_size_kb": 0, 00:13:06.692 "state": "online", 00:13:06.692 "raid_level": "raid1", 00:13:06.692 "superblock": true, 00:13:06.692 "num_base_bdevs": 4, 00:13:06.692 "num_base_bdevs_discovered": 3, 00:13:06.692 "num_base_bdevs_operational": 3, 00:13:06.692 "process": { 00:13:06.692 "type": "rebuild", 00:13:06.692 "target": "spare", 00:13:06.692 "progress": { 00:13:06.692 "blocks": 51200, 00:13:06.692 "percent": 80 00:13:06.692 } 00:13:06.692 }, 00:13:06.692 "base_bdevs_list": [ 00:13:06.692 { 00:13:06.692 "name": "spare", 00:13:06.692 "uuid": "4a1b1c6c-4e8e-5972-b73e-719c9e88c88f", 00:13:06.692 "is_configured": true, 00:13:06.692 "data_offset": 2048, 00:13:06.692 "data_size": 63488 00:13:06.692 }, 00:13:06.692 { 00:13:06.692 "name": null, 00:13:06.692 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:06.692 "is_configured": false, 00:13:06.692 "data_offset": 0, 00:13:06.692 "data_size": 63488 00:13:06.692 }, 00:13:06.692 { 00:13:06.692 "name": "BaseBdev3", 00:13:06.692 "uuid": "9b6eb3fc-dd72-5973-afbe-9ecd0fafecab", 00:13:06.692 "is_configured": true, 00:13:06.692 "data_offset": 2048, 00:13:06.692 "data_size": 63488 00:13:06.692 }, 00:13:06.692 { 00:13:06.692 "name": "BaseBdev4", 00:13:06.692 "uuid": "7dd75dc9-9415-55cc-bbe5-e0d8fdd0e6de", 00:13:06.692 "is_configured": true, 00:13:06.692 "data_offset": 2048, 00:13:06.692 "data_size": 63488 00:13:06.692 } 00:13:06.692 ] 00:13:06.692 }' 00:13:06.692 17:34:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:06.692 17:34:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:06.692 17:34:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:06.692 17:34:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:06.692 17:34:37 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:13:06.952 [2024-11-27 17:34:38.067100] bdev_raid.c:2896:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:13:06.952 [2024-11-27 17:34:38.067176] bdev_raid.c:2558:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:13:06.952 [2024-11-27 17:34:38.067274] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:07.522 17:34:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:07.522 17:34:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:07.522 17:34:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:07.522 17:34:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:07.522 17:34:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:07.522 17:34:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:07.522 17:34:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:07.522 17:34:38 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:07.522 17:34:38 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:07.522 17:34:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:07.522 17:34:38 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:07.522 17:34:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:07.522 "name": "raid_bdev1", 00:13:07.522 "uuid": "1bbbe449-6fda-4272-8c5b-fafa6297ec22", 00:13:07.522 "strip_size_kb": 0, 00:13:07.522 "state": "online", 00:13:07.522 "raid_level": "raid1", 00:13:07.522 "superblock": true, 00:13:07.522 "num_base_bdevs": 4, 00:13:07.522 "num_base_bdevs_discovered": 3, 00:13:07.522 "num_base_bdevs_operational": 3, 00:13:07.522 "base_bdevs_list": [ 00:13:07.522 { 00:13:07.522 "name": "spare", 00:13:07.522 "uuid": "4a1b1c6c-4e8e-5972-b73e-719c9e88c88f", 00:13:07.522 "is_configured": true, 00:13:07.522 "data_offset": 2048, 00:13:07.522 "data_size": 63488 00:13:07.522 }, 00:13:07.522 { 00:13:07.522 "name": null, 00:13:07.522 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:07.522 "is_configured": false, 00:13:07.522 "data_offset": 0, 00:13:07.522 "data_size": 63488 00:13:07.522 }, 00:13:07.522 { 00:13:07.522 "name": "BaseBdev3", 00:13:07.522 "uuid": "9b6eb3fc-dd72-5973-afbe-9ecd0fafecab", 00:13:07.522 "is_configured": true, 00:13:07.522 "data_offset": 2048, 00:13:07.522 "data_size": 63488 00:13:07.522 }, 00:13:07.522 { 00:13:07.522 "name": "BaseBdev4", 00:13:07.522 "uuid": "7dd75dc9-9415-55cc-bbe5-e0d8fdd0e6de", 00:13:07.522 "is_configured": true, 00:13:07.522 "data_offset": 2048, 00:13:07.522 "data_size": 63488 00:13:07.522 } 00:13:07.522 ] 00:13:07.522 }' 00:13:07.522 17:34:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:07.782 17:34:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:13:07.782 17:34:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:07.782 17:34:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:13:07.782 17:34:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@709 -- # break 00:13:07.782 17:34:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:13:07.782 17:34:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:07.782 17:34:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:13:07.782 17:34:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:13:07.782 17:34:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:07.782 17:34:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:07.782 17:34:38 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:07.782 17:34:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:07.782 17:34:38 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:07.782 17:34:38 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:07.782 17:34:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:07.782 "name": "raid_bdev1", 00:13:07.782 "uuid": "1bbbe449-6fda-4272-8c5b-fafa6297ec22", 00:13:07.782 "strip_size_kb": 0, 00:13:07.782 "state": "online", 00:13:07.782 "raid_level": "raid1", 00:13:07.782 "superblock": true, 00:13:07.782 "num_base_bdevs": 4, 00:13:07.782 "num_base_bdevs_discovered": 3, 00:13:07.782 "num_base_bdevs_operational": 3, 00:13:07.782 "base_bdevs_list": [ 00:13:07.782 { 00:13:07.782 "name": "spare", 00:13:07.782 "uuid": "4a1b1c6c-4e8e-5972-b73e-719c9e88c88f", 00:13:07.782 "is_configured": true, 00:13:07.782 "data_offset": 2048, 00:13:07.782 "data_size": 63488 00:13:07.782 }, 00:13:07.782 { 00:13:07.782 "name": null, 00:13:07.782 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:07.782 "is_configured": false, 00:13:07.782 "data_offset": 0, 00:13:07.782 "data_size": 63488 00:13:07.782 }, 00:13:07.782 { 00:13:07.782 "name": "BaseBdev3", 00:13:07.782 "uuid": "9b6eb3fc-dd72-5973-afbe-9ecd0fafecab", 00:13:07.782 "is_configured": true, 00:13:07.782 "data_offset": 2048, 00:13:07.782 "data_size": 63488 00:13:07.782 }, 00:13:07.782 { 00:13:07.782 "name": "BaseBdev4", 00:13:07.782 "uuid": "7dd75dc9-9415-55cc-bbe5-e0d8fdd0e6de", 00:13:07.782 "is_configured": true, 00:13:07.782 "data_offset": 2048, 00:13:07.782 "data_size": 63488 00:13:07.782 } 00:13:07.782 ] 00:13:07.782 }' 00:13:07.782 17:34:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:07.782 17:34:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:13:07.782 17:34:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:07.782 17:34:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:13:07.782 17:34:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:13:07.782 17:34:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:07.782 17:34:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:07.782 17:34:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:07.782 17:34:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:07.782 17:34:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:07.782 17:34:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:07.782 17:34:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:07.782 17:34:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:07.782 17:34:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:07.782 17:34:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:07.782 17:34:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:07.782 17:34:38 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:07.782 17:34:38 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:07.782 17:34:38 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:07.782 17:34:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:07.782 "name": "raid_bdev1", 00:13:07.782 "uuid": "1bbbe449-6fda-4272-8c5b-fafa6297ec22", 00:13:07.782 "strip_size_kb": 0, 00:13:07.782 "state": "online", 00:13:07.782 "raid_level": "raid1", 00:13:07.782 "superblock": true, 00:13:07.782 "num_base_bdevs": 4, 00:13:07.783 "num_base_bdevs_discovered": 3, 00:13:07.783 "num_base_bdevs_operational": 3, 00:13:07.783 "base_bdevs_list": [ 00:13:07.783 { 00:13:07.783 "name": "spare", 00:13:07.783 "uuid": "4a1b1c6c-4e8e-5972-b73e-719c9e88c88f", 00:13:07.783 "is_configured": true, 00:13:07.783 "data_offset": 2048, 00:13:07.783 "data_size": 63488 00:13:07.783 }, 00:13:07.783 { 00:13:07.783 "name": null, 00:13:07.783 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:07.783 "is_configured": false, 00:13:07.783 "data_offset": 0, 00:13:07.783 "data_size": 63488 00:13:07.783 }, 00:13:07.783 { 00:13:07.783 "name": "BaseBdev3", 00:13:07.783 "uuid": "9b6eb3fc-dd72-5973-afbe-9ecd0fafecab", 00:13:07.783 "is_configured": true, 00:13:07.783 "data_offset": 2048, 00:13:07.783 "data_size": 63488 00:13:07.783 }, 00:13:07.783 { 00:13:07.783 "name": "BaseBdev4", 00:13:07.783 "uuid": "7dd75dc9-9415-55cc-bbe5-e0d8fdd0e6de", 00:13:07.783 "is_configured": true, 00:13:07.783 "data_offset": 2048, 00:13:07.783 "data_size": 63488 00:13:07.783 } 00:13:07.783 ] 00:13:07.783 }' 00:13:07.783 17:34:38 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:07.783 17:34:38 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:08.352 17:34:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:13:08.352 17:34:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:08.352 17:34:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:08.352 [2024-11-27 17:34:39.344500] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:13:08.352 [2024-11-27 17:34:39.344530] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:13:08.352 [2024-11-27 17:34:39.344612] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:13:08.352 [2024-11-27 17:34:39.344691] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:13:08.352 [2024-11-27 17:34:39.344705] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:13:08.352 17:34:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:08.352 17:34:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:08.352 17:34:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:08.352 17:34:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:08.352 17:34:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # jq length 00:13:08.352 17:34:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:08.352 17:34:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:13:08.352 17:34:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:13:08.352 17:34:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:13:08.353 17:34:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:13:08.353 17:34:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:13:08.353 17:34:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:13:08.353 17:34:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # local bdev_list 00:13:08.353 17:34:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:13:08.353 17:34:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # local nbd_list 00:13:08.353 17:34:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@12 -- # local i 00:13:08.353 17:34:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:13:08.353 17:34:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:13:08.353 17:34:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:13:08.612 /dev/nbd0 00:13:08.612 17:34:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:13:08.612 17:34:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:13:08.612 17:34:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:13:08.612 17:34:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@869 -- # local i 00:13:08.612 17:34:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:13:08.612 17:34:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:13:08.612 17:34:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:13:08.612 17:34:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@873 -- # break 00:13:08.612 17:34:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:13:08.612 17:34:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:13:08.612 17:34:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:13:08.612 1+0 records in 00:13:08.612 1+0 records out 00:13:08.612 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000529109 s, 7.7 MB/s 00:13:08.612 17:34:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:08.612 17:34:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@886 -- # size=4096 00:13:08.612 17:34:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:08.612 17:34:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:13:08.612 17:34:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@889 -- # return 0 00:13:08.612 17:34:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:13:08.612 17:34:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:13:08.612 17:34:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:13:08.871 /dev/nbd1 00:13:08.871 17:34:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:13:08.871 17:34:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:13:08.871 17:34:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@868 -- # local nbd_name=nbd1 00:13:08.871 17:34:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@869 -- # local i 00:13:08.871 17:34:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:13:08.871 17:34:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:13:08.871 17:34:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@872 -- # grep -q -w nbd1 /proc/partitions 00:13:08.871 17:34:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@873 -- # break 00:13:08.871 17:34:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:13:08.871 17:34:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:13:08.871 17:34:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@885 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:13:08.871 1+0 records in 00:13:08.871 1+0 records out 00:13:08.871 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000620358 s, 6.6 MB/s 00:13:08.871 17:34:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:08.871 17:34:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@886 -- # size=4096 00:13:08.871 17:34:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:08.871 17:34:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:13:08.871 17:34:39 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@889 -- # return 0 00:13:08.871 17:34:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:13:08.871 17:34:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:13:08.871 17:34:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@738 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:13:08.871 17:34:39 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:13:08.871 17:34:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:13:08.871 17:34:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:13:08.871 17:34:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # local nbd_list 00:13:08.871 17:34:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@51 -- # local i 00:13:08.871 17:34:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:13:08.871 17:34:39 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:13:09.131 17:34:40 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:13:09.131 17:34:40 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:13:09.131 17:34:40 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:13:09.131 17:34:40 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:13:09.131 17:34:40 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:13:09.131 17:34:40 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:13:09.131 17:34:40 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:13:09.131 17:34:40 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:13:09.131 17:34:40 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:13:09.131 17:34:40 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:13:09.392 17:34:40 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:13:09.392 17:34:40 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:13:09.392 17:34:40 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:13:09.392 17:34:40 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:13:09.392 17:34:40 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:13:09.392 17:34:40 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:13:09.392 17:34:40 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:13:09.392 17:34:40 bdev_raid.raid_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:13:09.392 17:34:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:13:09.392 17:34:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:13:09.392 17:34:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:09.392 17:34:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:09.392 17:34:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:09.392 17:34:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:13:09.392 17:34:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:09.392 17:34:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:09.392 [2024-11-27 17:34:40.439599] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:13:09.392 [2024-11-27 17:34:40.439662] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:09.392 [2024-11-27 17:34:40.439682] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ae80 00:13:09.392 [2024-11-27 17:34:40.439695] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:09.392 [2024-11-27 17:34:40.441764] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:09.392 [2024-11-27 17:34:40.441806] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:13:09.392 [2024-11-27 17:34:40.441898] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:13:09.392 [2024-11-27 17:34:40.441944] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:13:09.392 [2024-11-27 17:34:40.442045] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:13:09.392 [2024-11-27 17:34:40.442139] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:13:09.392 spare 00:13:09.392 17:34:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:09.392 17:34:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:13:09.392 17:34:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:09.392 17:34:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:09.392 [2024-11-27 17:34:40.542032] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001580 00:13:09.392 [2024-11-27 17:34:40.542061] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:13:09.392 [2024-11-27 17:34:40.542375] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000caeb00 00:13:09.392 [2024-11-27 17:34:40.542506] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001580 00:13:09.392 [2024-11-27 17:34:40.542523] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001580 00:13:09.392 [2024-11-27 17:34:40.542639] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:09.392 17:34:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:09.392 17:34:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:13:09.392 17:34:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:09.392 17:34:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:09.392 17:34:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:09.392 17:34:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:09.392 17:34:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:09.392 17:34:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:09.392 17:34:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:09.392 17:34:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:09.392 17:34:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:09.393 17:34:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:09.393 17:34:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:09.393 17:34:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:09.393 17:34:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:09.393 17:34:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:09.653 17:34:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:09.653 "name": "raid_bdev1", 00:13:09.653 "uuid": "1bbbe449-6fda-4272-8c5b-fafa6297ec22", 00:13:09.653 "strip_size_kb": 0, 00:13:09.653 "state": "online", 00:13:09.653 "raid_level": "raid1", 00:13:09.653 "superblock": true, 00:13:09.653 "num_base_bdevs": 4, 00:13:09.653 "num_base_bdevs_discovered": 3, 00:13:09.653 "num_base_bdevs_operational": 3, 00:13:09.653 "base_bdevs_list": [ 00:13:09.653 { 00:13:09.653 "name": "spare", 00:13:09.653 "uuid": "4a1b1c6c-4e8e-5972-b73e-719c9e88c88f", 00:13:09.653 "is_configured": true, 00:13:09.653 "data_offset": 2048, 00:13:09.653 "data_size": 63488 00:13:09.653 }, 00:13:09.653 { 00:13:09.653 "name": null, 00:13:09.653 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:09.653 "is_configured": false, 00:13:09.653 "data_offset": 2048, 00:13:09.653 "data_size": 63488 00:13:09.653 }, 00:13:09.653 { 00:13:09.653 "name": "BaseBdev3", 00:13:09.653 "uuid": "9b6eb3fc-dd72-5973-afbe-9ecd0fafecab", 00:13:09.653 "is_configured": true, 00:13:09.653 "data_offset": 2048, 00:13:09.653 "data_size": 63488 00:13:09.653 }, 00:13:09.653 { 00:13:09.653 "name": "BaseBdev4", 00:13:09.653 "uuid": "7dd75dc9-9415-55cc-bbe5-e0d8fdd0e6de", 00:13:09.653 "is_configured": true, 00:13:09.653 "data_offset": 2048, 00:13:09.653 "data_size": 63488 00:13:09.653 } 00:13:09.653 ] 00:13:09.653 }' 00:13:09.653 17:34:40 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:09.653 17:34:40 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:09.913 17:34:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:13:09.913 17:34:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:09.913 17:34:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:13:09.913 17:34:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:13:09.913 17:34:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:09.913 17:34:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:09.913 17:34:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:09.913 17:34:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:09.913 17:34:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:09.913 17:34:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:09.913 17:34:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:09.913 "name": "raid_bdev1", 00:13:09.913 "uuid": "1bbbe449-6fda-4272-8c5b-fafa6297ec22", 00:13:09.913 "strip_size_kb": 0, 00:13:09.913 "state": "online", 00:13:09.913 "raid_level": "raid1", 00:13:09.913 "superblock": true, 00:13:09.913 "num_base_bdevs": 4, 00:13:09.913 "num_base_bdevs_discovered": 3, 00:13:09.913 "num_base_bdevs_operational": 3, 00:13:09.913 "base_bdevs_list": [ 00:13:09.913 { 00:13:09.913 "name": "spare", 00:13:09.913 "uuid": "4a1b1c6c-4e8e-5972-b73e-719c9e88c88f", 00:13:09.913 "is_configured": true, 00:13:09.913 "data_offset": 2048, 00:13:09.913 "data_size": 63488 00:13:09.913 }, 00:13:09.913 { 00:13:09.913 "name": null, 00:13:09.913 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:09.913 "is_configured": false, 00:13:09.913 "data_offset": 2048, 00:13:09.913 "data_size": 63488 00:13:09.913 }, 00:13:09.913 { 00:13:09.913 "name": "BaseBdev3", 00:13:09.913 "uuid": "9b6eb3fc-dd72-5973-afbe-9ecd0fafecab", 00:13:09.913 "is_configured": true, 00:13:09.913 "data_offset": 2048, 00:13:09.913 "data_size": 63488 00:13:09.913 }, 00:13:09.913 { 00:13:09.913 "name": "BaseBdev4", 00:13:09.913 "uuid": "7dd75dc9-9415-55cc-bbe5-e0d8fdd0e6de", 00:13:09.913 "is_configured": true, 00:13:09.913 "data_offset": 2048, 00:13:09.913 "data_size": 63488 00:13:09.913 } 00:13:09.913 ] 00:13:09.913 }' 00:13:09.913 17:34:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:10.173 17:34:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:13:10.173 17:34:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:10.173 17:34:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:13:10.173 17:34:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:10.173 17:34:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:10.173 17:34:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:10.173 17:34:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:13:10.173 17:34:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:10.173 17:34:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:13:10.173 17:34:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:13:10.173 17:34:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:10.173 17:34:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:10.173 [2024-11-27 17:34:41.206312] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:10.173 17:34:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:10.173 17:34:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:13:10.173 17:34:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:10.173 17:34:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:10.173 17:34:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:10.173 17:34:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:10.173 17:34:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:10.173 17:34:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:10.173 17:34:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:10.173 17:34:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:10.173 17:34:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:10.173 17:34:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:10.173 17:34:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:10.173 17:34:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:10.173 17:34:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:10.173 17:34:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:10.173 17:34:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:10.173 "name": "raid_bdev1", 00:13:10.173 "uuid": "1bbbe449-6fda-4272-8c5b-fafa6297ec22", 00:13:10.173 "strip_size_kb": 0, 00:13:10.173 "state": "online", 00:13:10.173 "raid_level": "raid1", 00:13:10.173 "superblock": true, 00:13:10.173 "num_base_bdevs": 4, 00:13:10.173 "num_base_bdevs_discovered": 2, 00:13:10.173 "num_base_bdevs_operational": 2, 00:13:10.173 "base_bdevs_list": [ 00:13:10.173 { 00:13:10.173 "name": null, 00:13:10.173 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:10.173 "is_configured": false, 00:13:10.173 "data_offset": 0, 00:13:10.173 "data_size": 63488 00:13:10.173 }, 00:13:10.173 { 00:13:10.173 "name": null, 00:13:10.173 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:10.173 "is_configured": false, 00:13:10.173 "data_offset": 2048, 00:13:10.173 "data_size": 63488 00:13:10.173 }, 00:13:10.173 { 00:13:10.173 "name": "BaseBdev3", 00:13:10.173 "uuid": "9b6eb3fc-dd72-5973-afbe-9ecd0fafecab", 00:13:10.173 "is_configured": true, 00:13:10.173 "data_offset": 2048, 00:13:10.173 "data_size": 63488 00:13:10.173 }, 00:13:10.173 { 00:13:10.173 "name": "BaseBdev4", 00:13:10.173 "uuid": "7dd75dc9-9415-55cc-bbe5-e0d8fdd0e6de", 00:13:10.173 "is_configured": true, 00:13:10.173 "data_offset": 2048, 00:13:10.173 "data_size": 63488 00:13:10.173 } 00:13:10.173 ] 00:13:10.173 }' 00:13:10.173 17:34:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:10.173 17:34:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:10.743 17:34:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:13:10.743 17:34:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:10.743 17:34:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:10.743 [2024-11-27 17:34:41.685616] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:13:10.743 [2024-11-27 17:34:41.685829] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (5) smaller than existing raid bdev raid_bdev1 (6) 00:13:10.743 [2024-11-27 17:34:41.685894] bdev_raid.c:3748:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:13:10.743 [2024-11-27 17:34:41.685962] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:13:10.743 [2024-11-27 17:34:41.689101] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000caebd0 00:13:10.743 [2024-11-27 17:34:41.691042] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:13:10.743 17:34:41 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:10.743 17:34:41 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@757 -- # sleep 1 00:13:11.681 17:34:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:11.682 17:34:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:11.682 17:34:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:11.682 17:34:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:11.682 17:34:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:11.682 17:34:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:11.682 17:34:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:11.682 17:34:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:11.682 17:34:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:11.682 17:34:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:11.682 17:34:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:11.682 "name": "raid_bdev1", 00:13:11.682 "uuid": "1bbbe449-6fda-4272-8c5b-fafa6297ec22", 00:13:11.682 "strip_size_kb": 0, 00:13:11.682 "state": "online", 00:13:11.682 "raid_level": "raid1", 00:13:11.682 "superblock": true, 00:13:11.682 "num_base_bdevs": 4, 00:13:11.682 "num_base_bdevs_discovered": 3, 00:13:11.682 "num_base_bdevs_operational": 3, 00:13:11.682 "process": { 00:13:11.682 "type": "rebuild", 00:13:11.682 "target": "spare", 00:13:11.682 "progress": { 00:13:11.682 "blocks": 20480, 00:13:11.682 "percent": 32 00:13:11.682 } 00:13:11.682 }, 00:13:11.682 "base_bdevs_list": [ 00:13:11.682 { 00:13:11.682 "name": "spare", 00:13:11.682 "uuid": "4a1b1c6c-4e8e-5972-b73e-719c9e88c88f", 00:13:11.682 "is_configured": true, 00:13:11.682 "data_offset": 2048, 00:13:11.682 "data_size": 63488 00:13:11.682 }, 00:13:11.682 { 00:13:11.682 "name": null, 00:13:11.682 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:11.682 "is_configured": false, 00:13:11.682 "data_offset": 2048, 00:13:11.682 "data_size": 63488 00:13:11.682 }, 00:13:11.682 { 00:13:11.682 "name": "BaseBdev3", 00:13:11.682 "uuid": "9b6eb3fc-dd72-5973-afbe-9ecd0fafecab", 00:13:11.682 "is_configured": true, 00:13:11.682 "data_offset": 2048, 00:13:11.682 "data_size": 63488 00:13:11.682 }, 00:13:11.682 { 00:13:11.682 "name": "BaseBdev4", 00:13:11.682 "uuid": "7dd75dc9-9415-55cc-bbe5-e0d8fdd0e6de", 00:13:11.682 "is_configured": true, 00:13:11.682 "data_offset": 2048, 00:13:11.682 "data_size": 63488 00:13:11.682 } 00:13:11.682 ] 00:13:11.682 }' 00:13:11.682 17:34:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:11.682 17:34:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:11.682 17:34:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:11.682 17:34:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:11.682 17:34:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:13:11.682 17:34:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:11.682 17:34:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:11.682 [2024-11-27 17:34:42.857748] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:11.942 [2024-11-27 17:34:42.894990] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:13:11.942 [2024-11-27 17:34:42.895094] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:11.942 [2024-11-27 17:34:42.895136] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:11.942 [2024-11-27 17:34:42.895167] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:13:11.942 17:34:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:11.942 17:34:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:13:11.942 17:34:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:11.942 17:34:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:11.942 17:34:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:11.942 17:34:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:11.942 17:34:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:11.942 17:34:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:11.942 17:34:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:11.942 17:34:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:11.942 17:34:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:11.942 17:34:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:11.942 17:34:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:11.942 17:34:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:11.942 17:34:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:11.942 17:34:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:11.942 17:34:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:11.942 "name": "raid_bdev1", 00:13:11.942 "uuid": "1bbbe449-6fda-4272-8c5b-fafa6297ec22", 00:13:11.942 "strip_size_kb": 0, 00:13:11.942 "state": "online", 00:13:11.942 "raid_level": "raid1", 00:13:11.942 "superblock": true, 00:13:11.942 "num_base_bdevs": 4, 00:13:11.942 "num_base_bdevs_discovered": 2, 00:13:11.942 "num_base_bdevs_operational": 2, 00:13:11.942 "base_bdevs_list": [ 00:13:11.942 { 00:13:11.942 "name": null, 00:13:11.942 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:11.942 "is_configured": false, 00:13:11.942 "data_offset": 0, 00:13:11.942 "data_size": 63488 00:13:11.942 }, 00:13:11.942 { 00:13:11.942 "name": null, 00:13:11.942 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:11.942 "is_configured": false, 00:13:11.942 "data_offset": 2048, 00:13:11.942 "data_size": 63488 00:13:11.942 }, 00:13:11.942 { 00:13:11.942 "name": "BaseBdev3", 00:13:11.942 "uuid": "9b6eb3fc-dd72-5973-afbe-9ecd0fafecab", 00:13:11.942 "is_configured": true, 00:13:11.942 "data_offset": 2048, 00:13:11.942 "data_size": 63488 00:13:11.942 }, 00:13:11.942 { 00:13:11.942 "name": "BaseBdev4", 00:13:11.942 "uuid": "7dd75dc9-9415-55cc-bbe5-e0d8fdd0e6de", 00:13:11.942 "is_configured": true, 00:13:11.942 "data_offset": 2048, 00:13:11.942 "data_size": 63488 00:13:11.942 } 00:13:11.942 ] 00:13:11.942 }' 00:13:11.942 17:34:42 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:11.942 17:34:42 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:12.211 17:34:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:13:12.211 17:34:43 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:12.211 17:34:43 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:12.211 [2024-11-27 17:34:43.373831] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:13:12.211 [2024-11-27 17:34:43.373886] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:12.211 [2024-11-27 17:34:43.373908] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b780 00:13:12.211 [2024-11-27 17:34:43.373918] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:12.211 [2024-11-27 17:34:43.374312] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:12.211 [2024-11-27 17:34:43.374332] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:13:12.211 [2024-11-27 17:34:43.374400] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:13:12.212 [2024-11-27 17:34:43.374416] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (5) smaller than existing raid bdev raid_bdev1 (6) 00:13:12.212 [2024-11-27 17:34:43.374424] bdev_raid.c:3748:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:13:12.212 [2024-11-27 17:34:43.374448] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:13:12.212 [2024-11-27 17:34:43.376966] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000caeca0 00:13:12.212 [2024-11-27 17:34:43.378716] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:13:12.212 spare 00:13:12.212 17:34:43 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:12.212 17:34:43 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@764 -- # sleep 1 00:13:13.198 17:34:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:13.198 17:34:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:13.198 17:34:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:13.198 17:34:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:13.198 17:34:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:13.458 17:34:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:13.458 17:34:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:13.458 17:34:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:13.458 17:34:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:13.458 17:34:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:13.458 17:34:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:13.458 "name": "raid_bdev1", 00:13:13.458 "uuid": "1bbbe449-6fda-4272-8c5b-fafa6297ec22", 00:13:13.458 "strip_size_kb": 0, 00:13:13.458 "state": "online", 00:13:13.458 "raid_level": "raid1", 00:13:13.458 "superblock": true, 00:13:13.458 "num_base_bdevs": 4, 00:13:13.458 "num_base_bdevs_discovered": 3, 00:13:13.458 "num_base_bdevs_operational": 3, 00:13:13.458 "process": { 00:13:13.458 "type": "rebuild", 00:13:13.458 "target": "spare", 00:13:13.458 "progress": { 00:13:13.458 "blocks": 20480, 00:13:13.458 "percent": 32 00:13:13.458 } 00:13:13.458 }, 00:13:13.458 "base_bdevs_list": [ 00:13:13.458 { 00:13:13.458 "name": "spare", 00:13:13.458 "uuid": "4a1b1c6c-4e8e-5972-b73e-719c9e88c88f", 00:13:13.458 "is_configured": true, 00:13:13.458 "data_offset": 2048, 00:13:13.458 "data_size": 63488 00:13:13.458 }, 00:13:13.458 { 00:13:13.458 "name": null, 00:13:13.458 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:13.458 "is_configured": false, 00:13:13.458 "data_offset": 2048, 00:13:13.458 "data_size": 63488 00:13:13.458 }, 00:13:13.458 { 00:13:13.458 "name": "BaseBdev3", 00:13:13.458 "uuid": "9b6eb3fc-dd72-5973-afbe-9ecd0fafecab", 00:13:13.458 "is_configured": true, 00:13:13.458 "data_offset": 2048, 00:13:13.458 "data_size": 63488 00:13:13.458 }, 00:13:13.458 { 00:13:13.458 "name": "BaseBdev4", 00:13:13.458 "uuid": "7dd75dc9-9415-55cc-bbe5-e0d8fdd0e6de", 00:13:13.458 "is_configured": true, 00:13:13.458 "data_offset": 2048, 00:13:13.458 "data_size": 63488 00:13:13.458 } 00:13:13.458 ] 00:13:13.458 }' 00:13:13.458 17:34:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:13.458 17:34:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:13.458 17:34:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:13.458 17:34:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:13.458 17:34:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:13:13.458 17:34:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:13.458 17:34:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:13.458 [2024-11-27 17:34:44.529429] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:13.458 [2024-11-27 17:34:44.582632] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:13:13.458 [2024-11-27 17:34:44.582685] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:13.458 [2024-11-27 17:34:44.582703] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:13.458 [2024-11-27 17:34:44.582711] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:13:13.458 17:34:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:13.458 17:34:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:13:13.458 17:34:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:13.458 17:34:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:13.458 17:34:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:13.458 17:34:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:13.458 17:34:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:13.458 17:34:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:13.458 17:34:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:13.458 17:34:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:13.458 17:34:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:13.458 17:34:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:13.458 17:34:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:13.458 17:34:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:13.458 17:34:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:13.458 17:34:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:13.458 17:34:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:13.458 "name": "raid_bdev1", 00:13:13.458 "uuid": "1bbbe449-6fda-4272-8c5b-fafa6297ec22", 00:13:13.458 "strip_size_kb": 0, 00:13:13.458 "state": "online", 00:13:13.458 "raid_level": "raid1", 00:13:13.458 "superblock": true, 00:13:13.458 "num_base_bdevs": 4, 00:13:13.458 "num_base_bdevs_discovered": 2, 00:13:13.458 "num_base_bdevs_operational": 2, 00:13:13.458 "base_bdevs_list": [ 00:13:13.458 { 00:13:13.458 "name": null, 00:13:13.458 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:13.458 "is_configured": false, 00:13:13.458 "data_offset": 0, 00:13:13.458 "data_size": 63488 00:13:13.458 }, 00:13:13.458 { 00:13:13.458 "name": null, 00:13:13.458 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:13.458 "is_configured": false, 00:13:13.458 "data_offset": 2048, 00:13:13.458 "data_size": 63488 00:13:13.458 }, 00:13:13.458 { 00:13:13.458 "name": "BaseBdev3", 00:13:13.458 "uuid": "9b6eb3fc-dd72-5973-afbe-9ecd0fafecab", 00:13:13.458 "is_configured": true, 00:13:13.458 "data_offset": 2048, 00:13:13.458 "data_size": 63488 00:13:13.458 }, 00:13:13.458 { 00:13:13.458 "name": "BaseBdev4", 00:13:13.458 "uuid": "7dd75dc9-9415-55cc-bbe5-e0d8fdd0e6de", 00:13:13.458 "is_configured": true, 00:13:13.458 "data_offset": 2048, 00:13:13.458 "data_size": 63488 00:13:13.458 } 00:13:13.458 ] 00:13:13.458 }' 00:13:13.458 17:34:44 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:13.458 17:34:44 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:14.029 17:34:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:13:14.029 17:34:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:14.029 17:34:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:13:14.029 17:34:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:13:14.029 17:34:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:14.029 17:34:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:14.029 17:34:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:14.029 17:34:45 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:14.029 17:34:45 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:14.029 17:34:45 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:14.029 17:34:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:14.029 "name": "raid_bdev1", 00:13:14.029 "uuid": "1bbbe449-6fda-4272-8c5b-fafa6297ec22", 00:13:14.029 "strip_size_kb": 0, 00:13:14.029 "state": "online", 00:13:14.029 "raid_level": "raid1", 00:13:14.029 "superblock": true, 00:13:14.029 "num_base_bdevs": 4, 00:13:14.029 "num_base_bdevs_discovered": 2, 00:13:14.029 "num_base_bdevs_operational": 2, 00:13:14.029 "base_bdevs_list": [ 00:13:14.029 { 00:13:14.029 "name": null, 00:13:14.029 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:14.029 "is_configured": false, 00:13:14.029 "data_offset": 0, 00:13:14.029 "data_size": 63488 00:13:14.029 }, 00:13:14.029 { 00:13:14.029 "name": null, 00:13:14.029 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:14.029 "is_configured": false, 00:13:14.029 "data_offset": 2048, 00:13:14.029 "data_size": 63488 00:13:14.029 }, 00:13:14.029 { 00:13:14.029 "name": "BaseBdev3", 00:13:14.029 "uuid": "9b6eb3fc-dd72-5973-afbe-9ecd0fafecab", 00:13:14.029 "is_configured": true, 00:13:14.029 "data_offset": 2048, 00:13:14.029 "data_size": 63488 00:13:14.029 }, 00:13:14.029 { 00:13:14.029 "name": "BaseBdev4", 00:13:14.029 "uuid": "7dd75dc9-9415-55cc-bbe5-e0d8fdd0e6de", 00:13:14.029 "is_configured": true, 00:13:14.029 "data_offset": 2048, 00:13:14.029 "data_size": 63488 00:13:14.029 } 00:13:14.029 ] 00:13:14.029 }' 00:13:14.029 17:34:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:14.029 17:34:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:13:14.029 17:34:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:14.029 17:34:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:13:14.029 17:34:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:13:14.029 17:34:45 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:14.029 17:34:45 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:14.029 17:34:45 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:14.029 17:34:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:13:14.029 17:34:45 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:14.029 17:34:45 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:14.029 [2024-11-27 17:34:45.193190] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:13:14.029 [2024-11-27 17:34:45.193289] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:14.029 [2024-11-27 17:34:45.193316] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000bd80 00:13:14.029 [2024-11-27 17:34:45.193325] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:14.029 [2024-11-27 17:34:45.193693] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:14.029 [2024-11-27 17:34:45.193717] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:13:14.029 [2024-11-27 17:34:45.193783] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:13:14.029 [2024-11-27 17:34:45.193796] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (6) 00:13:14.029 [2024-11-27 17:34:45.193816] bdev_raid.c:3709:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:13:14.029 [2024-11-27 17:34:45.193826] bdev_raid.c:3884:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:13:14.029 BaseBdev1 00:13:14.029 17:34:45 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:14.029 17:34:45 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@775 -- # sleep 1 00:13:15.412 17:34:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:13:15.412 17:34:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:15.412 17:34:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:15.412 17:34:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:15.412 17:34:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:15.412 17:34:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:15.412 17:34:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:15.412 17:34:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:15.412 17:34:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:15.412 17:34:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:15.412 17:34:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:15.412 17:34:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:15.412 17:34:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:15.412 17:34:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:15.412 17:34:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:15.412 17:34:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:15.412 "name": "raid_bdev1", 00:13:15.412 "uuid": "1bbbe449-6fda-4272-8c5b-fafa6297ec22", 00:13:15.412 "strip_size_kb": 0, 00:13:15.412 "state": "online", 00:13:15.412 "raid_level": "raid1", 00:13:15.412 "superblock": true, 00:13:15.412 "num_base_bdevs": 4, 00:13:15.412 "num_base_bdevs_discovered": 2, 00:13:15.412 "num_base_bdevs_operational": 2, 00:13:15.412 "base_bdevs_list": [ 00:13:15.412 { 00:13:15.412 "name": null, 00:13:15.412 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:15.412 "is_configured": false, 00:13:15.412 "data_offset": 0, 00:13:15.412 "data_size": 63488 00:13:15.412 }, 00:13:15.412 { 00:13:15.412 "name": null, 00:13:15.412 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:15.412 "is_configured": false, 00:13:15.412 "data_offset": 2048, 00:13:15.412 "data_size": 63488 00:13:15.412 }, 00:13:15.412 { 00:13:15.412 "name": "BaseBdev3", 00:13:15.412 "uuid": "9b6eb3fc-dd72-5973-afbe-9ecd0fafecab", 00:13:15.412 "is_configured": true, 00:13:15.412 "data_offset": 2048, 00:13:15.412 "data_size": 63488 00:13:15.412 }, 00:13:15.412 { 00:13:15.412 "name": "BaseBdev4", 00:13:15.412 "uuid": "7dd75dc9-9415-55cc-bbe5-e0d8fdd0e6de", 00:13:15.412 "is_configured": true, 00:13:15.412 "data_offset": 2048, 00:13:15.412 "data_size": 63488 00:13:15.412 } 00:13:15.412 ] 00:13:15.412 }' 00:13:15.412 17:34:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:15.412 17:34:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:15.673 17:34:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:13:15.673 17:34:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:15.673 17:34:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:13:15.673 17:34:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:13:15.673 17:34:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:15.673 17:34:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:15.673 17:34:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:15.673 17:34:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:15.673 17:34:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:15.673 17:34:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:15.673 17:34:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:15.673 "name": "raid_bdev1", 00:13:15.673 "uuid": "1bbbe449-6fda-4272-8c5b-fafa6297ec22", 00:13:15.673 "strip_size_kb": 0, 00:13:15.673 "state": "online", 00:13:15.673 "raid_level": "raid1", 00:13:15.673 "superblock": true, 00:13:15.673 "num_base_bdevs": 4, 00:13:15.673 "num_base_bdevs_discovered": 2, 00:13:15.673 "num_base_bdevs_operational": 2, 00:13:15.673 "base_bdevs_list": [ 00:13:15.673 { 00:13:15.673 "name": null, 00:13:15.673 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:15.673 "is_configured": false, 00:13:15.673 "data_offset": 0, 00:13:15.673 "data_size": 63488 00:13:15.673 }, 00:13:15.673 { 00:13:15.673 "name": null, 00:13:15.673 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:15.673 "is_configured": false, 00:13:15.673 "data_offset": 2048, 00:13:15.673 "data_size": 63488 00:13:15.673 }, 00:13:15.673 { 00:13:15.673 "name": "BaseBdev3", 00:13:15.673 "uuid": "9b6eb3fc-dd72-5973-afbe-9ecd0fafecab", 00:13:15.673 "is_configured": true, 00:13:15.673 "data_offset": 2048, 00:13:15.673 "data_size": 63488 00:13:15.673 }, 00:13:15.673 { 00:13:15.673 "name": "BaseBdev4", 00:13:15.673 "uuid": "7dd75dc9-9415-55cc-bbe5-e0d8fdd0e6de", 00:13:15.673 "is_configured": true, 00:13:15.673 "data_offset": 2048, 00:13:15.673 "data_size": 63488 00:13:15.673 } 00:13:15.673 ] 00:13:15.673 }' 00:13:15.673 17:34:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:15.673 17:34:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:13:15.673 17:34:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:15.673 17:34:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:13:15.673 17:34:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:13:15.673 17:34:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@650 -- # local es=0 00:13:15.673 17:34:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:13:15.673 17:34:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:13:15.673 17:34:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:13:15.673 17:34:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:13:15.673 17:34:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:13:15.673 17:34:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:13:15.673 17:34:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:15.673 17:34:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:15.673 [2024-11-27 17:34:46.762852] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:13:15.673 [2024-11-27 17:34:46.763022] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (6) 00:13:15.673 [2024-11-27 17:34:46.763078] bdev_raid.c:3709:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:13:15.673 request: 00:13:15.673 { 00:13:15.673 "base_bdev": "BaseBdev1", 00:13:15.673 "raid_bdev": "raid_bdev1", 00:13:15.673 "method": "bdev_raid_add_base_bdev", 00:13:15.673 "req_id": 1 00:13:15.673 } 00:13:15.673 Got JSON-RPC error response 00:13:15.673 response: 00:13:15.673 { 00:13:15.673 "code": -22, 00:13:15.673 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:13:15.673 } 00:13:15.673 17:34:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:13:15.673 17:34:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@653 -- # es=1 00:13:15.673 17:34:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:13:15.673 17:34:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:13:15.673 17:34:46 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:13:15.673 17:34:46 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@779 -- # sleep 1 00:13:16.613 17:34:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:13:16.613 17:34:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:16.613 17:34:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:16.613 17:34:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:16.613 17:34:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:16.613 17:34:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:16.613 17:34:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:16.613 17:34:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:16.613 17:34:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:16.613 17:34:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:16.613 17:34:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:16.613 17:34:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:16.613 17:34:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:16.613 17:34:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:16.613 17:34:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:16.873 17:34:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:16.873 "name": "raid_bdev1", 00:13:16.873 "uuid": "1bbbe449-6fda-4272-8c5b-fafa6297ec22", 00:13:16.873 "strip_size_kb": 0, 00:13:16.873 "state": "online", 00:13:16.873 "raid_level": "raid1", 00:13:16.873 "superblock": true, 00:13:16.873 "num_base_bdevs": 4, 00:13:16.873 "num_base_bdevs_discovered": 2, 00:13:16.873 "num_base_bdevs_operational": 2, 00:13:16.873 "base_bdevs_list": [ 00:13:16.873 { 00:13:16.873 "name": null, 00:13:16.873 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:16.873 "is_configured": false, 00:13:16.873 "data_offset": 0, 00:13:16.873 "data_size": 63488 00:13:16.873 }, 00:13:16.873 { 00:13:16.873 "name": null, 00:13:16.873 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:16.873 "is_configured": false, 00:13:16.873 "data_offset": 2048, 00:13:16.873 "data_size": 63488 00:13:16.873 }, 00:13:16.873 { 00:13:16.873 "name": "BaseBdev3", 00:13:16.873 "uuid": "9b6eb3fc-dd72-5973-afbe-9ecd0fafecab", 00:13:16.873 "is_configured": true, 00:13:16.873 "data_offset": 2048, 00:13:16.873 "data_size": 63488 00:13:16.873 }, 00:13:16.873 { 00:13:16.873 "name": "BaseBdev4", 00:13:16.873 "uuid": "7dd75dc9-9415-55cc-bbe5-e0d8fdd0e6de", 00:13:16.873 "is_configured": true, 00:13:16.873 "data_offset": 2048, 00:13:16.873 "data_size": 63488 00:13:16.873 } 00:13:16.873 ] 00:13:16.873 }' 00:13:16.873 17:34:47 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:16.873 17:34:47 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:17.134 17:34:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:13:17.134 17:34:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:17.134 17:34:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:13:17.134 17:34:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:13:17.134 17:34:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:17.134 17:34:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:17.134 17:34:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:17.134 17:34:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:17.134 17:34:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:17.134 17:34:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:17.394 17:34:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:17.394 "name": "raid_bdev1", 00:13:17.394 "uuid": "1bbbe449-6fda-4272-8c5b-fafa6297ec22", 00:13:17.394 "strip_size_kb": 0, 00:13:17.394 "state": "online", 00:13:17.394 "raid_level": "raid1", 00:13:17.394 "superblock": true, 00:13:17.394 "num_base_bdevs": 4, 00:13:17.394 "num_base_bdevs_discovered": 2, 00:13:17.394 "num_base_bdevs_operational": 2, 00:13:17.394 "base_bdevs_list": [ 00:13:17.394 { 00:13:17.394 "name": null, 00:13:17.394 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:17.394 "is_configured": false, 00:13:17.394 "data_offset": 0, 00:13:17.394 "data_size": 63488 00:13:17.394 }, 00:13:17.394 { 00:13:17.394 "name": null, 00:13:17.394 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:17.394 "is_configured": false, 00:13:17.394 "data_offset": 2048, 00:13:17.394 "data_size": 63488 00:13:17.394 }, 00:13:17.394 { 00:13:17.394 "name": "BaseBdev3", 00:13:17.394 "uuid": "9b6eb3fc-dd72-5973-afbe-9ecd0fafecab", 00:13:17.394 "is_configured": true, 00:13:17.394 "data_offset": 2048, 00:13:17.394 "data_size": 63488 00:13:17.394 }, 00:13:17.394 { 00:13:17.394 "name": "BaseBdev4", 00:13:17.394 "uuid": "7dd75dc9-9415-55cc-bbe5-e0d8fdd0e6de", 00:13:17.394 "is_configured": true, 00:13:17.394 "data_offset": 2048, 00:13:17.394 "data_size": 63488 00:13:17.394 } 00:13:17.394 ] 00:13:17.394 }' 00:13:17.394 17:34:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:17.394 17:34:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:13:17.394 17:34:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:17.394 17:34:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:13:17.394 17:34:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@784 -- # killprocess 88470 00:13:17.394 17:34:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@950 -- # '[' -z 88470 ']' 00:13:17.394 17:34:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@954 -- # kill -0 88470 00:13:17.394 17:34:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@955 -- # uname 00:13:17.394 17:34:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:13:17.395 17:34:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 88470 00:13:17.395 17:34:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:13:17.395 killing process with pid 88470 00:13:17.395 Received shutdown signal, test time was about 60.000000 seconds 00:13:17.395 00:13:17.395 Latency(us) 00:13:17.395 [2024-11-27T17:34:48.587Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:13:17.395 [2024-11-27T17:34:48.587Z] =================================================================================================================== 00:13:17.395 [2024-11-27T17:34:48.587Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:13:17.395 17:34:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:13:17.395 17:34:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@968 -- # echo 'killing process with pid 88470' 00:13:17.395 17:34:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@969 -- # kill 88470 00:13:17.395 [2024-11-27 17:34:48.469367] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:13:17.395 [2024-11-27 17:34:48.469474] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:13:17.395 [2024-11-27 17:34:48.469534] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:13:17.395 [2024-11-27 17:34:48.469545] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state offline 00:13:17.395 17:34:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@974 -- # wait 88470 00:13:17.395 [2024-11-27 17:34:48.520086] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:13:17.661 17:34:48 bdev_raid.raid_rebuild_test_sb -- bdev/bdev_raid.sh@786 -- # return 0 00:13:17.661 00:13:17.661 real 0m23.575s 00:13:17.661 user 0m28.846s 00:13:17.661 sys 0m4.035s 00:13:17.661 ************************************ 00:13:17.661 END TEST raid_rebuild_test_sb 00:13:17.661 ************************************ 00:13:17.661 17:34:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@1126 -- # xtrace_disable 00:13:17.661 17:34:48 bdev_raid.raid_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:17.661 17:34:48 bdev_raid -- bdev/bdev_raid.sh@980 -- # run_test raid_rebuild_test_io raid_rebuild_test raid1 4 false true true 00:13:17.661 17:34:48 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 7 -le 1 ']' 00:13:17.661 17:34:48 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:13:17.661 17:34:48 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:13:17.662 ************************************ 00:13:17.662 START TEST raid_rebuild_test_io 00:13:17.662 ************************************ 00:13:17.662 17:34:48 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@1125 -- # raid_rebuild_test raid1 4 false true true 00:13:17.662 17:34:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:13:17.662 17:34:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=4 00:13:17.662 17:34:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@571 -- # local superblock=false 00:13:17.662 17:34:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@572 -- # local background_io=true 00:13:17.662 17:34:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@573 -- # local verify=true 00:13:17.662 17:34:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:13:17.662 17:34:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:13:17.662 17:34:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:13:17.662 17:34:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:13:17.662 17:34:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:13:17.662 17:34:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:13:17.662 17:34:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:13:17.662 17:34:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:13:17.662 17:34:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev3 00:13:17.662 17:34:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:13:17.662 17:34:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:13:17.662 17:34:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev4 00:13:17.928 17:34:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:13:17.928 17:34:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:13:17.928 17:34:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:13:17.928 17:34:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:13:17.928 17:34:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:13:17.928 17:34:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@576 -- # local strip_size 00:13:17.928 17:34:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@577 -- # local create_arg 00:13:17.928 17:34:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:13:17.928 17:34:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@579 -- # local data_offset 00:13:17.929 17:34:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:13:17.929 17:34:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:13:17.929 17:34:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@592 -- # '[' false = true ']' 00:13:17.929 17:34:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@597 -- # raid_pid=89207 00:13:17.929 17:34:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:13:17.929 17:34:48 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@598 -- # waitforlisten 89207 00:13:17.929 17:34:48 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@831 -- # '[' -z 89207 ']' 00:13:17.929 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:13:17.929 17:34:48 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:13:17.929 17:34:48 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@836 -- # local max_retries=100 00:13:17.929 17:34:48 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:13:17.929 17:34:48 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@840 -- # xtrace_disable 00:13:17.929 17:34:48 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:17.929 I/O size of 3145728 is greater than zero copy threshold (65536). 00:13:17.929 Zero copy mechanism will not be used. 00:13:17.929 [2024-11-27 17:34:48.946026] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:13:17.929 [2024-11-27 17:34:48.946164] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid89207 ] 00:13:17.929 [2024-11-27 17:34:49.093132] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:13:18.188 [2024-11-27 17:34:49.138773] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:13:18.188 [2024-11-27 17:34:49.181342] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:13:18.188 [2024-11-27 17:34:49.181461] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:13:18.759 17:34:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:13:18.759 17:34:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@864 -- # return 0 00:13:18.759 17:34:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:13:18.759 17:34:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:13:18.759 17:34:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:18.759 17:34:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:18.759 BaseBdev1_malloc 00:13:18.759 17:34:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:18.759 17:34:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:13:18.759 17:34:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:18.759 17:34:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:18.759 [2024-11-27 17:34:49.779526] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:13:18.759 [2024-11-27 17:34:49.779642] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:18.759 [2024-11-27 17:34:49.779669] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:13:18.759 [2024-11-27 17:34:49.779694] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:18.759 [2024-11-27 17:34:49.781691] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:18.759 [2024-11-27 17:34:49.781731] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:13:18.759 BaseBdev1 00:13:18.759 17:34:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:18.759 17:34:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:13:18.759 17:34:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:13:18.759 17:34:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:18.759 17:34:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:18.759 BaseBdev2_malloc 00:13:18.759 17:34:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:18.759 17:34:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:13:18.759 17:34:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:18.759 17:34:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:18.759 [2024-11-27 17:34:49.824994] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:13:18.759 [2024-11-27 17:34:49.825094] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:18.759 [2024-11-27 17:34:49.825175] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:13:18.759 [2024-11-27 17:34:49.825200] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:18.759 [2024-11-27 17:34:49.829595] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:18.759 [2024-11-27 17:34:49.829649] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:13:18.759 BaseBdev2 00:13:18.759 17:34:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:18.759 17:34:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:13:18.759 17:34:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:13:18.759 17:34:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:18.759 17:34:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:18.759 BaseBdev3_malloc 00:13:18.759 17:34:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:18.759 17:34:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev3_malloc -p BaseBdev3 00:13:18.759 17:34:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:18.759 17:34:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:18.759 [2024-11-27 17:34:49.855801] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev3_malloc 00:13:18.759 [2024-11-27 17:34:49.855856] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:18.759 [2024-11-27 17:34:49.855882] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:13:18.759 [2024-11-27 17:34:49.855891] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:18.759 [2024-11-27 17:34:49.857856] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:18.759 [2024-11-27 17:34:49.857933] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:13:18.759 BaseBdev3 00:13:18.759 17:34:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:18.759 17:34:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:13:18.759 17:34:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:13:18.759 17:34:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:18.759 17:34:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:18.759 BaseBdev4_malloc 00:13:18.759 17:34:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:18.759 17:34:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev4_malloc -p BaseBdev4 00:13:18.759 17:34:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:18.759 17:34:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:18.759 [2024-11-27 17:34:49.884613] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev4_malloc 00:13:18.759 [2024-11-27 17:34:49.884658] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:18.759 [2024-11-27 17:34:49.884678] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:13:18.759 [2024-11-27 17:34:49.884685] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:18.759 [2024-11-27 17:34:49.886597] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:18.759 [2024-11-27 17:34:49.886631] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:13:18.759 BaseBdev4 00:13:18.759 17:34:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:18.759 17:34:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:13:18.759 17:34:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:18.759 17:34:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:18.759 spare_malloc 00:13:18.759 17:34:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:18.759 17:34:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:13:18.759 17:34:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:18.759 17:34:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:18.759 spare_delay 00:13:18.759 17:34:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:18.759 17:34:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:13:18.759 17:34:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:18.759 17:34:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:18.759 [2024-11-27 17:34:49.925084] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:13:18.759 [2024-11-27 17:34:49.925127] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:18.760 [2024-11-27 17:34:49.925158] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009c80 00:13:18.760 [2024-11-27 17:34:49.925167] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:18.760 [2024-11-27 17:34:49.927164] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:18.760 [2024-11-27 17:34:49.927196] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:13:18.760 spare 00:13:18.760 17:34:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:18.760 17:34:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 00:13:18.760 17:34:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:18.760 17:34:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:18.760 [2024-11-27 17:34:49.937135] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:13:18.760 [2024-11-27 17:34:49.938858] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:13:18.760 [2024-11-27 17:34:49.938919] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:13:18.760 [2024-11-27 17:34:49.938964] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:13:18.760 [2024-11-27 17:34:49.939035] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:13:18.760 [2024-11-27 17:34:49.939043] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 65536, blocklen 512 00:13:18.760 [2024-11-27 17:34:49.939290] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:13:18.760 [2024-11-27 17:34:49.939401] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:13:18.760 [2024-11-27 17:34:49.939412] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:13:18.760 [2024-11-27 17:34:49.939530] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:18.760 17:34:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:18.760 17:34:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:13:18.760 17:34:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:18.760 17:34:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:18.760 17:34:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:18.760 17:34:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:18.760 17:34:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:13:18.760 17:34:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:18.760 17:34:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:18.760 17:34:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:18.760 17:34:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:19.020 17:34:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:19.020 17:34:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:19.020 17:34:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:19.020 17:34:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:19.020 17:34:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:19.020 17:34:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:19.020 "name": "raid_bdev1", 00:13:19.020 "uuid": "83975ea4-b67b-4ad7-848b-5098630eeec8", 00:13:19.020 "strip_size_kb": 0, 00:13:19.020 "state": "online", 00:13:19.020 "raid_level": "raid1", 00:13:19.020 "superblock": false, 00:13:19.020 "num_base_bdevs": 4, 00:13:19.020 "num_base_bdevs_discovered": 4, 00:13:19.020 "num_base_bdevs_operational": 4, 00:13:19.020 "base_bdevs_list": [ 00:13:19.020 { 00:13:19.020 "name": "BaseBdev1", 00:13:19.020 "uuid": "d63f3038-0325-55a4-9dba-39657c264f7e", 00:13:19.020 "is_configured": true, 00:13:19.020 "data_offset": 0, 00:13:19.020 "data_size": 65536 00:13:19.020 }, 00:13:19.020 { 00:13:19.020 "name": "BaseBdev2", 00:13:19.020 "uuid": "b3650f48-0cae-5bac-add3-c51f5ed56b6d", 00:13:19.020 "is_configured": true, 00:13:19.020 "data_offset": 0, 00:13:19.020 "data_size": 65536 00:13:19.020 }, 00:13:19.020 { 00:13:19.020 "name": "BaseBdev3", 00:13:19.020 "uuid": "e4a3dacf-a371-52c5-a492-68182055a7d6", 00:13:19.020 "is_configured": true, 00:13:19.020 "data_offset": 0, 00:13:19.020 "data_size": 65536 00:13:19.020 }, 00:13:19.020 { 00:13:19.020 "name": "BaseBdev4", 00:13:19.020 "uuid": "23318573-4c35-5b68-bacd-cefdb3a7cc1a", 00:13:19.020 "is_configured": true, 00:13:19.020 "data_offset": 0, 00:13:19.020 "data_size": 65536 00:13:19.020 } 00:13:19.020 ] 00:13:19.020 }' 00:13:19.020 17:34:49 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:19.020 17:34:49 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:19.279 17:34:50 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:13:19.279 17:34:50 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:19.280 17:34:50 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:19.280 17:34:50 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:13:19.280 [2024-11-27 17:34:50.388607] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:13:19.280 17:34:50 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:19.280 17:34:50 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=65536 00:13:19.280 17:34:50 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:19.280 17:34:50 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:13:19.280 17:34:50 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:19.280 17:34:50 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:19.280 17:34:50 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:19.540 17:34:50 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@619 -- # data_offset=0 00:13:19.540 17:34:50 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@621 -- # '[' true = true ']' 00:13:19.540 17:34:50 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:13:19.540 17:34:50 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@623 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:13:19.540 17:34:50 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:19.540 17:34:50 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:19.540 [2024-11-27 17:34:50.484225] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:13:19.540 17:34:50 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:19.540 17:34:50 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:13:19.540 17:34:50 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:19.540 17:34:50 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:19.540 17:34:50 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:19.540 17:34:50 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:19.540 17:34:50 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:19.540 17:34:50 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:19.540 17:34:50 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:19.540 17:34:50 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:19.540 17:34:50 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:19.540 17:34:50 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:19.540 17:34:50 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:19.540 17:34:50 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:19.540 17:34:50 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:19.540 17:34:50 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:19.540 17:34:50 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:19.540 "name": "raid_bdev1", 00:13:19.540 "uuid": "83975ea4-b67b-4ad7-848b-5098630eeec8", 00:13:19.540 "strip_size_kb": 0, 00:13:19.540 "state": "online", 00:13:19.540 "raid_level": "raid1", 00:13:19.540 "superblock": false, 00:13:19.540 "num_base_bdevs": 4, 00:13:19.540 "num_base_bdevs_discovered": 3, 00:13:19.540 "num_base_bdevs_operational": 3, 00:13:19.540 "base_bdevs_list": [ 00:13:19.540 { 00:13:19.540 "name": null, 00:13:19.540 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:19.540 "is_configured": false, 00:13:19.540 "data_offset": 0, 00:13:19.540 "data_size": 65536 00:13:19.540 }, 00:13:19.540 { 00:13:19.540 "name": "BaseBdev2", 00:13:19.540 "uuid": "b3650f48-0cae-5bac-add3-c51f5ed56b6d", 00:13:19.540 "is_configured": true, 00:13:19.540 "data_offset": 0, 00:13:19.540 "data_size": 65536 00:13:19.540 }, 00:13:19.540 { 00:13:19.540 "name": "BaseBdev3", 00:13:19.540 "uuid": "e4a3dacf-a371-52c5-a492-68182055a7d6", 00:13:19.540 "is_configured": true, 00:13:19.540 "data_offset": 0, 00:13:19.540 "data_size": 65536 00:13:19.540 }, 00:13:19.540 { 00:13:19.540 "name": "BaseBdev4", 00:13:19.540 "uuid": "23318573-4c35-5b68-bacd-cefdb3a7cc1a", 00:13:19.540 "is_configured": true, 00:13:19.540 "data_offset": 0, 00:13:19.540 "data_size": 65536 00:13:19.540 } 00:13:19.540 ] 00:13:19.540 }' 00:13:19.540 17:34:50 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:19.541 17:34:50 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:19.541 [2024-11-27 17:34:50.574099] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002870 00:13:19.541 I/O size of 3145728 is greater than zero copy threshold (65536). 00:13:19.541 Zero copy mechanism will not be used. 00:13:19.541 Running I/O for 60 seconds... 00:13:19.801 17:34:50 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:13:19.801 17:34:50 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:19.801 17:34:50 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:19.801 [2024-11-27 17:34:50.935765] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:13:19.801 17:34:50 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:19.801 17:34:50 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@647 -- # sleep 1 00:13:19.801 [2024-11-27 17:34:50.981569] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002940 00:13:19.801 [2024-11-27 17:34:50.983520] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:13:20.061 [2024-11-27 17:34:51.111054] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:13:20.061 [2024-11-27 17:34:51.112420] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:13:20.321 [2024-11-27 17:34:51.322209] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:13:20.321 [2024-11-27 17:34:51.322549] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:13:20.581 170.00 IOPS, 510.00 MiB/s [2024-11-27T17:34:51.773Z] [2024-11-27 17:34:51.634221] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:13:20.581 [2024-11-27 17:34:51.756946] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:13:20.581 [2024-11-27 17:34:51.757126] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:13:20.841 17:34:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:20.841 17:34:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:20.841 17:34:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:20.841 17:34:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:20.841 17:34:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:20.841 17:34:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:20.841 17:34:51 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:20.841 17:34:51 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:20.841 17:34:51 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:20.841 [2024-11-27 17:34:52.005555] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 14336 offset_begin: 12288 offset_end: 18432 00:13:20.841 17:34:52 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:20.841 17:34:52 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:20.841 "name": "raid_bdev1", 00:13:20.841 "uuid": "83975ea4-b67b-4ad7-848b-5098630eeec8", 00:13:20.841 "strip_size_kb": 0, 00:13:20.841 "state": "online", 00:13:20.841 "raid_level": "raid1", 00:13:20.841 "superblock": false, 00:13:20.841 "num_base_bdevs": 4, 00:13:20.841 "num_base_bdevs_discovered": 4, 00:13:20.841 "num_base_bdevs_operational": 4, 00:13:20.841 "process": { 00:13:20.841 "type": "rebuild", 00:13:20.841 "target": "spare", 00:13:20.841 "progress": { 00:13:20.841 "blocks": 12288, 00:13:20.841 "percent": 18 00:13:20.841 } 00:13:20.841 }, 00:13:20.841 "base_bdevs_list": [ 00:13:20.841 { 00:13:20.841 "name": "spare", 00:13:20.841 "uuid": "4d960f8d-8c22-588c-bf06-97e590c629ad", 00:13:20.841 "is_configured": true, 00:13:20.841 "data_offset": 0, 00:13:20.841 "data_size": 65536 00:13:20.841 }, 00:13:20.841 { 00:13:20.841 "name": "BaseBdev2", 00:13:20.841 "uuid": "b3650f48-0cae-5bac-add3-c51f5ed56b6d", 00:13:20.841 "is_configured": true, 00:13:20.841 "data_offset": 0, 00:13:20.841 "data_size": 65536 00:13:20.841 }, 00:13:20.841 { 00:13:20.841 "name": "BaseBdev3", 00:13:20.841 "uuid": "e4a3dacf-a371-52c5-a492-68182055a7d6", 00:13:20.841 "is_configured": true, 00:13:20.841 "data_offset": 0, 00:13:20.841 "data_size": 65536 00:13:20.841 }, 00:13:20.841 { 00:13:20.841 "name": "BaseBdev4", 00:13:20.841 "uuid": "23318573-4c35-5b68-bacd-cefdb3a7cc1a", 00:13:20.841 "is_configured": true, 00:13:20.841 "data_offset": 0, 00:13:20.841 "data_size": 65536 00:13:20.841 } 00:13:20.841 ] 00:13:20.841 }' 00:13:21.101 17:34:52 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:21.101 17:34:52 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:21.101 17:34:52 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:21.101 17:34:52 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:21.101 17:34:52 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:13:21.101 17:34:52 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:21.101 17:34:52 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:21.101 [2024-11-27 17:34:52.137220] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:21.101 [2024-11-27 17:34:52.222265] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:13:21.361 [2024-11-27 17:34:52.329337] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:13:21.361 [2024-11-27 17:34:52.332582] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:21.361 [2024-11-27 17:34:52.332682] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:21.361 [2024-11-27 17:34:52.332700] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:13:21.361 [2024-11-27 17:34:52.349658] bdev_raid.c:1970:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 0 raid_ch: 0x60d000002870 00:13:21.361 17:34:52 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:21.361 17:34:52 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:13:21.361 17:34:52 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:21.361 17:34:52 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:21.361 17:34:52 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:21.361 17:34:52 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:21.361 17:34:52 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:21.361 17:34:52 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:21.361 17:34:52 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:21.361 17:34:52 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:21.361 17:34:52 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:21.361 17:34:52 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:21.361 17:34:52 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:21.361 17:34:52 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:21.361 17:34:52 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:21.361 17:34:52 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:21.361 17:34:52 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:21.361 "name": "raid_bdev1", 00:13:21.361 "uuid": "83975ea4-b67b-4ad7-848b-5098630eeec8", 00:13:21.361 "strip_size_kb": 0, 00:13:21.361 "state": "online", 00:13:21.361 "raid_level": "raid1", 00:13:21.361 "superblock": false, 00:13:21.361 "num_base_bdevs": 4, 00:13:21.361 "num_base_bdevs_discovered": 3, 00:13:21.361 "num_base_bdevs_operational": 3, 00:13:21.361 "base_bdevs_list": [ 00:13:21.361 { 00:13:21.361 "name": null, 00:13:21.361 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:21.361 "is_configured": false, 00:13:21.361 "data_offset": 0, 00:13:21.361 "data_size": 65536 00:13:21.361 }, 00:13:21.361 { 00:13:21.361 "name": "BaseBdev2", 00:13:21.361 "uuid": "b3650f48-0cae-5bac-add3-c51f5ed56b6d", 00:13:21.361 "is_configured": true, 00:13:21.361 "data_offset": 0, 00:13:21.361 "data_size": 65536 00:13:21.361 }, 00:13:21.361 { 00:13:21.361 "name": "BaseBdev3", 00:13:21.361 "uuid": "e4a3dacf-a371-52c5-a492-68182055a7d6", 00:13:21.361 "is_configured": true, 00:13:21.361 "data_offset": 0, 00:13:21.361 "data_size": 65536 00:13:21.361 }, 00:13:21.361 { 00:13:21.361 "name": "BaseBdev4", 00:13:21.361 "uuid": "23318573-4c35-5b68-bacd-cefdb3a7cc1a", 00:13:21.361 "is_configured": true, 00:13:21.361 "data_offset": 0, 00:13:21.361 "data_size": 65536 00:13:21.361 } 00:13:21.361 ] 00:13:21.361 }' 00:13:21.361 17:34:52 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:21.361 17:34:52 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:21.881 158.00 IOPS, 474.00 MiB/s [2024-11-27T17:34:53.073Z] 17:34:52 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:13:21.881 17:34:52 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:21.881 17:34:52 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:13:21.881 17:34:52 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:13:21.881 17:34:52 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:21.881 17:34:52 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:21.881 17:34:52 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:21.881 17:34:52 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:21.881 17:34:52 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:21.881 17:34:52 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:21.881 17:34:52 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:21.881 "name": "raid_bdev1", 00:13:21.881 "uuid": "83975ea4-b67b-4ad7-848b-5098630eeec8", 00:13:21.881 "strip_size_kb": 0, 00:13:21.881 "state": "online", 00:13:21.881 "raid_level": "raid1", 00:13:21.881 "superblock": false, 00:13:21.881 "num_base_bdevs": 4, 00:13:21.881 "num_base_bdevs_discovered": 3, 00:13:21.881 "num_base_bdevs_operational": 3, 00:13:21.881 "base_bdevs_list": [ 00:13:21.881 { 00:13:21.881 "name": null, 00:13:21.881 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:21.881 "is_configured": false, 00:13:21.881 "data_offset": 0, 00:13:21.881 "data_size": 65536 00:13:21.881 }, 00:13:21.881 { 00:13:21.881 "name": "BaseBdev2", 00:13:21.881 "uuid": "b3650f48-0cae-5bac-add3-c51f5ed56b6d", 00:13:21.881 "is_configured": true, 00:13:21.881 "data_offset": 0, 00:13:21.881 "data_size": 65536 00:13:21.881 }, 00:13:21.881 { 00:13:21.881 "name": "BaseBdev3", 00:13:21.881 "uuid": "e4a3dacf-a371-52c5-a492-68182055a7d6", 00:13:21.881 "is_configured": true, 00:13:21.881 "data_offset": 0, 00:13:21.881 "data_size": 65536 00:13:21.881 }, 00:13:21.881 { 00:13:21.881 "name": "BaseBdev4", 00:13:21.881 "uuid": "23318573-4c35-5b68-bacd-cefdb3a7cc1a", 00:13:21.881 "is_configured": true, 00:13:21.881 "data_offset": 0, 00:13:21.881 "data_size": 65536 00:13:21.881 } 00:13:21.881 ] 00:13:21.881 }' 00:13:21.881 17:34:52 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:21.881 17:34:52 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:13:21.881 17:34:52 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:21.881 17:34:52 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:13:21.882 17:34:52 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:13:21.882 17:34:52 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:21.882 17:34:52 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:21.882 [2024-11-27 17:34:52.943397] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:13:21.882 17:34:52 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:21.882 17:34:52 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@663 -- # sleep 1 00:13:21.882 [2024-11-27 17:34:52.996482] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002a10 00:13:21.882 [2024-11-27 17:34:52.998410] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:13:22.142 [2024-11-27 17:34:53.105982] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:13:22.142 [2024-11-27 17:34:53.106324] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:13:22.401 [2024-11-27 17:34:53.328893] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:13:22.401 [2024-11-27 17:34:53.329281] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:13:22.662 171.00 IOPS, 513.00 MiB/s [2024-11-27T17:34:53.854Z] [2024-11-27 17:34:53.650234] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:13:22.662 [2024-11-27 17:34:53.778419] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:13:22.920 17:34:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:22.920 17:34:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:22.920 17:34:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:22.920 17:34:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:22.920 17:34:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:22.920 17:34:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:22.920 17:34:53 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:22.920 17:34:53 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:22.920 17:34:53 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:22.920 17:34:54 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:22.920 [2024-11-27 17:34:54.007183] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 14336 offset_begin: 12288 offset_end: 18432 00:13:22.920 17:34:54 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:22.920 "name": "raid_bdev1", 00:13:22.920 "uuid": "83975ea4-b67b-4ad7-848b-5098630eeec8", 00:13:22.920 "strip_size_kb": 0, 00:13:22.920 "state": "online", 00:13:22.920 "raid_level": "raid1", 00:13:22.920 "superblock": false, 00:13:22.920 "num_base_bdevs": 4, 00:13:22.920 "num_base_bdevs_discovered": 4, 00:13:22.920 "num_base_bdevs_operational": 4, 00:13:22.920 "process": { 00:13:22.920 "type": "rebuild", 00:13:22.920 "target": "spare", 00:13:22.920 "progress": { 00:13:22.920 "blocks": 12288, 00:13:22.920 "percent": 18 00:13:22.920 } 00:13:22.920 }, 00:13:22.920 "base_bdevs_list": [ 00:13:22.920 { 00:13:22.920 "name": "spare", 00:13:22.920 "uuid": "4d960f8d-8c22-588c-bf06-97e590c629ad", 00:13:22.920 "is_configured": true, 00:13:22.920 "data_offset": 0, 00:13:22.920 "data_size": 65536 00:13:22.920 }, 00:13:22.920 { 00:13:22.920 "name": "BaseBdev2", 00:13:22.920 "uuid": "b3650f48-0cae-5bac-add3-c51f5ed56b6d", 00:13:22.920 "is_configured": true, 00:13:22.920 "data_offset": 0, 00:13:22.920 "data_size": 65536 00:13:22.920 }, 00:13:22.920 { 00:13:22.920 "name": "BaseBdev3", 00:13:22.920 "uuid": "e4a3dacf-a371-52c5-a492-68182055a7d6", 00:13:22.920 "is_configured": true, 00:13:22.920 "data_offset": 0, 00:13:22.920 "data_size": 65536 00:13:22.920 }, 00:13:22.920 { 00:13:22.920 "name": "BaseBdev4", 00:13:22.920 "uuid": "23318573-4c35-5b68-bacd-cefdb3a7cc1a", 00:13:22.920 "is_configured": true, 00:13:22.920 "data_offset": 0, 00:13:22.920 "data_size": 65536 00:13:22.920 } 00:13:22.920 ] 00:13:22.920 }' 00:13:22.921 17:34:54 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:22.921 17:34:54 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:22.921 17:34:54 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:22.921 17:34:54 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:22.921 17:34:54 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@666 -- # '[' false = true ']' 00:13:22.921 17:34:54 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=4 00:13:22.921 17:34:54 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:13:22.921 17:34:54 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@693 -- # '[' 4 -gt 2 ']' 00:13:22.921 17:34:54 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@695 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:13:22.921 17:34:54 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:22.921 17:34:54 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:22.921 [2024-11-27 17:34:54.109450] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:13:23.180 [2024-11-27 17:34:54.111731] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:13:23.180 [2024-11-27 17:34:54.331838] bdev_raid.c:1970:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 1 raid_ch: 0x60d000002870 00:13:23.180 [2024-11-27 17:34:54.331926] bdev_raid.c:1970:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 1 raid_ch: 0x60d000002a10 00:13:23.180 17:34:54 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:23.180 17:34:54 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@698 -- # base_bdevs[1]= 00:13:23.180 17:34:54 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@699 -- # (( num_base_bdevs_operational-- )) 00:13:23.180 17:34:54 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@702 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:23.180 17:34:54 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:23.180 17:34:54 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:23.180 17:34:54 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:23.180 17:34:54 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:23.180 17:34:54 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:23.180 17:34:54 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:23.180 17:34:54 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:23.180 17:34:54 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:23.440 17:34:54 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:23.441 17:34:54 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:23.441 "name": "raid_bdev1", 00:13:23.441 "uuid": "83975ea4-b67b-4ad7-848b-5098630eeec8", 00:13:23.441 "strip_size_kb": 0, 00:13:23.441 "state": "online", 00:13:23.441 "raid_level": "raid1", 00:13:23.441 "superblock": false, 00:13:23.441 "num_base_bdevs": 4, 00:13:23.441 "num_base_bdevs_discovered": 3, 00:13:23.441 "num_base_bdevs_operational": 3, 00:13:23.441 "process": { 00:13:23.441 "type": "rebuild", 00:13:23.441 "target": "spare", 00:13:23.441 "progress": { 00:13:23.441 "blocks": 18432, 00:13:23.441 "percent": 28 00:13:23.441 } 00:13:23.441 }, 00:13:23.441 "base_bdevs_list": [ 00:13:23.441 { 00:13:23.441 "name": "spare", 00:13:23.441 "uuid": "4d960f8d-8c22-588c-bf06-97e590c629ad", 00:13:23.441 "is_configured": true, 00:13:23.441 "data_offset": 0, 00:13:23.441 "data_size": 65536 00:13:23.441 }, 00:13:23.441 { 00:13:23.441 "name": null, 00:13:23.441 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:23.441 "is_configured": false, 00:13:23.441 "data_offset": 0, 00:13:23.441 "data_size": 65536 00:13:23.441 }, 00:13:23.441 { 00:13:23.441 "name": "BaseBdev3", 00:13:23.441 "uuid": "e4a3dacf-a371-52c5-a492-68182055a7d6", 00:13:23.441 "is_configured": true, 00:13:23.441 "data_offset": 0, 00:13:23.441 "data_size": 65536 00:13:23.441 }, 00:13:23.441 { 00:13:23.441 "name": "BaseBdev4", 00:13:23.441 "uuid": "23318573-4c35-5b68-bacd-cefdb3a7cc1a", 00:13:23.441 "is_configured": true, 00:13:23.441 "data_offset": 0, 00:13:23.441 "data_size": 65536 00:13:23.441 } 00:13:23.441 ] 00:13:23.441 }' 00:13:23.441 17:34:54 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:23.441 17:34:54 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:23.441 17:34:54 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:23.441 [2024-11-27 17:34:54.457410] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 20480 offset_begin: 18432 offset_end: 24576 00:13:23.441 17:34:54 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:23.441 17:34:54 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@706 -- # local timeout=402 00:13:23.441 17:34:54 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:23.441 17:34:54 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:23.441 17:34:54 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:23.441 17:34:54 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:23.441 17:34:54 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:23.441 17:34:54 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:23.441 17:34:54 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:23.441 17:34:54 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:23.441 17:34:54 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:23.441 17:34:54 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:23.441 17:34:54 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:23.441 17:34:54 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:23.441 "name": "raid_bdev1", 00:13:23.441 "uuid": "83975ea4-b67b-4ad7-848b-5098630eeec8", 00:13:23.441 "strip_size_kb": 0, 00:13:23.441 "state": "online", 00:13:23.441 "raid_level": "raid1", 00:13:23.441 "superblock": false, 00:13:23.441 "num_base_bdevs": 4, 00:13:23.441 "num_base_bdevs_discovered": 3, 00:13:23.441 "num_base_bdevs_operational": 3, 00:13:23.441 "process": { 00:13:23.441 "type": "rebuild", 00:13:23.441 "target": "spare", 00:13:23.441 "progress": { 00:13:23.441 "blocks": 20480, 00:13:23.441 "percent": 31 00:13:23.441 } 00:13:23.441 }, 00:13:23.441 "base_bdevs_list": [ 00:13:23.441 { 00:13:23.441 "name": "spare", 00:13:23.441 "uuid": "4d960f8d-8c22-588c-bf06-97e590c629ad", 00:13:23.441 "is_configured": true, 00:13:23.441 "data_offset": 0, 00:13:23.441 "data_size": 65536 00:13:23.441 }, 00:13:23.441 { 00:13:23.441 "name": null, 00:13:23.441 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:23.441 "is_configured": false, 00:13:23.441 "data_offset": 0, 00:13:23.441 "data_size": 65536 00:13:23.441 }, 00:13:23.441 { 00:13:23.441 "name": "BaseBdev3", 00:13:23.441 "uuid": "e4a3dacf-a371-52c5-a492-68182055a7d6", 00:13:23.441 "is_configured": true, 00:13:23.441 "data_offset": 0, 00:13:23.441 "data_size": 65536 00:13:23.441 }, 00:13:23.441 { 00:13:23.441 "name": "BaseBdev4", 00:13:23.441 "uuid": "23318573-4c35-5b68-bacd-cefdb3a7cc1a", 00:13:23.441 "is_configured": true, 00:13:23.441 "data_offset": 0, 00:13:23.441 "data_size": 65536 00:13:23.441 } 00:13:23.441 ] 00:13:23.441 }' 00:13:23.441 17:34:54 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:23.441 17:34:54 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:23.441 17:34:54 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:23.441 17:34:54 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:23.441 17:34:54 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:13:23.701 152.75 IOPS, 458.25 MiB/s [2024-11-27T17:34:54.893Z] [2024-11-27 17:34:54.673163] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 22528 offset_begin: 18432 offset_end: 24576 00:13:23.701 [2024-11-27 17:34:54.673669] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 22528 offset_begin: 18432 offset_end: 24576 00:13:23.961 [2024-11-27 17:34:55.002195] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 26624 offset_begin: 24576 offset_end: 30720 00:13:24.221 [2024-11-27 17:34:55.227310] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 28672 offset_begin: 24576 offset_end: 30720 00:13:24.481 17:34:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:24.481 17:34:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:24.481 17:34:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:24.481 17:34:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:24.481 17:34:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:24.481 17:34:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:24.481 17:34:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:24.481 17:34:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:24.481 17:34:55 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:24.481 17:34:55 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:24.481 132.00 IOPS, 396.00 MiB/s [2024-11-27T17:34:55.673Z] 17:34:55 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:24.482 17:34:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:24.482 "name": "raid_bdev1", 00:13:24.482 "uuid": "83975ea4-b67b-4ad7-848b-5098630eeec8", 00:13:24.482 "strip_size_kb": 0, 00:13:24.482 "state": "online", 00:13:24.482 "raid_level": "raid1", 00:13:24.482 "superblock": false, 00:13:24.482 "num_base_bdevs": 4, 00:13:24.482 "num_base_bdevs_discovered": 3, 00:13:24.482 "num_base_bdevs_operational": 3, 00:13:24.482 "process": { 00:13:24.482 "type": "rebuild", 00:13:24.482 "target": "spare", 00:13:24.482 "progress": { 00:13:24.482 "blocks": 34816, 00:13:24.482 "percent": 53 00:13:24.482 } 00:13:24.482 }, 00:13:24.482 "base_bdevs_list": [ 00:13:24.482 { 00:13:24.482 "name": "spare", 00:13:24.482 "uuid": "4d960f8d-8c22-588c-bf06-97e590c629ad", 00:13:24.482 "is_configured": true, 00:13:24.482 "data_offset": 0, 00:13:24.482 "data_size": 65536 00:13:24.482 }, 00:13:24.482 { 00:13:24.482 "name": null, 00:13:24.482 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:24.482 "is_configured": false, 00:13:24.482 "data_offset": 0, 00:13:24.482 "data_size": 65536 00:13:24.482 }, 00:13:24.482 { 00:13:24.482 "name": "BaseBdev3", 00:13:24.482 "uuid": "e4a3dacf-a371-52c5-a492-68182055a7d6", 00:13:24.482 "is_configured": true, 00:13:24.482 "data_offset": 0, 00:13:24.482 "data_size": 65536 00:13:24.482 }, 00:13:24.482 { 00:13:24.482 "name": "BaseBdev4", 00:13:24.482 "uuid": "23318573-4c35-5b68-bacd-cefdb3a7cc1a", 00:13:24.482 "is_configured": true, 00:13:24.482 "data_offset": 0, 00:13:24.482 "data_size": 65536 00:13:24.482 } 00:13:24.482 ] 00:13:24.482 }' 00:13:24.482 17:34:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:24.482 17:34:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:24.482 17:34:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:24.742 17:34:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:24.742 17:34:55 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:13:25.003 [2024-11-27 17:34:56.093016] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 45056 offset_begin: 43008 offset_end: 49152 00:13:25.262 [2024-11-27 17:34:56.200875] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 47104 offset_begin: 43008 offset_end: 49152 00:13:25.522 115.83 IOPS, 347.50 MiB/s [2024-11-27T17:34:56.714Z] [2024-11-27 17:34:56.611207] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 53248 offset_begin: 49152 offset_end: 55296 00:13:25.782 17:34:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:25.782 17:34:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:25.782 17:34:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:25.782 17:34:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:25.782 17:34:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:25.782 17:34:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:25.782 17:34:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:25.782 17:34:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:25.782 17:34:56 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:25.782 17:34:56 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:25.782 17:34:56 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:25.782 17:34:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:25.782 "name": "raid_bdev1", 00:13:25.782 "uuid": "83975ea4-b67b-4ad7-848b-5098630eeec8", 00:13:25.782 "strip_size_kb": 0, 00:13:25.782 "state": "online", 00:13:25.782 "raid_level": "raid1", 00:13:25.782 "superblock": false, 00:13:25.782 "num_base_bdevs": 4, 00:13:25.782 "num_base_bdevs_discovered": 3, 00:13:25.782 "num_base_bdevs_operational": 3, 00:13:25.782 "process": { 00:13:25.782 "type": "rebuild", 00:13:25.782 "target": "spare", 00:13:25.782 "progress": { 00:13:25.782 "blocks": 55296, 00:13:25.782 "percent": 84 00:13:25.782 } 00:13:25.782 }, 00:13:25.782 "base_bdevs_list": [ 00:13:25.782 { 00:13:25.783 "name": "spare", 00:13:25.783 "uuid": "4d960f8d-8c22-588c-bf06-97e590c629ad", 00:13:25.783 "is_configured": true, 00:13:25.783 "data_offset": 0, 00:13:25.783 "data_size": 65536 00:13:25.783 }, 00:13:25.783 { 00:13:25.783 "name": null, 00:13:25.783 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:25.783 "is_configured": false, 00:13:25.783 "data_offset": 0, 00:13:25.783 "data_size": 65536 00:13:25.783 }, 00:13:25.783 { 00:13:25.783 "name": "BaseBdev3", 00:13:25.783 "uuid": "e4a3dacf-a371-52c5-a492-68182055a7d6", 00:13:25.783 "is_configured": true, 00:13:25.783 "data_offset": 0, 00:13:25.783 "data_size": 65536 00:13:25.783 }, 00:13:25.783 { 00:13:25.783 "name": "BaseBdev4", 00:13:25.783 "uuid": "23318573-4c35-5b68-bacd-cefdb3a7cc1a", 00:13:25.783 "is_configured": true, 00:13:25.783 "data_offset": 0, 00:13:25.783 "data_size": 65536 00:13:25.783 } 00:13:25.783 ] 00:13:25.783 }' 00:13:25.783 17:34:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:25.783 17:34:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:25.783 17:34:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:25.783 [2024-11-27 17:34:56.825671] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 57344 offset_begin: 55296 offset_end: 61440 00:13:25.783 17:34:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:25.783 17:34:56 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:13:25.783 [2024-11-27 17:34:56.935495] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 59392 offset_begin: 55296 offset_end: 61440 00:13:25.783 [2024-11-27 17:34:56.935709] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 59392 offset_begin: 55296 offset_end: 61440 00:13:26.352 [2024-11-27 17:34:57.271691] bdev_raid.c:2896:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:13:26.352 [2024-11-27 17:34:57.376527] bdev_raid.c:2558:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:13:26.352 [2024-11-27 17:34:57.379564] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:26.873 104.43 IOPS, 313.29 MiB/s [2024-11-27T17:34:58.065Z] 17:34:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:26.873 17:34:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:26.873 17:34:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:26.873 17:34:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:26.873 17:34:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:26.873 17:34:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:26.873 17:34:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:26.873 17:34:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:26.873 17:34:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:26.873 17:34:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:26.873 17:34:57 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:26.873 17:34:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:26.873 "name": "raid_bdev1", 00:13:26.873 "uuid": "83975ea4-b67b-4ad7-848b-5098630eeec8", 00:13:26.873 "strip_size_kb": 0, 00:13:26.873 "state": "online", 00:13:26.873 "raid_level": "raid1", 00:13:26.873 "superblock": false, 00:13:26.873 "num_base_bdevs": 4, 00:13:26.873 "num_base_bdevs_discovered": 3, 00:13:26.873 "num_base_bdevs_operational": 3, 00:13:26.873 "base_bdevs_list": [ 00:13:26.873 { 00:13:26.873 "name": "spare", 00:13:26.873 "uuid": "4d960f8d-8c22-588c-bf06-97e590c629ad", 00:13:26.873 "is_configured": true, 00:13:26.873 "data_offset": 0, 00:13:26.873 "data_size": 65536 00:13:26.873 }, 00:13:26.873 { 00:13:26.873 "name": null, 00:13:26.873 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:26.873 "is_configured": false, 00:13:26.873 "data_offset": 0, 00:13:26.873 "data_size": 65536 00:13:26.873 }, 00:13:26.873 { 00:13:26.873 "name": "BaseBdev3", 00:13:26.873 "uuid": "e4a3dacf-a371-52c5-a492-68182055a7d6", 00:13:26.873 "is_configured": true, 00:13:26.873 "data_offset": 0, 00:13:26.873 "data_size": 65536 00:13:26.873 }, 00:13:26.873 { 00:13:26.873 "name": "BaseBdev4", 00:13:26.873 "uuid": "23318573-4c35-5b68-bacd-cefdb3a7cc1a", 00:13:26.873 "is_configured": true, 00:13:26.873 "data_offset": 0, 00:13:26.873 "data_size": 65536 00:13:26.873 } 00:13:26.873 ] 00:13:26.873 }' 00:13:26.873 17:34:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:26.873 17:34:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:13:26.873 17:34:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:26.873 17:34:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:13:26.873 17:34:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@709 -- # break 00:13:26.873 17:34:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:13:26.873 17:34:57 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:26.873 17:34:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:13:26.873 17:34:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:13:26.873 17:34:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:26.873 17:34:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:26.873 17:34:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:26.873 17:34:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:26.873 17:34:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:26.873 17:34:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:26.873 17:34:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:26.873 "name": "raid_bdev1", 00:13:26.873 "uuid": "83975ea4-b67b-4ad7-848b-5098630eeec8", 00:13:26.873 "strip_size_kb": 0, 00:13:26.873 "state": "online", 00:13:26.873 "raid_level": "raid1", 00:13:26.873 "superblock": false, 00:13:26.873 "num_base_bdevs": 4, 00:13:26.873 "num_base_bdevs_discovered": 3, 00:13:26.873 "num_base_bdevs_operational": 3, 00:13:26.873 "base_bdevs_list": [ 00:13:26.873 { 00:13:26.873 "name": "spare", 00:13:26.873 "uuid": "4d960f8d-8c22-588c-bf06-97e590c629ad", 00:13:26.873 "is_configured": true, 00:13:26.873 "data_offset": 0, 00:13:26.873 "data_size": 65536 00:13:26.873 }, 00:13:26.873 { 00:13:26.873 "name": null, 00:13:26.873 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:26.873 "is_configured": false, 00:13:26.873 "data_offset": 0, 00:13:26.873 "data_size": 65536 00:13:26.873 }, 00:13:26.873 { 00:13:26.873 "name": "BaseBdev3", 00:13:26.873 "uuid": "e4a3dacf-a371-52c5-a492-68182055a7d6", 00:13:26.873 "is_configured": true, 00:13:26.873 "data_offset": 0, 00:13:26.873 "data_size": 65536 00:13:26.873 }, 00:13:26.873 { 00:13:26.873 "name": "BaseBdev4", 00:13:26.873 "uuid": "23318573-4c35-5b68-bacd-cefdb3a7cc1a", 00:13:26.873 "is_configured": true, 00:13:26.873 "data_offset": 0, 00:13:26.873 "data_size": 65536 00:13:26.873 } 00:13:26.873 ] 00:13:26.873 }' 00:13:26.873 17:34:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:27.133 17:34:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:13:27.133 17:34:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:27.133 17:34:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:13:27.133 17:34:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:13:27.133 17:34:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:27.133 17:34:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:27.133 17:34:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:27.133 17:34:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:27.133 17:34:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:27.134 17:34:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:27.134 17:34:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:27.134 17:34:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:27.134 17:34:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:27.134 17:34:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:27.134 17:34:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:27.134 17:34:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:27.134 17:34:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:27.134 17:34:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:27.134 17:34:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:27.134 "name": "raid_bdev1", 00:13:27.134 "uuid": "83975ea4-b67b-4ad7-848b-5098630eeec8", 00:13:27.134 "strip_size_kb": 0, 00:13:27.134 "state": "online", 00:13:27.134 "raid_level": "raid1", 00:13:27.134 "superblock": false, 00:13:27.134 "num_base_bdevs": 4, 00:13:27.134 "num_base_bdevs_discovered": 3, 00:13:27.134 "num_base_bdevs_operational": 3, 00:13:27.134 "base_bdevs_list": [ 00:13:27.134 { 00:13:27.134 "name": "spare", 00:13:27.134 "uuid": "4d960f8d-8c22-588c-bf06-97e590c629ad", 00:13:27.134 "is_configured": true, 00:13:27.134 "data_offset": 0, 00:13:27.134 "data_size": 65536 00:13:27.134 }, 00:13:27.134 { 00:13:27.134 "name": null, 00:13:27.134 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:27.134 "is_configured": false, 00:13:27.134 "data_offset": 0, 00:13:27.134 "data_size": 65536 00:13:27.134 }, 00:13:27.134 { 00:13:27.134 "name": "BaseBdev3", 00:13:27.134 "uuid": "e4a3dacf-a371-52c5-a492-68182055a7d6", 00:13:27.134 "is_configured": true, 00:13:27.134 "data_offset": 0, 00:13:27.134 "data_size": 65536 00:13:27.134 }, 00:13:27.134 { 00:13:27.134 "name": "BaseBdev4", 00:13:27.134 "uuid": "23318573-4c35-5b68-bacd-cefdb3a7cc1a", 00:13:27.134 "is_configured": true, 00:13:27.134 "data_offset": 0, 00:13:27.134 "data_size": 65536 00:13:27.134 } 00:13:27.134 ] 00:13:27.134 }' 00:13:27.134 17:34:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:27.134 17:34:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:27.394 96.50 IOPS, 289.50 MiB/s [2024-11-27T17:34:58.586Z] 17:34:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:13:27.394 17:34:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:27.394 17:34:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:27.394 [2024-11-27 17:34:58.578562] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:13:27.394 [2024-11-27 17:34:58.578631] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:13:27.655 00:13:27.655 Latency(us) 00:13:27.655 [2024-11-27T17:34:58.847Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:13:27.655 Job: raid_bdev1 (Core Mask 0x1, workload: randrw, percentage: 50, depth: 2, IO size: 3145728) 00:13:27.655 raid_bdev1 : 8.11 95.44 286.32 0.00 0.00 13819.62 295.13 112183.90 00:13:27.655 [2024-11-27T17:34:58.847Z] =================================================================================================================== 00:13:27.655 [2024-11-27T17:34:58.847Z] Total : 95.44 286.32 0.00 0.00 13819.62 295.13 112183.90 00:13:27.655 [2024-11-27 17:34:58.673339] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:27.655 [2024-11-27 17:34:58.673370] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:13:27.655 [2024-11-27 17:34:58.673462] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:13:27.655 [2024-11-27 17:34:58.673473] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:13:27.655 { 00:13:27.655 "results": [ 00:13:27.655 { 00:13:27.655 "job": "raid_bdev1", 00:13:27.655 "core_mask": "0x1", 00:13:27.655 "workload": "randrw", 00:13:27.655 "percentage": 50, 00:13:27.655 "status": "finished", 00:13:27.655 "queue_depth": 2, 00:13:27.655 "io_size": 3145728, 00:13:27.655 "runtime": 8.109784, 00:13:27.655 "iops": 95.44027313181215, 00:13:27.655 "mibps": 286.32081939543644, 00:13:27.655 "io_failed": 0, 00:13:27.655 "io_timeout": 0, 00:13:27.655 "avg_latency_us": 13819.615043498865, 00:13:27.655 "min_latency_us": 295.12663755458516, 00:13:27.655 "max_latency_us": 112183.89519650655 00:13:27.655 } 00:13:27.655 ], 00:13:27.655 "core_count": 1 00:13:27.655 } 00:13:27.655 17:34:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:27.655 17:34:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:27.655 17:34:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:27.655 17:34:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@720 -- # jq length 00:13:27.655 17:34:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:27.655 17:34:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:27.655 17:34:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:13:27.655 17:34:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:13:27.655 17:34:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@723 -- # '[' true = true ']' 00:13:27.655 17:34:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@725 -- # nbd_start_disks /var/tmp/spdk.sock spare /dev/nbd0 00:13:27.655 17:34:58 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:13:27.655 17:34:58 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # bdev_list=('spare') 00:13:27.655 17:34:58 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:13:27.655 17:34:58 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:13:27.655 17:34:58 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:13:27.655 17:34:58 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@12 -- # local i 00:13:27.655 17:34:58 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:13:27.655 17:34:58 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:13:27.655 17:34:58 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd0 00:13:27.915 /dev/nbd0 00:13:27.915 17:34:58 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:13:27.915 17:34:58 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:13:27.915 17:34:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:13:27.915 17:34:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@869 -- # local i 00:13:27.915 17:34:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:13:27.915 17:34:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:13:27.915 17:34:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:13:27.915 17:34:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@873 -- # break 00:13:27.915 17:34:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:13:27.915 17:34:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:13:27.915 17:34:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:13:27.915 1+0 records in 00:13:27.915 1+0 records out 00:13:27.915 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000541716 s, 7.6 MB/s 00:13:27.915 17:34:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:27.915 17:34:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@886 -- # size=4096 00:13:27.915 17:34:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:27.915 17:34:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:13:27.915 17:34:58 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@889 -- # return 0 00:13:27.915 17:34:58 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:13:27.915 17:34:58 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:13:27.915 17:34:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@726 -- # for bdev in "${base_bdevs[@]:1}" 00:13:27.915 17:34:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@727 -- # '[' -z '' ']' 00:13:27.915 17:34:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@728 -- # continue 00:13:27.915 17:34:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@726 -- # for bdev in "${base_bdevs[@]:1}" 00:13:27.915 17:34:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@727 -- # '[' -z BaseBdev3 ']' 00:13:27.915 17:34:58 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@730 -- # nbd_start_disks /var/tmp/spdk.sock BaseBdev3 /dev/nbd1 00:13:27.915 17:34:58 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:13:27.915 17:34:58 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev3') 00:13:27.915 17:34:58 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:13:27.915 17:34:58 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd1') 00:13:27.915 17:34:58 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:13:27.915 17:34:58 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@12 -- # local i 00:13:27.915 17:34:58 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:13:27.915 17:34:58 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:13:27.915 17:34:58 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev3 /dev/nbd1 00:13:28.176 /dev/nbd1 00:13:28.176 17:34:59 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:13:28.176 17:34:59 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:13:28.176 17:34:59 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@868 -- # local nbd_name=nbd1 00:13:28.176 17:34:59 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@869 -- # local i 00:13:28.176 17:34:59 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:13:28.176 17:34:59 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:13:28.176 17:34:59 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@872 -- # grep -q -w nbd1 /proc/partitions 00:13:28.176 17:34:59 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@873 -- # break 00:13:28.176 17:34:59 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:13:28.176 17:34:59 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:13:28.176 17:34:59 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@885 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:13:28.176 1+0 records in 00:13:28.176 1+0 records out 00:13:28.176 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000506545 s, 8.1 MB/s 00:13:28.176 17:34:59 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:28.176 17:34:59 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@886 -- # size=4096 00:13:28.176 17:34:59 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:28.176 17:34:59 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:13:28.176 17:34:59 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@889 -- # return 0 00:13:28.176 17:34:59 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:13:28.176 17:34:59 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:13:28.176 17:34:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@731 -- # cmp -i 0 /dev/nbd0 /dev/nbd1 00:13:28.176 17:34:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@732 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd1 00:13:28.176 17:34:59 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:13:28.176 17:34:59 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd1') 00:13:28.176 17:34:59 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:13:28.176 17:34:59 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@51 -- # local i 00:13:28.176 17:34:59 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:13:28.176 17:34:59 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:13:28.435 17:34:59 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:13:28.435 17:34:59 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:13:28.435 17:34:59 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:13:28.435 17:34:59 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:13:28.435 17:34:59 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:13:28.435 17:34:59 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:13:28.435 17:34:59 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@41 -- # break 00:13:28.435 17:34:59 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@45 -- # return 0 00:13:28.435 17:34:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@726 -- # for bdev in "${base_bdevs[@]:1}" 00:13:28.435 17:34:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@727 -- # '[' -z BaseBdev4 ']' 00:13:28.435 17:34:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@730 -- # nbd_start_disks /var/tmp/spdk.sock BaseBdev4 /dev/nbd1 00:13:28.435 17:34:59 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:13:28.435 17:34:59 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev4') 00:13:28.435 17:34:59 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:13:28.435 17:34:59 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd1') 00:13:28.435 17:34:59 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:13:28.435 17:34:59 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@12 -- # local i 00:13:28.435 17:34:59 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:13:28.435 17:34:59 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:13:28.435 17:34:59 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev4 /dev/nbd1 00:13:28.694 /dev/nbd1 00:13:28.694 17:34:59 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:13:28.694 17:34:59 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:13:28.694 17:34:59 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@868 -- # local nbd_name=nbd1 00:13:28.694 17:34:59 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@869 -- # local i 00:13:28.694 17:34:59 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:13:28.694 17:34:59 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:13:28.694 17:34:59 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@872 -- # grep -q -w nbd1 /proc/partitions 00:13:28.694 17:34:59 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@873 -- # break 00:13:28.694 17:34:59 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:13:28.695 17:34:59 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:13:28.695 17:34:59 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@885 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:13:28.695 1+0 records in 00:13:28.695 1+0 records out 00:13:28.695 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000352845 s, 11.6 MB/s 00:13:28.695 17:34:59 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:28.695 17:34:59 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@886 -- # size=4096 00:13:28.695 17:34:59 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:28.695 17:34:59 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:13:28.695 17:34:59 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@889 -- # return 0 00:13:28.695 17:34:59 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:13:28.695 17:34:59 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:13:28.695 17:34:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@731 -- # cmp -i 0 /dev/nbd0 /dev/nbd1 00:13:28.695 17:34:59 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@732 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd1 00:13:28.695 17:34:59 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:13:28.695 17:34:59 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd1') 00:13:28.695 17:34:59 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:13:28.695 17:34:59 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@51 -- # local i 00:13:28.695 17:34:59 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:13:28.695 17:34:59 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:13:28.954 17:35:00 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:13:28.954 17:35:00 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:13:28.954 17:35:00 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:13:28.954 17:35:00 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:13:28.954 17:35:00 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:13:28.954 17:35:00 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:13:28.954 17:35:00 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@41 -- # break 00:13:28.954 17:35:00 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@45 -- # return 0 00:13:28.954 17:35:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@734 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:13:28.954 17:35:00 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:13:28.954 17:35:00 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:13:28.954 17:35:00 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:13:28.954 17:35:00 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@51 -- # local i 00:13:28.954 17:35:00 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:13:28.954 17:35:00 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:13:29.213 17:35:00 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:13:29.213 17:35:00 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:13:29.213 17:35:00 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:13:29.213 17:35:00 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:13:29.213 17:35:00 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:13:29.213 17:35:00 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:13:29.213 17:35:00 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@41 -- # break 00:13:29.213 17:35:00 bdev_raid.raid_rebuild_test_io -- bdev/nbd_common.sh@45 -- # return 0 00:13:29.213 17:35:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@743 -- # '[' false = true ']' 00:13:29.213 17:35:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@784 -- # killprocess 89207 00:13:29.213 17:35:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@950 -- # '[' -z 89207 ']' 00:13:29.213 17:35:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@954 -- # kill -0 89207 00:13:29.213 17:35:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@955 -- # uname 00:13:29.213 17:35:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:13:29.213 17:35:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 89207 00:13:29.213 17:35:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:13:29.213 17:35:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:13:29.213 17:35:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@968 -- # echo 'killing process with pid 89207' 00:13:29.213 killing process with pid 89207 00:13:29.213 Received shutdown signal, test time was about 9.754983 seconds 00:13:29.213 00:13:29.213 Latency(us) 00:13:29.213 [2024-11-27T17:35:00.405Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:13:29.213 [2024-11-27T17:35:00.405Z] =================================================================================================================== 00:13:29.213 [2024-11-27T17:35:00.405Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:13:29.213 17:35:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@969 -- # kill 89207 00:13:29.213 [2024-11-27 17:35:00.312375] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:13:29.213 17:35:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@974 -- # wait 89207 00:13:29.213 [2024-11-27 17:35:00.357104] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:13:29.472 17:35:00 bdev_raid.raid_rebuild_test_io -- bdev/bdev_raid.sh@786 -- # return 0 00:13:29.472 00:13:29.472 real 0m11.750s 00:13:29.472 user 0m15.113s 00:13:29.472 sys 0m1.822s 00:13:29.472 17:35:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@1126 -- # xtrace_disable 00:13:29.472 ************************************ 00:13:29.472 END TEST raid_rebuild_test_io 00:13:29.472 ************************************ 00:13:29.472 17:35:00 bdev_raid.raid_rebuild_test_io -- common/autotest_common.sh@10 -- # set +x 00:13:29.472 17:35:00 bdev_raid -- bdev/bdev_raid.sh@981 -- # run_test raid_rebuild_test_sb_io raid_rebuild_test raid1 4 true true true 00:13:29.472 17:35:00 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 7 -le 1 ']' 00:13:29.472 17:35:00 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:13:29.472 17:35:00 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:13:29.732 ************************************ 00:13:29.732 START TEST raid_rebuild_test_sb_io 00:13:29.732 ************************************ 00:13:29.732 17:35:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@1125 -- # raid_rebuild_test raid1 4 true true true 00:13:29.732 17:35:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:13:29.732 17:35:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=4 00:13:29.732 17:35:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:13:29.732 17:35:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@572 -- # local background_io=true 00:13:29.732 17:35:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@573 -- # local verify=true 00:13:29.732 17:35:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:13:29.732 17:35:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:13:29.732 17:35:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:13:29.732 17:35:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:13:29.732 17:35:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:13:29.732 17:35:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:13:29.732 17:35:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:13:29.732 17:35:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:13:29.732 17:35:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev3 00:13:29.732 17:35:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:13:29.732 17:35:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:13:29.732 17:35:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@576 -- # echo BaseBdev4 00:13:29.732 17:35:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:13:29.732 17:35:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:13:29.732 17:35:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:13:29.732 17:35:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:13:29.732 17:35:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:13:29.732 17:35:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@576 -- # local strip_size 00:13:29.732 17:35:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@577 -- # local create_arg 00:13:29.732 17:35:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:13:29.732 17:35:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@579 -- # local data_offset 00:13:29.732 17:35:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:13:29.732 17:35:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:13:29.732 17:35:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:13:29.732 17:35:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:13:29.732 17:35:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@597 -- # raid_pid=89599 00:13:29.732 17:35:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:13:29.732 17:35:00 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@598 -- # waitforlisten 89599 00:13:29.732 17:35:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@831 -- # '[' -z 89599 ']' 00:13:29.732 17:35:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:13:29.732 17:35:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@836 -- # local max_retries=100 00:13:29.732 17:35:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:13:29.732 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:13:29.732 17:35:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@840 -- # xtrace_disable 00:13:29.732 17:35:00 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:29.732 [2024-11-27 17:35:00.763307] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:13:29.732 [2024-11-27 17:35:00.763478] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.ealI/O size of 3145728 is greater than zero copy threshold (65536). 00:13:29.732 Zero copy mechanism will not be used. 00:13:29.732 :6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid89599 ] 00:13:29.733 [2024-11-27 17:35:00.908537] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:13:29.991 [2024-11-27 17:35:00.953386] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:13:29.991 [2024-11-27 17:35:00.996344] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:13:29.991 [2024-11-27 17:35:00.996450] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:13:30.561 17:35:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:13:30.561 17:35:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@864 -- # return 0 00:13:30.561 17:35:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:13:30.561 17:35:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:13:30.561 17:35:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:30.561 17:35:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:30.561 BaseBdev1_malloc 00:13:30.561 17:35:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:30.561 17:35:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:13:30.561 17:35:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:30.561 17:35:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:30.561 [2024-11-27 17:35:01.595225] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:13:30.561 [2024-11-27 17:35:01.595348] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:30.561 [2024-11-27 17:35:01.595390] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:13:30.561 [2024-11-27 17:35:01.595422] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:30.561 [2024-11-27 17:35:01.597414] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:30.561 [2024-11-27 17:35:01.597479] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:13:30.561 BaseBdev1 00:13:30.561 17:35:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:30.561 17:35:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:13:30.561 17:35:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:13:30.561 17:35:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:30.561 17:35:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:30.561 BaseBdev2_malloc 00:13:30.561 17:35:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:30.561 17:35:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:13:30.561 17:35:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:30.561 17:35:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:30.561 [2024-11-27 17:35:01.641122] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:13:30.561 [2024-11-27 17:35:01.641233] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:30.561 [2024-11-27 17:35:01.641278] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:13:30.561 [2024-11-27 17:35:01.641299] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:30.561 [2024-11-27 17:35:01.646120] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:30.561 [2024-11-27 17:35:01.646206] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:13:30.561 BaseBdev2 00:13:30.561 17:35:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:30.561 17:35:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:13:30.561 17:35:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:13:30.561 17:35:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:30.561 17:35:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:30.561 BaseBdev3_malloc 00:13:30.561 17:35:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:30.561 17:35:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev3_malloc -p BaseBdev3 00:13:30.561 17:35:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:30.561 17:35:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:30.561 [2024-11-27 17:35:01.672593] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev3_malloc 00:13:30.561 [2024-11-27 17:35:01.672650] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:30.561 [2024-11-27 17:35:01.672676] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:13:30.561 [2024-11-27 17:35:01.672684] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:30.561 [2024-11-27 17:35:01.674728] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:30.561 [2024-11-27 17:35:01.674765] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:13:30.561 BaseBdev3 00:13:30.561 17:35:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:30.561 17:35:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:13:30.561 17:35:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:13:30.561 17:35:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:30.561 17:35:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:30.561 BaseBdev4_malloc 00:13:30.561 17:35:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:30.561 17:35:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev4_malloc -p BaseBdev4 00:13:30.561 17:35:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:30.562 17:35:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:30.562 [2024-11-27 17:35:01.701251] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev4_malloc 00:13:30.562 [2024-11-27 17:35:01.701296] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:30.562 [2024-11-27 17:35:01.701316] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:13:30.562 [2024-11-27 17:35:01.701324] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:30.562 [2024-11-27 17:35:01.703286] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:30.562 [2024-11-27 17:35:01.703319] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:13:30.562 BaseBdev4 00:13:30.562 17:35:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:30.562 17:35:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:13:30.562 17:35:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:30.562 17:35:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:30.562 spare_malloc 00:13:30.562 17:35:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:30.562 17:35:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:13:30.562 17:35:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:30.562 17:35:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:30.562 spare_delay 00:13:30.562 17:35:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:30.562 17:35:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:13:30.562 17:35:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:30.562 17:35:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:30.562 [2024-11-27 17:35:01.741678] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:13:30.562 [2024-11-27 17:35:01.741723] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:30.562 [2024-11-27 17:35:01.741740] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009c80 00:13:30.562 [2024-11-27 17:35:01.741748] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:30.562 [2024-11-27 17:35:01.743748] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:30.562 [2024-11-27 17:35:01.743833] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:13:30.562 spare 00:13:30.562 17:35:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:30.562 17:35:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 00:13:30.562 17:35:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:30.562 17:35:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:30.822 [2024-11-27 17:35:01.753732] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:13:30.822 [2024-11-27 17:35:01.755546] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:13:30.822 [2024-11-27 17:35:01.755607] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:13:30.822 [2024-11-27 17:35:01.755653] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:13:30.822 [2024-11-27 17:35:01.755818] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:13:30.822 [2024-11-27 17:35:01.755829] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:13:30.822 [2024-11-27 17:35:01.756064] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:13:30.822 [2024-11-27 17:35:01.756198] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:13:30.822 [2024-11-27 17:35:01.756210] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:13:30.822 [2024-11-27 17:35:01.756325] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:30.822 17:35:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:30.822 17:35:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 4 00:13:30.822 17:35:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:30.822 17:35:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:30.822 17:35:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:30.822 17:35:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:30.822 17:35:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:13:30.822 17:35:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:30.822 17:35:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:30.822 17:35:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:30.822 17:35:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:30.822 17:35:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:30.822 17:35:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:30.822 17:35:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:30.822 17:35:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:30.822 17:35:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:30.822 17:35:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:30.822 "name": "raid_bdev1", 00:13:30.822 "uuid": "da8a894f-f09b-4dea-ad41-aabdb7da2f87", 00:13:30.822 "strip_size_kb": 0, 00:13:30.822 "state": "online", 00:13:30.822 "raid_level": "raid1", 00:13:30.822 "superblock": true, 00:13:30.822 "num_base_bdevs": 4, 00:13:30.822 "num_base_bdevs_discovered": 4, 00:13:30.822 "num_base_bdevs_operational": 4, 00:13:30.822 "base_bdevs_list": [ 00:13:30.822 { 00:13:30.822 "name": "BaseBdev1", 00:13:30.822 "uuid": "ddbe946f-7005-59a5-9eaa-8136d9b85801", 00:13:30.822 "is_configured": true, 00:13:30.822 "data_offset": 2048, 00:13:30.822 "data_size": 63488 00:13:30.822 }, 00:13:30.822 { 00:13:30.822 "name": "BaseBdev2", 00:13:30.822 "uuid": "0c03df5b-02ae-55cb-8b40-2d7414997ed4", 00:13:30.822 "is_configured": true, 00:13:30.822 "data_offset": 2048, 00:13:30.822 "data_size": 63488 00:13:30.822 }, 00:13:30.822 { 00:13:30.822 "name": "BaseBdev3", 00:13:30.822 "uuid": "e9304de8-d30a-5749-92d8-8df9c95094de", 00:13:30.823 "is_configured": true, 00:13:30.823 "data_offset": 2048, 00:13:30.823 "data_size": 63488 00:13:30.823 }, 00:13:30.823 { 00:13:30.823 "name": "BaseBdev4", 00:13:30.823 "uuid": "69df1b46-7233-5fb6-b5b6-39ecbfd83e27", 00:13:30.823 "is_configured": true, 00:13:30.823 "data_offset": 2048, 00:13:30.823 "data_size": 63488 00:13:30.823 } 00:13:30.823 ] 00:13:30.823 }' 00:13:30.823 17:35:01 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:30.823 17:35:01 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:31.081 17:35:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:13:31.081 17:35:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:13:31.081 17:35:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:31.081 17:35:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:31.081 [2024-11-27 17:35:02.265105] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:13:31.341 17:35:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:31.341 17:35:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=63488 00:13:31.341 17:35:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:31.341 17:35:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:31.341 17:35:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:31.341 17:35:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:13:31.341 17:35:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:31.341 17:35:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@619 -- # data_offset=2048 00:13:31.341 17:35:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@621 -- # '[' true = true ']' 00:13:31.341 17:35:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:13:31.341 17:35:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@623 -- # /home/vagrant/spdk_repo/spdk/examples/bdev/bdevperf/bdevperf.py perform_tests 00:13:31.341 17:35:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:31.341 17:35:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:31.341 [2024-11-27 17:35:02.356614] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:13:31.341 17:35:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:31.341 17:35:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:13:31.341 17:35:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:31.341 17:35:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:31.341 17:35:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:31.341 17:35:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:31.341 17:35:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:31.341 17:35:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:31.341 17:35:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:31.341 17:35:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:31.341 17:35:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:31.341 17:35:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:31.341 17:35:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:31.341 17:35:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:31.341 17:35:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:31.341 17:35:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:31.341 17:35:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:31.341 "name": "raid_bdev1", 00:13:31.341 "uuid": "da8a894f-f09b-4dea-ad41-aabdb7da2f87", 00:13:31.341 "strip_size_kb": 0, 00:13:31.341 "state": "online", 00:13:31.341 "raid_level": "raid1", 00:13:31.341 "superblock": true, 00:13:31.341 "num_base_bdevs": 4, 00:13:31.341 "num_base_bdevs_discovered": 3, 00:13:31.341 "num_base_bdevs_operational": 3, 00:13:31.341 "base_bdevs_list": [ 00:13:31.341 { 00:13:31.341 "name": null, 00:13:31.341 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:31.341 "is_configured": false, 00:13:31.341 "data_offset": 0, 00:13:31.341 "data_size": 63488 00:13:31.341 }, 00:13:31.341 { 00:13:31.341 "name": "BaseBdev2", 00:13:31.341 "uuid": "0c03df5b-02ae-55cb-8b40-2d7414997ed4", 00:13:31.341 "is_configured": true, 00:13:31.341 "data_offset": 2048, 00:13:31.341 "data_size": 63488 00:13:31.341 }, 00:13:31.341 { 00:13:31.341 "name": "BaseBdev3", 00:13:31.341 "uuid": "e9304de8-d30a-5749-92d8-8df9c95094de", 00:13:31.341 "is_configured": true, 00:13:31.341 "data_offset": 2048, 00:13:31.341 "data_size": 63488 00:13:31.341 }, 00:13:31.341 { 00:13:31.341 "name": "BaseBdev4", 00:13:31.341 "uuid": "69df1b46-7233-5fb6-b5b6-39ecbfd83e27", 00:13:31.341 "is_configured": true, 00:13:31.341 "data_offset": 2048, 00:13:31.341 "data_size": 63488 00:13:31.341 } 00:13:31.341 ] 00:13:31.341 }' 00:13:31.341 17:35:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:31.341 17:35:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:31.342 [2024-11-27 17:35:02.442521] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002870 00:13:31.342 I/O size of 3145728 is greater than zero copy threshold (65536). 00:13:31.342 Zero copy mechanism will not be used. 00:13:31.342 Running I/O for 60 seconds... 00:13:31.911 17:35:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:13:31.912 17:35:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:31.912 17:35:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:31.912 [2024-11-27 17:35:02.815727] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:13:31.912 17:35:02 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:31.912 17:35:02 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@647 -- # sleep 1 00:13:31.912 [2024-11-27 17:35:02.869118] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002940 00:13:31.912 [2024-11-27 17:35:02.871166] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:13:31.912 [2024-11-27 17:35:02.980726] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:13:31.912 [2024-11-27 17:35:02.981941] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:13:32.170 [2024-11-27 17:35:03.188540] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:13:32.170 [2024-11-27 17:35:03.189095] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:13:32.430 180.00 IOPS, 540.00 MiB/s [2024-11-27T17:35:03.622Z] [2024-11-27 17:35:03.531886] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:13:32.690 [2024-11-27 17:35:03.669443] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:13:32.690 [2024-11-27 17:35:03.670220] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 10240 offset_begin: 6144 offset_end: 12288 00:13:32.690 17:35:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:32.690 17:35:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:32.690 17:35:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:32.690 17:35:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:32.690 17:35:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:32.690 17:35:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:32.690 17:35:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:32.690 17:35:03 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:32.690 17:35:03 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:32.690 17:35:03 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:32.951 17:35:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:32.951 "name": "raid_bdev1", 00:13:32.951 "uuid": "da8a894f-f09b-4dea-ad41-aabdb7da2f87", 00:13:32.951 "strip_size_kb": 0, 00:13:32.951 "state": "online", 00:13:32.951 "raid_level": "raid1", 00:13:32.951 "superblock": true, 00:13:32.951 "num_base_bdevs": 4, 00:13:32.951 "num_base_bdevs_discovered": 4, 00:13:32.951 "num_base_bdevs_operational": 4, 00:13:32.951 "process": { 00:13:32.951 "type": "rebuild", 00:13:32.951 "target": "spare", 00:13:32.951 "progress": { 00:13:32.951 "blocks": 10240, 00:13:32.951 "percent": 16 00:13:32.951 } 00:13:32.951 }, 00:13:32.951 "base_bdevs_list": [ 00:13:32.951 { 00:13:32.951 "name": "spare", 00:13:32.951 "uuid": "713e7c40-9159-54c2-9e60-a67bcff7b97c", 00:13:32.951 "is_configured": true, 00:13:32.951 "data_offset": 2048, 00:13:32.951 "data_size": 63488 00:13:32.951 }, 00:13:32.951 { 00:13:32.951 "name": "BaseBdev2", 00:13:32.951 "uuid": "0c03df5b-02ae-55cb-8b40-2d7414997ed4", 00:13:32.951 "is_configured": true, 00:13:32.951 "data_offset": 2048, 00:13:32.951 "data_size": 63488 00:13:32.951 }, 00:13:32.951 { 00:13:32.951 "name": "BaseBdev3", 00:13:32.951 "uuid": "e9304de8-d30a-5749-92d8-8df9c95094de", 00:13:32.951 "is_configured": true, 00:13:32.951 "data_offset": 2048, 00:13:32.951 "data_size": 63488 00:13:32.951 }, 00:13:32.951 { 00:13:32.951 "name": "BaseBdev4", 00:13:32.951 "uuid": "69df1b46-7233-5fb6-b5b6-39ecbfd83e27", 00:13:32.951 "is_configured": true, 00:13:32.951 "data_offset": 2048, 00:13:32.951 "data_size": 63488 00:13:32.951 } 00:13:32.951 ] 00:13:32.951 }' 00:13:32.951 17:35:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:32.951 17:35:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:32.951 17:35:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:32.951 17:35:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:32.951 17:35:03 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:13:32.951 17:35:03 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:32.951 17:35:03 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:32.951 [2024-11-27 17:35:03.980556] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:32.951 [2024-11-27 17:35:04.033358] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:13:32.951 [2024-11-27 17:35:04.043521] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:32.951 [2024-11-27 17:35:04.043567] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:32.951 [2024-11-27 17:35:04.043582] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:13:32.951 [2024-11-27 17:35:04.055017] bdev_raid.c:1970:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 0 raid_ch: 0x60d000002870 00:13:32.951 17:35:04 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:32.951 17:35:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:13:32.951 17:35:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:32.951 17:35:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:32.951 17:35:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:32.951 17:35:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:32.951 17:35:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:32.951 17:35:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:32.951 17:35:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:32.951 17:35:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:32.951 17:35:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:32.951 17:35:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:32.951 17:35:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:32.951 17:35:04 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:32.951 17:35:04 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:32.951 17:35:04 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:32.951 17:35:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:32.951 "name": "raid_bdev1", 00:13:32.951 "uuid": "da8a894f-f09b-4dea-ad41-aabdb7da2f87", 00:13:32.951 "strip_size_kb": 0, 00:13:32.951 "state": "online", 00:13:32.951 "raid_level": "raid1", 00:13:32.951 "superblock": true, 00:13:32.951 "num_base_bdevs": 4, 00:13:32.951 "num_base_bdevs_discovered": 3, 00:13:32.951 "num_base_bdevs_operational": 3, 00:13:32.951 "base_bdevs_list": [ 00:13:32.951 { 00:13:32.951 "name": null, 00:13:32.951 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:32.951 "is_configured": false, 00:13:32.951 "data_offset": 0, 00:13:32.951 "data_size": 63488 00:13:32.951 }, 00:13:32.951 { 00:13:32.951 "name": "BaseBdev2", 00:13:32.951 "uuid": "0c03df5b-02ae-55cb-8b40-2d7414997ed4", 00:13:32.951 "is_configured": true, 00:13:32.951 "data_offset": 2048, 00:13:32.951 "data_size": 63488 00:13:32.951 }, 00:13:32.951 { 00:13:32.951 "name": "BaseBdev3", 00:13:32.951 "uuid": "e9304de8-d30a-5749-92d8-8df9c95094de", 00:13:32.951 "is_configured": true, 00:13:32.951 "data_offset": 2048, 00:13:32.951 "data_size": 63488 00:13:32.951 }, 00:13:32.951 { 00:13:32.951 "name": "BaseBdev4", 00:13:32.951 "uuid": "69df1b46-7233-5fb6-b5b6-39ecbfd83e27", 00:13:32.951 "is_configured": true, 00:13:32.951 "data_offset": 2048, 00:13:32.951 "data_size": 63488 00:13:32.951 } 00:13:32.951 ] 00:13:32.951 }' 00:13:32.951 17:35:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:32.951 17:35:04 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:33.520 175.00 IOPS, 525.00 MiB/s [2024-11-27T17:35:04.712Z] 17:35:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:13:33.520 17:35:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:33.520 17:35:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:13:33.520 17:35:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:13:33.520 17:35:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:33.520 17:35:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:33.520 17:35:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:33.520 17:35:04 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:33.520 17:35:04 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:33.520 17:35:04 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:33.520 17:35:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:33.520 "name": "raid_bdev1", 00:13:33.520 "uuid": "da8a894f-f09b-4dea-ad41-aabdb7da2f87", 00:13:33.520 "strip_size_kb": 0, 00:13:33.520 "state": "online", 00:13:33.520 "raid_level": "raid1", 00:13:33.520 "superblock": true, 00:13:33.520 "num_base_bdevs": 4, 00:13:33.520 "num_base_bdevs_discovered": 3, 00:13:33.520 "num_base_bdevs_operational": 3, 00:13:33.520 "base_bdevs_list": [ 00:13:33.520 { 00:13:33.520 "name": null, 00:13:33.520 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:33.520 "is_configured": false, 00:13:33.520 "data_offset": 0, 00:13:33.520 "data_size": 63488 00:13:33.520 }, 00:13:33.520 { 00:13:33.520 "name": "BaseBdev2", 00:13:33.520 "uuid": "0c03df5b-02ae-55cb-8b40-2d7414997ed4", 00:13:33.520 "is_configured": true, 00:13:33.520 "data_offset": 2048, 00:13:33.520 "data_size": 63488 00:13:33.520 }, 00:13:33.520 { 00:13:33.520 "name": "BaseBdev3", 00:13:33.520 "uuid": "e9304de8-d30a-5749-92d8-8df9c95094de", 00:13:33.520 "is_configured": true, 00:13:33.520 "data_offset": 2048, 00:13:33.520 "data_size": 63488 00:13:33.520 }, 00:13:33.520 { 00:13:33.520 "name": "BaseBdev4", 00:13:33.520 "uuid": "69df1b46-7233-5fb6-b5b6-39ecbfd83e27", 00:13:33.520 "is_configured": true, 00:13:33.520 "data_offset": 2048, 00:13:33.520 "data_size": 63488 00:13:33.520 } 00:13:33.520 ] 00:13:33.520 }' 00:13:33.520 17:35:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:33.520 17:35:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:13:33.520 17:35:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:33.520 17:35:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:13:33.520 17:35:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:13:33.520 17:35:04 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:33.520 17:35:04 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:33.520 [2024-11-27 17:35:04.668872] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:13:33.520 17:35:04 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:33.520 17:35:04 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@663 -- # sleep 1 00:13:33.784 [2024-11-27 17:35:04.712284] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002a10 00:13:33.784 [2024-11-27 17:35:04.714270] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:13:33.784 [2024-11-27 17:35:04.835394] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:13:33.784 [2024-11-27 17:35:04.835847] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 2048 offset_begin: 0 offset_end: 6144 00:13:33.784 [2024-11-27 17:35:04.957462] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:13:33.784 [2024-11-27 17:35:04.958099] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 4096 offset_begin: 0 offset_end: 6144 00:13:34.361 [2024-11-27 17:35:05.290724] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 8192 offset_begin: 6144 offset_end: 12288 00:13:34.620 152.00 IOPS, 456.00 MiB/s [2024-11-27T17:35:05.812Z] [2024-11-27 17:35:05.647988] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 14336 offset_begin: 12288 offset_end: 18432 00:13:34.620 17:35:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:34.620 17:35:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:34.620 17:35:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:34.620 17:35:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:34.620 17:35:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:34.620 17:35:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:34.620 17:35:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:34.620 17:35:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:34.620 17:35:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:34.620 17:35:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:34.620 17:35:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:34.620 "name": "raid_bdev1", 00:13:34.620 "uuid": "da8a894f-f09b-4dea-ad41-aabdb7da2f87", 00:13:34.620 "strip_size_kb": 0, 00:13:34.620 "state": "online", 00:13:34.620 "raid_level": "raid1", 00:13:34.620 "superblock": true, 00:13:34.620 "num_base_bdevs": 4, 00:13:34.620 "num_base_bdevs_discovered": 4, 00:13:34.620 "num_base_bdevs_operational": 4, 00:13:34.620 "process": { 00:13:34.620 "type": "rebuild", 00:13:34.620 "target": "spare", 00:13:34.620 "progress": { 00:13:34.620 "blocks": 14336, 00:13:34.620 "percent": 22 00:13:34.620 } 00:13:34.620 }, 00:13:34.620 "base_bdevs_list": [ 00:13:34.620 { 00:13:34.620 "name": "spare", 00:13:34.620 "uuid": "713e7c40-9159-54c2-9e60-a67bcff7b97c", 00:13:34.620 "is_configured": true, 00:13:34.620 "data_offset": 2048, 00:13:34.620 "data_size": 63488 00:13:34.620 }, 00:13:34.620 { 00:13:34.620 "name": "BaseBdev2", 00:13:34.620 "uuid": "0c03df5b-02ae-55cb-8b40-2d7414997ed4", 00:13:34.620 "is_configured": true, 00:13:34.620 "data_offset": 2048, 00:13:34.620 "data_size": 63488 00:13:34.620 }, 00:13:34.620 { 00:13:34.620 "name": "BaseBdev3", 00:13:34.620 "uuid": "e9304de8-d30a-5749-92d8-8df9c95094de", 00:13:34.620 "is_configured": true, 00:13:34.620 "data_offset": 2048, 00:13:34.620 "data_size": 63488 00:13:34.620 }, 00:13:34.620 { 00:13:34.620 "name": "BaseBdev4", 00:13:34.620 "uuid": "69df1b46-7233-5fb6-b5b6-39ecbfd83e27", 00:13:34.620 "is_configured": true, 00:13:34.620 "data_offset": 2048, 00:13:34.620 "data_size": 63488 00:13:34.620 } 00:13:34.620 ] 00:13:34.620 }' 00:13:34.620 17:35:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:34.620 17:35:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:34.620 17:35:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:34.880 17:35:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:34.880 17:35:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:13:34.880 17:35:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:13:34.880 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:13:34.880 17:35:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=4 00:13:34.880 17:35:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:13:34.880 17:35:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@693 -- # '[' 4 -gt 2 ']' 00:13:34.880 17:35:05 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@695 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:13:34.880 17:35:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:34.880 17:35:05 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:34.880 [2024-11-27 17:35:05.843766] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:13:34.880 [2024-11-27 17:35:05.864163] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:13:34.880 [2024-11-27 17:35:05.864785] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:13:34.880 [2024-11-27 17:35:06.066395] bdev_raid.c:1970:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 1 raid_ch: 0x60d000002870 00:13:34.880 [2024-11-27 17:35:06.066469] bdev_raid.c:1970:raid_bdev_channel_remove_base_bdev: *DEBUG*: slot: 1 raid_ch: 0x60d000002a10 00:13:35.141 17:35:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:35.141 17:35:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@698 -- # base_bdevs[1]= 00:13:35.141 17:35:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@699 -- # (( num_base_bdevs_operational-- )) 00:13:35.141 17:35:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@702 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:35.141 17:35:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:35.141 17:35:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:35.141 17:35:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:35.141 17:35:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:35.141 [2024-11-27 17:35:06.075471] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 16384 offset_begin: 12288 offset_end: 18432 00:13:35.141 17:35:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:35.141 17:35:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:35.141 17:35:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:35.141 17:35:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:35.141 17:35:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:35.141 17:35:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:35.141 "name": "raid_bdev1", 00:13:35.141 "uuid": "da8a894f-f09b-4dea-ad41-aabdb7da2f87", 00:13:35.141 "strip_size_kb": 0, 00:13:35.141 "state": "online", 00:13:35.141 "raid_level": "raid1", 00:13:35.141 "superblock": true, 00:13:35.141 "num_base_bdevs": 4, 00:13:35.141 "num_base_bdevs_discovered": 3, 00:13:35.141 "num_base_bdevs_operational": 3, 00:13:35.141 "process": { 00:13:35.141 "type": "rebuild", 00:13:35.141 "target": "spare", 00:13:35.141 "progress": { 00:13:35.141 "blocks": 16384, 00:13:35.141 "percent": 25 00:13:35.141 } 00:13:35.141 }, 00:13:35.141 "base_bdevs_list": [ 00:13:35.141 { 00:13:35.141 "name": "spare", 00:13:35.141 "uuid": "713e7c40-9159-54c2-9e60-a67bcff7b97c", 00:13:35.141 "is_configured": true, 00:13:35.141 "data_offset": 2048, 00:13:35.141 "data_size": 63488 00:13:35.141 }, 00:13:35.141 { 00:13:35.141 "name": null, 00:13:35.141 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:35.141 "is_configured": false, 00:13:35.141 "data_offset": 0, 00:13:35.141 "data_size": 63488 00:13:35.141 }, 00:13:35.141 { 00:13:35.141 "name": "BaseBdev3", 00:13:35.141 "uuid": "e9304de8-d30a-5749-92d8-8df9c95094de", 00:13:35.141 "is_configured": true, 00:13:35.141 "data_offset": 2048, 00:13:35.141 "data_size": 63488 00:13:35.141 }, 00:13:35.141 { 00:13:35.141 "name": "BaseBdev4", 00:13:35.141 "uuid": "69df1b46-7233-5fb6-b5b6-39ecbfd83e27", 00:13:35.141 "is_configured": true, 00:13:35.141 "data_offset": 2048, 00:13:35.141 "data_size": 63488 00:13:35.141 } 00:13:35.141 ] 00:13:35.141 }' 00:13:35.141 17:35:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:35.141 17:35:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:35.142 17:35:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:35.142 17:35:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:35.142 17:35:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@706 -- # local timeout=414 00:13:35.142 17:35:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:35.142 17:35:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:35.142 17:35:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:35.142 17:35:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:35.142 17:35:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:35.142 17:35:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:35.142 17:35:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:35.142 17:35:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:35.142 17:35:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:35.142 17:35:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:35.142 17:35:06 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:35.142 17:35:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:35.142 "name": "raid_bdev1", 00:13:35.142 "uuid": "da8a894f-f09b-4dea-ad41-aabdb7da2f87", 00:13:35.142 "strip_size_kb": 0, 00:13:35.142 "state": "online", 00:13:35.142 "raid_level": "raid1", 00:13:35.142 "superblock": true, 00:13:35.142 "num_base_bdevs": 4, 00:13:35.142 "num_base_bdevs_discovered": 3, 00:13:35.142 "num_base_bdevs_operational": 3, 00:13:35.142 "process": { 00:13:35.142 "type": "rebuild", 00:13:35.142 "target": "spare", 00:13:35.142 "progress": { 00:13:35.142 "blocks": 16384, 00:13:35.142 "percent": 25 00:13:35.142 } 00:13:35.142 }, 00:13:35.142 "base_bdevs_list": [ 00:13:35.142 { 00:13:35.142 "name": "spare", 00:13:35.142 "uuid": "713e7c40-9159-54c2-9e60-a67bcff7b97c", 00:13:35.142 "is_configured": true, 00:13:35.142 "data_offset": 2048, 00:13:35.142 "data_size": 63488 00:13:35.142 }, 00:13:35.142 { 00:13:35.142 "name": null, 00:13:35.142 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:35.142 "is_configured": false, 00:13:35.142 "data_offset": 0, 00:13:35.142 "data_size": 63488 00:13:35.142 }, 00:13:35.142 { 00:13:35.142 "name": "BaseBdev3", 00:13:35.142 "uuid": "e9304de8-d30a-5749-92d8-8df9c95094de", 00:13:35.142 "is_configured": true, 00:13:35.142 "data_offset": 2048, 00:13:35.142 "data_size": 63488 00:13:35.142 }, 00:13:35.142 { 00:13:35.142 "name": "BaseBdev4", 00:13:35.142 "uuid": "69df1b46-7233-5fb6-b5b6-39ecbfd83e27", 00:13:35.142 "is_configured": true, 00:13:35.142 "data_offset": 2048, 00:13:35.142 "data_size": 63488 00:13:35.142 } 00:13:35.142 ] 00:13:35.142 }' 00:13:35.142 17:35:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:35.142 17:35:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:35.142 17:35:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:35.403 17:35:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:35.403 17:35:06 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:13:35.403 124.75 IOPS, 374.25 MiB/s [2024-11-27T17:35:06.595Z] [2024-11-27 17:35:06.506712] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 22528 offset_begin: 18432 offset_end: 24576 00:13:35.403 [2024-11-27 17:35:06.507194] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 22528 offset_begin: 18432 offset_end: 24576 00:13:35.974 [2024-11-27 17:35:06.933253] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 28672 offset_begin: 24576 offset_end: 30720 00:13:36.234 17:35:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:36.234 17:35:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:36.234 17:35:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:36.234 17:35:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:36.234 17:35:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:36.234 17:35:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:36.234 17:35:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:36.234 17:35:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:36.234 17:35:07 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:36.234 17:35:07 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:36.234 17:35:07 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:36.234 17:35:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:36.234 "name": "raid_bdev1", 00:13:36.234 "uuid": "da8a894f-f09b-4dea-ad41-aabdb7da2f87", 00:13:36.234 "strip_size_kb": 0, 00:13:36.234 "state": "online", 00:13:36.234 "raid_level": "raid1", 00:13:36.234 "superblock": true, 00:13:36.234 "num_base_bdevs": 4, 00:13:36.234 "num_base_bdevs_discovered": 3, 00:13:36.235 "num_base_bdevs_operational": 3, 00:13:36.235 "process": { 00:13:36.235 "type": "rebuild", 00:13:36.235 "target": "spare", 00:13:36.235 "progress": { 00:13:36.235 "blocks": 34816, 00:13:36.235 "percent": 54 00:13:36.235 } 00:13:36.235 }, 00:13:36.235 "base_bdevs_list": [ 00:13:36.235 { 00:13:36.235 "name": "spare", 00:13:36.235 "uuid": "713e7c40-9159-54c2-9e60-a67bcff7b97c", 00:13:36.235 "is_configured": true, 00:13:36.235 "data_offset": 2048, 00:13:36.235 "data_size": 63488 00:13:36.235 }, 00:13:36.235 { 00:13:36.235 "name": null, 00:13:36.235 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:36.235 "is_configured": false, 00:13:36.235 "data_offset": 0, 00:13:36.235 "data_size": 63488 00:13:36.235 }, 00:13:36.235 { 00:13:36.235 "name": "BaseBdev3", 00:13:36.235 "uuid": "e9304de8-d30a-5749-92d8-8df9c95094de", 00:13:36.235 "is_configured": true, 00:13:36.235 "data_offset": 2048, 00:13:36.235 "data_size": 63488 00:13:36.235 }, 00:13:36.235 { 00:13:36.235 "name": "BaseBdev4", 00:13:36.235 "uuid": "69df1b46-7233-5fb6-b5b6-39ecbfd83e27", 00:13:36.235 "is_configured": true, 00:13:36.235 "data_offset": 2048, 00:13:36.235 "data_size": 63488 00:13:36.235 } 00:13:36.235 ] 00:13:36.235 }' 00:13:36.235 17:35:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:36.495 109.40 IOPS, 328.20 MiB/s [2024-11-27T17:35:07.687Z] 17:35:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:36.495 17:35:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:36.495 [2024-11-27 17:35:07.471077] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 38912 offset_begin: 36864 offset_end: 43008 00:13:36.495 17:35:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:36.495 17:35:07 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:13:36.756 [2024-11-27 17:35:07.794169] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 45056 offset_begin: 43008 offset_end: 49152 00:13:36.756 [2024-11-27 17:35:07.900455] bdev_raid.c: 859:raid_bdev_submit_rw_request: *DEBUG*: split: process_offset: 47104 offset_begin: 43008 offset_end: 49152 00:13:37.326 99.50 IOPS, 298.50 MiB/s [2024-11-27T17:35:08.518Z] 17:35:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:37.326 17:35:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:37.326 17:35:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:37.326 17:35:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:37.326 17:35:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:37.326 17:35:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:37.326 17:35:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:37.326 17:35:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:37.326 17:35:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:37.326 17:35:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:37.586 17:35:08 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:37.586 17:35:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:37.586 "name": "raid_bdev1", 00:13:37.586 "uuid": "da8a894f-f09b-4dea-ad41-aabdb7da2f87", 00:13:37.586 "strip_size_kb": 0, 00:13:37.586 "state": "online", 00:13:37.586 "raid_level": "raid1", 00:13:37.586 "superblock": true, 00:13:37.586 "num_base_bdevs": 4, 00:13:37.586 "num_base_bdevs_discovered": 3, 00:13:37.586 "num_base_bdevs_operational": 3, 00:13:37.586 "process": { 00:13:37.586 "type": "rebuild", 00:13:37.586 "target": "spare", 00:13:37.586 "progress": { 00:13:37.586 "blocks": 55296, 00:13:37.586 "percent": 87 00:13:37.586 } 00:13:37.586 }, 00:13:37.586 "base_bdevs_list": [ 00:13:37.586 { 00:13:37.586 "name": "spare", 00:13:37.586 "uuid": "713e7c40-9159-54c2-9e60-a67bcff7b97c", 00:13:37.586 "is_configured": true, 00:13:37.586 "data_offset": 2048, 00:13:37.586 "data_size": 63488 00:13:37.586 }, 00:13:37.586 { 00:13:37.586 "name": null, 00:13:37.586 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:37.586 "is_configured": false, 00:13:37.586 "data_offset": 0, 00:13:37.586 "data_size": 63488 00:13:37.586 }, 00:13:37.586 { 00:13:37.586 "name": "BaseBdev3", 00:13:37.586 "uuid": "e9304de8-d30a-5749-92d8-8df9c95094de", 00:13:37.586 "is_configured": true, 00:13:37.586 "data_offset": 2048, 00:13:37.586 "data_size": 63488 00:13:37.586 }, 00:13:37.586 { 00:13:37.586 "name": "BaseBdev4", 00:13:37.586 "uuid": "69df1b46-7233-5fb6-b5b6-39ecbfd83e27", 00:13:37.586 "is_configured": true, 00:13:37.586 "data_offset": 2048, 00:13:37.586 "data_size": 63488 00:13:37.586 } 00:13:37.586 ] 00:13:37.586 }' 00:13:37.586 17:35:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:37.586 17:35:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:37.586 17:35:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:37.586 17:35:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:37.586 17:35:08 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@711 -- # sleep 1 00:13:37.846 [2024-11-27 17:35:08.863753] bdev_raid.c:2896:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:13:37.846 [2024-11-27 17:35:08.968496] bdev_raid.c:2558:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:13:37.846 [2024-11-27 17:35:08.971776] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:38.676 89.86 IOPS, 269.57 MiB/s [2024-11-27T17:35:09.868Z] 17:35:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:13:38.676 17:35:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:38.676 17:35:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:38.676 17:35:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:38.676 17:35:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:38.676 17:35:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:38.676 17:35:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:38.676 17:35:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:38.676 17:35:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:38.676 17:35:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:38.676 17:35:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:38.676 17:35:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:38.676 "name": "raid_bdev1", 00:13:38.676 "uuid": "da8a894f-f09b-4dea-ad41-aabdb7da2f87", 00:13:38.676 "strip_size_kb": 0, 00:13:38.676 "state": "online", 00:13:38.676 "raid_level": "raid1", 00:13:38.676 "superblock": true, 00:13:38.676 "num_base_bdevs": 4, 00:13:38.676 "num_base_bdevs_discovered": 3, 00:13:38.676 "num_base_bdevs_operational": 3, 00:13:38.676 "base_bdevs_list": [ 00:13:38.676 { 00:13:38.676 "name": "spare", 00:13:38.676 "uuid": "713e7c40-9159-54c2-9e60-a67bcff7b97c", 00:13:38.676 "is_configured": true, 00:13:38.676 "data_offset": 2048, 00:13:38.676 "data_size": 63488 00:13:38.676 }, 00:13:38.676 { 00:13:38.676 "name": null, 00:13:38.676 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:38.676 "is_configured": false, 00:13:38.676 "data_offset": 0, 00:13:38.676 "data_size": 63488 00:13:38.676 }, 00:13:38.676 { 00:13:38.676 "name": "BaseBdev3", 00:13:38.676 "uuid": "e9304de8-d30a-5749-92d8-8df9c95094de", 00:13:38.676 "is_configured": true, 00:13:38.676 "data_offset": 2048, 00:13:38.676 "data_size": 63488 00:13:38.676 }, 00:13:38.676 { 00:13:38.676 "name": "BaseBdev4", 00:13:38.676 "uuid": "69df1b46-7233-5fb6-b5b6-39ecbfd83e27", 00:13:38.676 "is_configured": true, 00:13:38.676 "data_offset": 2048, 00:13:38.676 "data_size": 63488 00:13:38.676 } 00:13:38.676 ] 00:13:38.676 }' 00:13:38.676 17:35:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:38.676 17:35:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:13:38.676 17:35:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:38.676 17:35:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:13:38.676 17:35:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@709 -- # break 00:13:38.676 17:35:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:13:38.676 17:35:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:38.676 17:35:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:13:38.676 17:35:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:13:38.676 17:35:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:38.677 17:35:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:38.677 17:35:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:38.677 17:35:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:38.677 17:35:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:38.677 17:35:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:38.677 17:35:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:38.677 "name": "raid_bdev1", 00:13:38.677 "uuid": "da8a894f-f09b-4dea-ad41-aabdb7da2f87", 00:13:38.677 "strip_size_kb": 0, 00:13:38.677 "state": "online", 00:13:38.677 "raid_level": "raid1", 00:13:38.677 "superblock": true, 00:13:38.677 "num_base_bdevs": 4, 00:13:38.677 "num_base_bdevs_discovered": 3, 00:13:38.677 "num_base_bdevs_operational": 3, 00:13:38.677 "base_bdevs_list": [ 00:13:38.677 { 00:13:38.677 "name": "spare", 00:13:38.677 "uuid": "713e7c40-9159-54c2-9e60-a67bcff7b97c", 00:13:38.677 "is_configured": true, 00:13:38.677 "data_offset": 2048, 00:13:38.677 "data_size": 63488 00:13:38.677 }, 00:13:38.677 { 00:13:38.677 "name": null, 00:13:38.677 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:38.677 "is_configured": false, 00:13:38.677 "data_offset": 0, 00:13:38.677 "data_size": 63488 00:13:38.677 }, 00:13:38.677 { 00:13:38.677 "name": "BaseBdev3", 00:13:38.677 "uuid": "e9304de8-d30a-5749-92d8-8df9c95094de", 00:13:38.677 "is_configured": true, 00:13:38.677 "data_offset": 2048, 00:13:38.677 "data_size": 63488 00:13:38.677 }, 00:13:38.677 { 00:13:38.677 "name": "BaseBdev4", 00:13:38.677 "uuid": "69df1b46-7233-5fb6-b5b6-39ecbfd83e27", 00:13:38.677 "is_configured": true, 00:13:38.677 "data_offset": 2048, 00:13:38.677 "data_size": 63488 00:13:38.677 } 00:13:38.677 ] 00:13:38.677 }' 00:13:38.677 17:35:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:38.937 17:35:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:13:38.937 17:35:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:38.937 17:35:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:13:38.937 17:35:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:13:38.937 17:35:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:38.937 17:35:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:38.937 17:35:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:38.937 17:35:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:38.937 17:35:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:38.937 17:35:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:38.937 17:35:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:38.937 17:35:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:38.937 17:35:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:38.937 17:35:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:38.937 17:35:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:38.937 17:35:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:38.937 17:35:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:38.937 17:35:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:38.937 17:35:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:38.937 "name": "raid_bdev1", 00:13:38.937 "uuid": "da8a894f-f09b-4dea-ad41-aabdb7da2f87", 00:13:38.937 "strip_size_kb": 0, 00:13:38.937 "state": "online", 00:13:38.937 "raid_level": "raid1", 00:13:38.937 "superblock": true, 00:13:38.937 "num_base_bdevs": 4, 00:13:38.937 "num_base_bdevs_discovered": 3, 00:13:38.937 "num_base_bdevs_operational": 3, 00:13:38.937 "base_bdevs_list": [ 00:13:38.937 { 00:13:38.937 "name": "spare", 00:13:38.937 "uuid": "713e7c40-9159-54c2-9e60-a67bcff7b97c", 00:13:38.937 "is_configured": true, 00:13:38.937 "data_offset": 2048, 00:13:38.937 "data_size": 63488 00:13:38.937 }, 00:13:38.937 { 00:13:38.937 "name": null, 00:13:38.937 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:38.937 "is_configured": false, 00:13:38.937 "data_offset": 0, 00:13:38.937 "data_size": 63488 00:13:38.937 }, 00:13:38.937 { 00:13:38.937 "name": "BaseBdev3", 00:13:38.937 "uuid": "e9304de8-d30a-5749-92d8-8df9c95094de", 00:13:38.937 "is_configured": true, 00:13:38.937 "data_offset": 2048, 00:13:38.937 "data_size": 63488 00:13:38.937 }, 00:13:38.937 { 00:13:38.937 "name": "BaseBdev4", 00:13:38.937 "uuid": "69df1b46-7233-5fb6-b5b6-39ecbfd83e27", 00:13:38.937 "is_configured": true, 00:13:38.937 "data_offset": 2048, 00:13:38.937 "data_size": 63488 00:13:38.937 } 00:13:38.937 ] 00:13:38.937 }' 00:13:38.937 17:35:09 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:38.937 17:35:09 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:39.198 17:35:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:13:39.198 17:35:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:39.198 17:35:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:39.198 [2024-11-27 17:35:10.320428] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:13:39.198 [2024-11-27 17:35:10.320512] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:13:39.198 00:13:39.198 Latency(us) 00:13:39.198 [2024-11-27T17:35:10.390Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:13:39.198 Job: raid_bdev1 (Core Mask 0x1, workload: randrw, percentage: 50, depth: 2, IO size: 3145728) 00:13:39.198 raid_bdev1 : 7.93 82.84 248.52 0.00 0.00 17397.26 264.72 111268.11 00:13:39.198 [2024-11-27T17:35:10.390Z] =================================================================================================================== 00:13:39.198 [2024-11-27T17:35:10.390Z] Total : 82.84 248.52 0.00 0.00 17397.26 264.72 111268.11 00:13:39.198 [2024-11-27 17:35:10.363358] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:39.198 [2024-11-27 17:35:10.363429] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:13:39.198 [2024-11-27 17:35:10.363546] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:13:39.198 [2024-11-27 17:35:10.363655] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:13:39.198 { 00:13:39.198 "results": [ 00:13:39.198 { 00:13:39.198 "job": "raid_bdev1", 00:13:39.198 "core_mask": "0x1", 00:13:39.198 "workload": "randrw", 00:13:39.198 "percentage": 50, 00:13:39.198 "status": "finished", 00:13:39.198 "queue_depth": 2, 00:13:39.198 "io_size": 3145728, 00:13:39.198 "runtime": 7.930925, 00:13:39.198 "iops": 82.84027399073878, 00:13:39.198 "mibps": 248.52082197221634, 00:13:39.198 "io_failed": 0, 00:13:39.198 "io_timeout": 0, 00:13:39.198 "avg_latency_us": 17397.255871268768, 00:13:39.198 "min_latency_us": 264.71965065502184, 00:13:39.198 "max_latency_us": 111268.10829694323 00:13:39.198 } 00:13:39.198 ], 00:13:39.198 "core_count": 1 00:13:39.198 } 00:13:39.198 17:35:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:39.198 17:35:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:39.198 17:35:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@720 -- # jq length 00:13:39.198 17:35:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:39.198 17:35:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:39.198 17:35:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:39.458 17:35:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:13:39.458 17:35:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:13:39.458 17:35:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@723 -- # '[' true = true ']' 00:13:39.458 17:35:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@725 -- # nbd_start_disks /var/tmp/spdk.sock spare /dev/nbd0 00:13:39.458 17:35:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:13:39.458 17:35:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # bdev_list=('spare') 00:13:39.458 17:35:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:13:39.458 17:35:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:13:39.458 17:35:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:13:39.458 17:35:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@12 -- # local i 00:13:39.458 17:35:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:13:39.458 17:35:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:13:39.458 17:35:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd0 00:13:39.458 /dev/nbd0 00:13:39.458 17:35:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:13:39.458 17:35:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:13:39.458 17:35:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:13:39.458 17:35:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@869 -- # local i 00:13:39.458 17:35:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:13:39.458 17:35:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:13:39.458 17:35:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:13:39.458 17:35:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@873 -- # break 00:13:39.458 17:35:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:13:39.718 17:35:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:13:39.718 17:35:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:13:39.718 1+0 records in 00:13:39.718 1+0 records out 00:13:39.718 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000336838 s, 12.2 MB/s 00:13:39.718 17:35:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:39.718 17:35:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@886 -- # size=4096 00:13:39.718 17:35:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:39.718 17:35:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:13:39.718 17:35:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@889 -- # return 0 00:13:39.718 17:35:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:13:39.718 17:35:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:13:39.718 17:35:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@726 -- # for bdev in "${base_bdevs[@]:1}" 00:13:39.718 17:35:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@727 -- # '[' -z '' ']' 00:13:39.718 17:35:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@728 -- # continue 00:13:39.718 17:35:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@726 -- # for bdev in "${base_bdevs[@]:1}" 00:13:39.718 17:35:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@727 -- # '[' -z BaseBdev3 ']' 00:13:39.718 17:35:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@730 -- # nbd_start_disks /var/tmp/spdk.sock BaseBdev3 /dev/nbd1 00:13:39.718 17:35:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:13:39.718 17:35:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev3') 00:13:39.718 17:35:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:13:39.718 17:35:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd1') 00:13:39.718 17:35:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:13:39.718 17:35:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@12 -- # local i 00:13:39.718 17:35:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:13:39.718 17:35:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:13:39.718 17:35:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev3 /dev/nbd1 00:13:39.718 /dev/nbd1 00:13:39.718 17:35:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:13:39.718 17:35:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:13:39.718 17:35:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@868 -- # local nbd_name=nbd1 00:13:39.718 17:35:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@869 -- # local i 00:13:39.718 17:35:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:13:39.718 17:35:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:13:39.718 17:35:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@872 -- # grep -q -w nbd1 /proc/partitions 00:13:39.718 17:35:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@873 -- # break 00:13:39.718 17:35:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:13:39.718 17:35:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:13:39.718 17:35:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@885 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:13:39.718 1+0 records in 00:13:39.718 1+0 records out 00:13:39.718 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000496569 s, 8.2 MB/s 00:13:39.718 17:35:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:39.718 17:35:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@886 -- # size=4096 00:13:39.718 17:35:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:39.718 17:35:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:13:39.718 17:35:10 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@889 -- # return 0 00:13:39.718 17:35:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:13:39.979 17:35:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:13:39.979 17:35:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@731 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:13:39.979 17:35:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@732 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd1 00:13:39.979 17:35:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:13:39.979 17:35:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd1') 00:13:39.979 17:35:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:13:39.979 17:35:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@51 -- # local i 00:13:39.979 17:35:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:13:39.979 17:35:10 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:13:39.979 17:35:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:13:39.979 17:35:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:13:39.979 17:35:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:13:40.239 17:35:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:13:40.239 17:35:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:13:40.239 17:35:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:13:40.239 17:35:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@41 -- # break 00:13:40.239 17:35:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@45 -- # return 0 00:13:40.239 17:35:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@726 -- # for bdev in "${base_bdevs[@]:1}" 00:13:40.239 17:35:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@727 -- # '[' -z BaseBdev4 ']' 00:13:40.239 17:35:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@730 -- # nbd_start_disks /var/tmp/spdk.sock BaseBdev4 /dev/nbd1 00:13:40.239 17:35:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:13:40.239 17:35:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev4') 00:13:40.239 17:35:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@10 -- # local bdev_list 00:13:40.239 17:35:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd1') 00:13:40.239 17:35:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@11 -- # local nbd_list 00:13:40.239 17:35:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@12 -- # local i 00:13:40.239 17:35:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:13:40.239 17:35:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:13:40.239 17:35:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev4 /dev/nbd1 00:13:40.239 /dev/nbd1 00:13:40.239 17:35:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:13:40.239 17:35:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:13:40.239 17:35:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@868 -- # local nbd_name=nbd1 00:13:40.239 17:35:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@869 -- # local i 00:13:40.239 17:35:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:13:40.239 17:35:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:13:40.239 17:35:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@872 -- # grep -q -w nbd1 /proc/partitions 00:13:40.239 17:35:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@873 -- # break 00:13:40.239 17:35:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:13:40.239 17:35:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:13:40.239 17:35:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@885 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:13:40.239 1+0 records in 00:13:40.239 1+0 records out 00:13:40.239 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000402512 s, 10.2 MB/s 00:13:40.239 17:35:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:40.239 17:35:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@886 -- # size=4096 00:13:40.239 17:35:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:13:40.239 17:35:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:13:40.239 17:35:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@889 -- # return 0 00:13:40.239 17:35:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:13:40.239 17:35:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:13:40.239 17:35:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@731 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:13:40.499 17:35:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@732 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd1 00:13:40.499 17:35:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:13:40.499 17:35:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd1') 00:13:40.499 17:35:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:13:40.499 17:35:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@51 -- # local i 00:13:40.499 17:35:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:13:40.499 17:35:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:13:40.499 17:35:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:13:40.499 17:35:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:13:40.499 17:35:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:13:40.499 17:35:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:13:40.500 17:35:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:13:40.500 17:35:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:13:40.500 17:35:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@41 -- # break 00:13:40.500 17:35:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@45 -- # return 0 00:13:40.500 17:35:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@734 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:13:40.500 17:35:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:13:40.500 17:35:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:13:40.500 17:35:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@50 -- # local nbd_list 00:13:40.500 17:35:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@51 -- # local i 00:13:40.500 17:35:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:13:40.500 17:35:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:13:40.760 17:35:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:13:40.760 17:35:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:13:40.760 17:35:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:13:40.760 17:35:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:13:40.760 17:35:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:13:40.760 17:35:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:13:40.760 17:35:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@41 -- # break 00:13:40.760 17:35:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/nbd_common.sh@45 -- # return 0 00:13:40.760 17:35:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:13:40.760 17:35:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:13:40.760 17:35:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:40.760 17:35:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:40.760 17:35:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:40.760 17:35:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:13:40.760 17:35:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:40.760 17:35:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:40.760 [2024-11-27 17:35:11.892814] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:13:40.760 [2024-11-27 17:35:11.892875] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:40.760 [2024-11-27 17:35:11.892899] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b180 00:13:40.760 [2024-11-27 17:35:11.892907] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:40.760 [2024-11-27 17:35:11.895108] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:40.760 [2024-11-27 17:35:11.895155] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:13:40.760 [2024-11-27 17:35:11.895240] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:13:40.760 [2024-11-27 17:35:11.895278] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:13:40.760 [2024-11-27 17:35:11.895408] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:13:40.760 [2024-11-27 17:35:11.895497] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:13:40.760 spare 00:13:40.760 17:35:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:40.760 17:35:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:13:40.760 17:35:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:40.760 17:35:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:41.020 [2024-11-27 17:35:11.995393] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001580 00:13:41.020 [2024-11-27 17:35:11.995415] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 63488, blocklen 512 00:13:41.020 [2024-11-27 17:35:11.995661] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000337b0 00:13:41.020 [2024-11-27 17:35:11.995788] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001580 00:13:41.020 [2024-11-27 17:35:11.995805] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001580 00:13:41.020 [2024-11-27 17:35:11.995918] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:41.020 17:35:11 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:41.020 17:35:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 3 00:13:41.020 17:35:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:41.020 17:35:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:41.020 17:35:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:41.020 17:35:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:41.020 17:35:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:41.020 17:35:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:41.020 17:35:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:41.020 17:35:11 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:41.020 17:35:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:41.020 17:35:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:41.020 17:35:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:41.020 17:35:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:41.020 17:35:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:41.020 17:35:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:41.020 17:35:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:41.020 "name": "raid_bdev1", 00:13:41.020 "uuid": "da8a894f-f09b-4dea-ad41-aabdb7da2f87", 00:13:41.020 "strip_size_kb": 0, 00:13:41.020 "state": "online", 00:13:41.020 "raid_level": "raid1", 00:13:41.020 "superblock": true, 00:13:41.020 "num_base_bdevs": 4, 00:13:41.020 "num_base_bdevs_discovered": 3, 00:13:41.020 "num_base_bdevs_operational": 3, 00:13:41.020 "base_bdevs_list": [ 00:13:41.020 { 00:13:41.020 "name": "spare", 00:13:41.020 "uuid": "713e7c40-9159-54c2-9e60-a67bcff7b97c", 00:13:41.020 "is_configured": true, 00:13:41.020 "data_offset": 2048, 00:13:41.020 "data_size": 63488 00:13:41.020 }, 00:13:41.020 { 00:13:41.020 "name": null, 00:13:41.020 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:41.020 "is_configured": false, 00:13:41.020 "data_offset": 2048, 00:13:41.020 "data_size": 63488 00:13:41.020 }, 00:13:41.020 { 00:13:41.020 "name": "BaseBdev3", 00:13:41.020 "uuid": "e9304de8-d30a-5749-92d8-8df9c95094de", 00:13:41.020 "is_configured": true, 00:13:41.020 "data_offset": 2048, 00:13:41.020 "data_size": 63488 00:13:41.020 }, 00:13:41.020 { 00:13:41.020 "name": "BaseBdev4", 00:13:41.020 "uuid": "69df1b46-7233-5fb6-b5b6-39ecbfd83e27", 00:13:41.020 "is_configured": true, 00:13:41.020 "data_offset": 2048, 00:13:41.020 "data_size": 63488 00:13:41.020 } 00:13:41.020 ] 00:13:41.020 }' 00:13:41.020 17:35:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:41.020 17:35:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:41.280 17:35:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:13:41.280 17:35:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:41.280 17:35:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:13:41.280 17:35:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:13:41.280 17:35:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:41.280 17:35:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:41.280 17:35:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:41.280 17:35:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:41.280 17:35:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:41.280 17:35:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:41.540 17:35:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:41.540 "name": "raid_bdev1", 00:13:41.540 "uuid": "da8a894f-f09b-4dea-ad41-aabdb7da2f87", 00:13:41.540 "strip_size_kb": 0, 00:13:41.540 "state": "online", 00:13:41.540 "raid_level": "raid1", 00:13:41.540 "superblock": true, 00:13:41.540 "num_base_bdevs": 4, 00:13:41.540 "num_base_bdevs_discovered": 3, 00:13:41.540 "num_base_bdevs_operational": 3, 00:13:41.540 "base_bdevs_list": [ 00:13:41.540 { 00:13:41.540 "name": "spare", 00:13:41.540 "uuid": "713e7c40-9159-54c2-9e60-a67bcff7b97c", 00:13:41.540 "is_configured": true, 00:13:41.540 "data_offset": 2048, 00:13:41.540 "data_size": 63488 00:13:41.540 }, 00:13:41.540 { 00:13:41.540 "name": null, 00:13:41.540 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:41.540 "is_configured": false, 00:13:41.540 "data_offset": 2048, 00:13:41.540 "data_size": 63488 00:13:41.540 }, 00:13:41.540 { 00:13:41.540 "name": "BaseBdev3", 00:13:41.540 "uuid": "e9304de8-d30a-5749-92d8-8df9c95094de", 00:13:41.540 "is_configured": true, 00:13:41.540 "data_offset": 2048, 00:13:41.540 "data_size": 63488 00:13:41.540 }, 00:13:41.540 { 00:13:41.540 "name": "BaseBdev4", 00:13:41.540 "uuid": "69df1b46-7233-5fb6-b5b6-39ecbfd83e27", 00:13:41.540 "is_configured": true, 00:13:41.540 "data_offset": 2048, 00:13:41.540 "data_size": 63488 00:13:41.540 } 00:13:41.540 ] 00:13:41.540 }' 00:13:41.540 17:35:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:41.540 17:35:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:13:41.540 17:35:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:41.540 17:35:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:13:41.540 17:35:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:41.540 17:35:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:41.540 17:35:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:41.540 17:35:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:13:41.540 17:35:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:41.540 17:35:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:13:41.540 17:35:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:13:41.540 17:35:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:41.540 17:35:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:41.540 [2024-11-27 17:35:12.627639] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:41.540 17:35:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:41.540 17:35:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:13:41.540 17:35:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:41.540 17:35:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:41.540 17:35:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:41.540 17:35:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:41.540 17:35:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:41.540 17:35:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:41.540 17:35:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:41.540 17:35:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:41.540 17:35:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:41.540 17:35:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:41.540 17:35:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:41.541 17:35:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:41.541 17:35:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:41.541 17:35:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:41.541 17:35:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:41.541 "name": "raid_bdev1", 00:13:41.541 "uuid": "da8a894f-f09b-4dea-ad41-aabdb7da2f87", 00:13:41.541 "strip_size_kb": 0, 00:13:41.541 "state": "online", 00:13:41.541 "raid_level": "raid1", 00:13:41.541 "superblock": true, 00:13:41.541 "num_base_bdevs": 4, 00:13:41.541 "num_base_bdevs_discovered": 2, 00:13:41.541 "num_base_bdevs_operational": 2, 00:13:41.541 "base_bdevs_list": [ 00:13:41.541 { 00:13:41.541 "name": null, 00:13:41.541 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:41.541 "is_configured": false, 00:13:41.541 "data_offset": 0, 00:13:41.541 "data_size": 63488 00:13:41.541 }, 00:13:41.541 { 00:13:41.541 "name": null, 00:13:41.541 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:41.541 "is_configured": false, 00:13:41.541 "data_offset": 2048, 00:13:41.541 "data_size": 63488 00:13:41.541 }, 00:13:41.541 { 00:13:41.541 "name": "BaseBdev3", 00:13:41.541 "uuid": "e9304de8-d30a-5749-92d8-8df9c95094de", 00:13:41.541 "is_configured": true, 00:13:41.541 "data_offset": 2048, 00:13:41.541 "data_size": 63488 00:13:41.541 }, 00:13:41.541 { 00:13:41.541 "name": "BaseBdev4", 00:13:41.541 "uuid": "69df1b46-7233-5fb6-b5b6-39ecbfd83e27", 00:13:41.541 "is_configured": true, 00:13:41.541 "data_offset": 2048, 00:13:41.541 "data_size": 63488 00:13:41.541 } 00:13:41.541 ] 00:13:41.541 }' 00:13:41.541 17:35:12 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:41.541 17:35:12 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:42.109 17:35:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:13:42.109 17:35:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:42.109 17:35:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:42.109 [2024-11-27 17:35:13.070966] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:13:42.109 [2024-11-27 17:35:13.071132] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (5) smaller than existing raid bdev raid_bdev1 (6) 00:13:42.109 [2024-11-27 17:35:13.071199] bdev_raid.c:3748:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:13:42.109 [2024-11-27 17:35:13.071267] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:13:42.109 [2024-11-27 17:35:13.074810] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000033880 00:13:42.109 [2024-11-27 17:35:13.076656] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:13:42.109 17:35:13 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:42.109 17:35:13 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@757 -- # sleep 1 00:13:43.048 17:35:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:43.048 17:35:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:43.048 17:35:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:43.048 17:35:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:43.048 17:35:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:43.048 17:35:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:43.048 17:35:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:43.048 17:35:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:43.048 17:35:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:43.048 17:35:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:43.048 17:35:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:43.048 "name": "raid_bdev1", 00:13:43.048 "uuid": "da8a894f-f09b-4dea-ad41-aabdb7da2f87", 00:13:43.048 "strip_size_kb": 0, 00:13:43.048 "state": "online", 00:13:43.048 "raid_level": "raid1", 00:13:43.048 "superblock": true, 00:13:43.048 "num_base_bdevs": 4, 00:13:43.048 "num_base_bdevs_discovered": 3, 00:13:43.048 "num_base_bdevs_operational": 3, 00:13:43.048 "process": { 00:13:43.048 "type": "rebuild", 00:13:43.048 "target": "spare", 00:13:43.048 "progress": { 00:13:43.048 "blocks": 20480, 00:13:43.048 "percent": 32 00:13:43.048 } 00:13:43.048 }, 00:13:43.048 "base_bdevs_list": [ 00:13:43.048 { 00:13:43.049 "name": "spare", 00:13:43.049 "uuid": "713e7c40-9159-54c2-9e60-a67bcff7b97c", 00:13:43.049 "is_configured": true, 00:13:43.049 "data_offset": 2048, 00:13:43.049 "data_size": 63488 00:13:43.049 }, 00:13:43.049 { 00:13:43.049 "name": null, 00:13:43.049 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:43.049 "is_configured": false, 00:13:43.049 "data_offset": 2048, 00:13:43.049 "data_size": 63488 00:13:43.049 }, 00:13:43.049 { 00:13:43.049 "name": "BaseBdev3", 00:13:43.049 "uuid": "e9304de8-d30a-5749-92d8-8df9c95094de", 00:13:43.049 "is_configured": true, 00:13:43.049 "data_offset": 2048, 00:13:43.049 "data_size": 63488 00:13:43.049 }, 00:13:43.049 { 00:13:43.049 "name": "BaseBdev4", 00:13:43.049 "uuid": "69df1b46-7233-5fb6-b5b6-39ecbfd83e27", 00:13:43.049 "is_configured": true, 00:13:43.049 "data_offset": 2048, 00:13:43.049 "data_size": 63488 00:13:43.049 } 00:13:43.049 ] 00:13:43.049 }' 00:13:43.049 17:35:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:43.049 17:35:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:43.049 17:35:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:43.049 17:35:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:43.049 17:35:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:13:43.309 17:35:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:43.309 17:35:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:43.309 [2024-11-27 17:35:14.243427] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:43.309 [2024-11-27 17:35:14.280592] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:13:43.309 [2024-11-27 17:35:14.280644] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:43.309 [2024-11-27 17:35:14.280663] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:43.309 [2024-11-27 17:35:14.280670] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:13:43.309 17:35:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:43.309 17:35:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:13:43.309 17:35:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:43.309 17:35:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:43.309 17:35:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:43.309 17:35:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:43.309 17:35:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:43.309 17:35:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:43.309 17:35:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:43.309 17:35:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:43.309 17:35:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:43.309 17:35:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:43.309 17:35:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:43.309 17:35:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:43.309 17:35:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:43.309 17:35:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:43.309 17:35:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:43.309 "name": "raid_bdev1", 00:13:43.309 "uuid": "da8a894f-f09b-4dea-ad41-aabdb7da2f87", 00:13:43.309 "strip_size_kb": 0, 00:13:43.309 "state": "online", 00:13:43.309 "raid_level": "raid1", 00:13:43.309 "superblock": true, 00:13:43.309 "num_base_bdevs": 4, 00:13:43.309 "num_base_bdevs_discovered": 2, 00:13:43.309 "num_base_bdevs_operational": 2, 00:13:43.309 "base_bdevs_list": [ 00:13:43.309 { 00:13:43.309 "name": null, 00:13:43.309 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:43.309 "is_configured": false, 00:13:43.309 "data_offset": 0, 00:13:43.309 "data_size": 63488 00:13:43.309 }, 00:13:43.309 { 00:13:43.309 "name": null, 00:13:43.309 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:43.309 "is_configured": false, 00:13:43.309 "data_offset": 2048, 00:13:43.309 "data_size": 63488 00:13:43.309 }, 00:13:43.309 { 00:13:43.309 "name": "BaseBdev3", 00:13:43.309 "uuid": "e9304de8-d30a-5749-92d8-8df9c95094de", 00:13:43.309 "is_configured": true, 00:13:43.309 "data_offset": 2048, 00:13:43.309 "data_size": 63488 00:13:43.309 }, 00:13:43.309 { 00:13:43.309 "name": "BaseBdev4", 00:13:43.309 "uuid": "69df1b46-7233-5fb6-b5b6-39ecbfd83e27", 00:13:43.309 "is_configured": true, 00:13:43.309 "data_offset": 2048, 00:13:43.309 "data_size": 63488 00:13:43.309 } 00:13:43.309 ] 00:13:43.309 }' 00:13:43.309 17:35:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:43.309 17:35:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:43.569 17:35:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:13:43.569 17:35:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:43.569 17:35:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:43.569 [2024-11-27 17:35:14.755409] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:13:43.569 [2024-11-27 17:35:14.755502] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:43.569 [2024-11-27 17:35:14.755546] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ba80 00:13:43.569 [2024-11-27 17:35:14.755575] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:43.569 [2024-11-27 17:35:14.755989] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:43.569 [2024-11-27 17:35:14.756056] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:13:43.569 [2024-11-27 17:35:14.756165] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:13:43.569 [2024-11-27 17:35:14.756206] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (5) smaller than existing raid bdev raid_bdev1 (6) 00:13:43.569 [2024-11-27 17:35:14.756243] bdev_raid.c:3748:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:13:43.569 [2024-11-27 17:35:14.756308] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:13:43.830 [2024-11-27 17:35:14.759442] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000033950 00:13:43.830 spare 00:13:43.830 17:35:14 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:43.830 17:35:14 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@764 -- # sleep 1 00:13:43.830 [2024-11-27 17:35:14.761305] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:13:44.770 17:35:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:13:44.770 17:35:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:44.770 17:35:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:13:44.770 17:35:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=spare 00:13:44.770 17:35:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:44.770 17:35:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:44.770 17:35:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:44.770 17:35:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:44.770 17:35:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:44.770 17:35:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:44.770 17:35:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:44.770 "name": "raid_bdev1", 00:13:44.770 "uuid": "da8a894f-f09b-4dea-ad41-aabdb7da2f87", 00:13:44.770 "strip_size_kb": 0, 00:13:44.770 "state": "online", 00:13:44.770 "raid_level": "raid1", 00:13:44.770 "superblock": true, 00:13:44.770 "num_base_bdevs": 4, 00:13:44.770 "num_base_bdevs_discovered": 3, 00:13:44.770 "num_base_bdevs_operational": 3, 00:13:44.770 "process": { 00:13:44.770 "type": "rebuild", 00:13:44.770 "target": "spare", 00:13:44.770 "progress": { 00:13:44.770 "blocks": 20480, 00:13:44.770 "percent": 32 00:13:44.770 } 00:13:44.770 }, 00:13:44.770 "base_bdevs_list": [ 00:13:44.770 { 00:13:44.770 "name": "spare", 00:13:44.770 "uuid": "713e7c40-9159-54c2-9e60-a67bcff7b97c", 00:13:44.770 "is_configured": true, 00:13:44.771 "data_offset": 2048, 00:13:44.771 "data_size": 63488 00:13:44.771 }, 00:13:44.771 { 00:13:44.771 "name": null, 00:13:44.771 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:44.771 "is_configured": false, 00:13:44.771 "data_offset": 2048, 00:13:44.771 "data_size": 63488 00:13:44.771 }, 00:13:44.771 { 00:13:44.771 "name": "BaseBdev3", 00:13:44.771 "uuid": "e9304de8-d30a-5749-92d8-8df9c95094de", 00:13:44.771 "is_configured": true, 00:13:44.771 "data_offset": 2048, 00:13:44.771 "data_size": 63488 00:13:44.771 }, 00:13:44.771 { 00:13:44.771 "name": "BaseBdev4", 00:13:44.771 "uuid": "69df1b46-7233-5fb6-b5b6-39ecbfd83e27", 00:13:44.771 "is_configured": true, 00:13:44.771 "data_offset": 2048, 00:13:44.771 "data_size": 63488 00:13:44.771 } 00:13:44.771 ] 00:13:44.771 }' 00:13:44.771 17:35:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:44.771 17:35:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:13:44.771 17:35:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:44.771 17:35:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:13:44.771 17:35:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:13:44.771 17:35:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:44.771 17:35:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:44.771 [2024-11-27 17:35:15.902450] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:45.031 [2024-11-27 17:35:15.965278] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:13:45.031 [2024-11-27 17:35:15.965337] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:45.031 [2024-11-27 17:35:15.965352] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:13:45.031 [2024-11-27 17:35:15.965361] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:13:45.031 17:35:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:45.031 17:35:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:13:45.031 17:35:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:45.031 17:35:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:45.031 17:35:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:45.031 17:35:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:45.031 17:35:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:45.031 17:35:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:45.031 17:35:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:45.031 17:35:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:45.031 17:35:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:45.031 17:35:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:45.031 17:35:15 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:45.031 17:35:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:45.031 17:35:15 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:45.031 17:35:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:45.031 17:35:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:45.031 "name": "raid_bdev1", 00:13:45.031 "uuid": "da8a894f-f09b-4dea-ad41-aabdb7da2f87", 00:13:45.031 "strip_size_kb": 0, 00:13:45.031 "state": "online", 00:13:45.031 "raid_level": "raid1", 00:13:45.031 "superblock": true, 00:13:45.031 "num_base_bdevs": 4, 00:13:45.031 "num_base_bdevs_discovered": 2, 00:13:45.031 "num_base_bdevs_operational": 2, 00:13:45.031 "base_bdevs_list": [ 00:13:45.031 { 00:13:45.031 "name": null, 00:13:45.031 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:45.031 "is_configured": false, 00:13:45.031 "data_offset": 0, 00:13:45.031 "data_size": 63488 00:13:45.031 }, 00:13:45.031 { 00:13:45.031 "name": null, 00:13:45.031 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:45.031 "is_configured": false, 00:13:45.031 "data_offset": 2048, 00:13:45.031 "data_size": 63488 00:13:45.031 }, 00:13:45.031 { 00:13:45.031 "name": "BaseBdev3", 00:13:45.031 "uuid": "e9304de8-d30a-5749-92d8-8df9c95094de", 00:13:45.031 "is_configured": true, 00:13:45.031 "data_offset": 2048, 00:13:45.031 "data_size": 63488 00:13:45.031 }, 00:13:45.031 { 00:13:45.031 "name": "BaseBdev4", 00:13:45.031 "uuid": "69df1b46-7233-5fb6-b5b6-39ecbfd83e27", 00:13:45.031 "is_configured": true, 00:13:45.031 "data_offset": 2048, 00:13:45.031 "data_size": 63488 00:13:45.031 } 00:13:45.031 ] 00:13:45.031 }' 00:13:45.031 17:35:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:45.031 17:35:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:45.291 17:35:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:13:45.291 17:35:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:45.291 17:35:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:13:45.291 17:35:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:13:45.291 17:35:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:45.291 17:35:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:45.291 17:35:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:45.291 17:35:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:45.291 17:35:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:45.291 17:35:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:45.291 17:35:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:45.291 "name": "raid_bdev1", 00:13:45.291 "uuid": "da8a894f-f09b-4dea-ad41-aabdb7da2f87", 00:13:45.291 "strip_size_kb": 0, 00:13:45.291 "state": "online", 00:13:45.291 "raid_level": "raid1", 00:13:45.291 "superblock": true, 00:13:45.291 "num_base_bdevs": 4, 00:13:45.291 "num_base_bdevs_discovered": 2, 00:13:45.291 "num_base_bdevs_operational": 2, 00:13:45.291 "base_bdevs_list": [ 00:13:45.291 { 00:13:45.291 "name": null, 00:13:45.291 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:45.291 "is_configured": false, 00:13:45.291 "data_offset": 0, 00:13:45.291 "data_size": 63488 00:13:45.291 }, 00:13:45.291 { 00:13:45.291 "name": null, 00:13:45.291 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:45.291 "is_configured": false, 00:13:45.291 "data_offset": 2048, 00:13:45.291 "data_size": 63488 00:13:45.291 }, 00:13:45.291 { 00:13:45.291 "name": "BaseBdev3", 00:13:45.291 "uuid": "e9304de8-d30a-5749-92d8-8df9c95094de", 00:13:45.291 "is_configured": true, 00:13:45.291 "data_offset": 2048, 00:13:45.291 "data_size": 63488 00:13:45.291 }, 00:13:45.291 { 00:13:45.291 "name": "BaseBdev4", 00:13:45.291 "uuid": "69df1b46-7233-5fb6-b5b6-39ecbfd83e27", 00:13:45.291 "is_configured": true, 00:13:45.291 "data_offset": 2048, 00:13:45.291 "data_size": 63488 00:13:45.291 } 00:13:45.291 ] 00:13:45.291 }' 00:13:45.291 17:35:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:45.551 17:35:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:13:45.551 17:35:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:45.551 17:35:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:13:45.551 17:35:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:13:45.551 17:35:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:45.551 17:35:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:45.551 17:35:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:45.551 17:35:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:13:45.551 17:35:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:45.551 17:35:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:45.551 [2024-11-27 17:35:16.559787] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:13:45.551 [2024-11-27 17:35:16.559886] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:13:45.551 [2024-11-27 17:35:16.559912] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000c080 00:13:45.551 [2024-11-27 17:35:16.559923] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:13:45.551 [2024-11-27 17:35:16.560298] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:13:45.551 [2024-11-27 17:35:16.560318] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:13:45.551 [2024-11-27 17:35:16.560391] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:13:45.551 [2024-11-27 17:35:16.560406] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (6) 00:13:45.551 [2024-11-27 17:35:16.560414] bdev_raid.c:3709:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:13:45.551 [2024-11-27 17:35:16.560424] bdev_raid.c:3884:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:13:45.551 BaseBdev1 00:13:45.551 17:35:16 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:45.551 17:35:16 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@775 -- # sleep 1 00:13:46.491 17:35:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:13:46.491 17:35:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:46.491 17:35:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:46.491 17:35:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:46.491 17:35:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:46.491 17:35:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:46.491 17:35:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:46.491 17:35:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:46.491 17:35:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:46.491 17:35:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:46.491 17:35:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:46.491 17:35:17 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:46.491 17:35:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:46.491 17:35:17 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:46.491 17:35:17 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:46.491 17:35:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:46.491 "name": "raid_bdev1", 00:13:46.491 "uuid": "da8a894f-f09b-4dea-ad41-aabdb7da2f87", 00:13:46.491 "strip_size_kb": 0, 00:13:46.491 "state": "online", 00:13:46.491 "raid_level": "raid1", 00:13:46.491 "superblock": true, 00:13:46.491 "num_base_bdevs": 4, 00:13:46.491 "num_base_bdevs_discovered": 2, 00:13:46.491 "num_base_bdevs_operational": 2, 00:13:46.491 "base_bdevs_list": [ 00:13:46.491 { 00:13:46.491 "name": null, 00:13:46.491 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:46.491 "is_configured": false, 00:13:46.491 "data_offset": 0, 00:13:46.491 "data_size": 63488 00:13:46.491 }, 00:13:46.491 { 00:13:46.491 "name": null, 00:13:46.491 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:46.491 "is_configured": false, 00:13:46.491 "data_offset": 2048, 00:13:46.491 "data_size": 63488 00:13:46.491 }, 00:13:46.491 { 00:13:46.491 "name": "BaseBdev3", 00:13:46.491 "uuid": "e9304de8-d30a-5749-92d8-8df9c95094de", 00:13:46.491 "is_configured": true, 00:13:46.491 "data_offset": 2048, 00:13:46.491 "data_size": 63488 00:13:46.491 }, 00:13:46.491 { 00:13:46.491 "name": "BaseBdev4", 00:13:46.491 "uuid": "69df1b46-7233-5fb6-b5b6-39ecbfd83e27", 00:13:46.491 "is_configured": true, 00:13:46.491 "data_offset": 2048, 00:13:46.491 "data_size": 63488 00:13:46.491 } 00:13:46.491 ] 00:13:46.491 }' 00:13:46.491 17:35:17 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:46.491 17:35:17 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:47.059 17:35:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:13:47.059 17:35:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:47.059 17:35:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:13:47.059 17:35:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:13:47.059 17:35:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:47.059 17:35:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:47.059 17:35:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:47.059 17:35:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:47.059 17:35:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:47.059 17:35:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:47.059 17:35:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:47.059 "name": "raid_bdev1", 00:13:47.059 "uuid": "da8a894f-f09b-4dea-ad41-aabdb7da2f87", 00:13:47.059 "strip_size_kb": 0, 00:13:47.059 "state": "online", 00:13:47.059 "raid_level": "raid1", 00:13:47.059 "superblock": true, 00:13:47.059 "num_base_bdevs": 4, 00:13:47.059 "num_base_bdevs_discovered": 2, 00:13:47.059 "num_base_bdevs_operational": 2, 00:13:47.059 "base_bdevs_list": [ 00:13:47.059 { 00:13:47.059 "name": null, 00:13:47.059 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:47.059 "is_configured": false, 00:13:47.059 "data_offset": 0, 00:13:47.059 "data_size": 63488 00:13:47.059 }, 00:13:47.059 { 00:13:47.059 "name": null, 00:13:47.059 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:47.059 "is_configured": false, 00:13:47.059 "data_offset": 2048, 00:13:47.059 "data_size": 63488 00:13:47.059 }, 00:13:47.059 { 00:13:47.059 "name": "BaseBdev3", 00:13:47.059 "uuid": "e9304de8-d30a-5749-92d8-8df9c95094de", 00:13:47.059 "is_configured": true, 00:13:47.059 "data_offset": 2048, 00:13:47.059 "data_size": 63488 00:13:47.059 }, 00:13:47.059 { 00:13:47.059 "name": "BaseBdev4", 00:13:47.059 "uuid": "69df1b46-7233-5fb6-b5b6-39ecbfd83e27", 00:13:47.059 "is_configured": true, 00:13:47.059 "data_offset": 2048, 00:13:47.059 "data_size": 63488 00:13:47.059 } 00:13:47.059 ] 00:13:47.059 }' 00:13:47.059 17:35:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:47.059 17:35:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:13:47.059 17:35:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:47.059 17:35:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:13:47.059 17:35:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:13:47.059 17:35:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@650 -- # local es=0 00:13:47.059 17:35:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:13:47.059 17:35:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:13:47.059 17:35:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:13:47.059 17:35:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:13:47.059 17:35:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:13:47.059 17:35:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:13:47.059 17:35:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:47.059 17:35:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:47.059 [2024-11-27 17:35:18.181923] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:13:47.059 [2024-11-27 17:35:18.182037] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (6) 00:13:47.059 [2024-11-27 17:35:18.182050] bdev_raid.c:3709:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:13:47.059 request: 00:13:47.059 { 00:13:47.059 "base_bdev": "BaseBdev1", 00:13:47.059 "raid_bdev": "raid_bdev1", 00:13:47.059 "method": "bdev_raid_add_base_bdev", 00:13:47.059 "req_id": 1 00:13:47.059 } 00:13:47.059 Got JSON-RPC error response 00:13:47.059 response: 00:13:47.059 { 00:13:47.059 "code": -22, 00:13:47.059 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:13:47.059 } 00:13:47.059 17:35:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:13:47.059 17:35:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@653 -- # es=1 00:13:47.059 17:35:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:13:47.059 17:35:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:13:47.059 17:35:18 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:13:47.059 17:35:18 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@779 -- # sleep 1 00:13:48.439 17:35:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:13:48.439 17:35:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:13:48.439 17:35:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:48.439 17:35:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:13:48.439 17:35:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:13:48.439 17:35:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:48.439 17:35:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:48.439 17:35:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:48.439 17:35:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:48.439 17:35:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:48.439 17:35:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:48.439 17:35:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:48.439 17:35:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:48.439 17:35:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:48.439 17:35:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:48.439 17:35:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:48.439 "name": "raid_bdev1", 00:13:48.439 "uuid": "da8a894f-f09b-4dea-ad41-aabdb7da2f87", 00:13:48.439 "strip_size_kb": 0, 00:13:48.439 "state": "online", 00:13:48.439 "raid_level": "raid1", 00:13:48.439 "superblock": true, 00:13:48.439 "num_base_bdevs": 4, 00:13:48.439 "num_base_bdevs_discovered": 2, 00:13:48.439 "num_base_bdevs_operational": 2, 00:13:48.439 "base_bdevs_list": [ 00:13:48.439 { 00:13:48.439 "name": null, 00:13:48.439 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:48.439 "is_configured": false, 00:13:48.439 "data_offset": 0, 00:13:48.439 "data_size": 63488 00:13:48.439 }, 00:13:48.439 { 00:13:48.439 "name": null, 00:13:48.439 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:48.439 "is_configured": false, 00:13:48.439 "data_offset": 2048, 00:13:48.439 "data_size": 63488 00:13:48.439 }, 00:13:48.439 { 00:13:48.439 "name": "BaseBdev3", 00:13:48.439 "uuid": "e9304de8-d30a-5749-92d8-8df9c95094de", 00:13:48.439 "is_configured": true, 00:13:48.439 "data_offset": 2048, 00:13:48.439 "data_size": 63488 00:13:48.439 }, 00:13:48.439 { 00:13:48.439 "name": "BaseBdev4", 00:13:48.439 "uuid": "69df1b46-7233-5fb6-b5b6-39ecbfd83e27", 00:13:48.439 "is_configured": true, 00:13:48.439 "data_offset": 2048, 00:13:48.439 "data_size": 63488 00:13:48.439 } 00:13:48.439 ] 00:13:48.439 }' 00:13:48.439 17:35:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:48.439 17:35:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:48.699 17:35:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:13:48.699 17:35:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:13:48.699 17:35:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:13:48.699 17:35:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@171 -- # local target=none 00:13:48.699 17:35:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:13:48.699 17:35:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:13:48.699 17:35:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:48.699 17:35:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:48.699 17:35:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:48.699 17:35:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:48.699 17:35:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:13:48.699 "name": "raid_bdev1", 00:13:48.699 "uuid": "da8a894f-f09b-4dea-ad41-aabdb7da2f87", 00:13:48.699 "strip_size_kb": 0, 00:13:48.699 "state": "online", 00:13:48.699 "raid_level": "raid1", 00:13:48.699 "superblock": true, 00:13:48.699 "num_base_bdevs": 4, 00:13:48.699 "num_base_bdevs_discovered": 2, 00:13:48.699 "num_base_bdevs_operational": 2, 00:13:48.699 "base_bdevs_list": [ 00:13:48.699 { 00:13:48.699 "name": null, 00:13:48.699 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:48.699 "is_configured": false, 00:13:48.699 "data_offset": 0, 00:13:48.699 "data_size": 63488 00:13:48.699 }, 00:13:48.699 { 00:13:48.699 "name": null, 00:13:48.699 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:48.699 "is_configured": false, 00:13:48.699 "data_offset": 2048, 00:13:48.699 "data_size": 63488 00:13:48.699 }, 00:13:48.699 { 00:13:48.699 "name": "BaseBdev3", 00:13:48.699 "uuid": "e9304de8-d30a-5749-92d8-8df9c95094de", 00:13:48.699 "is_configured": true, 00:13:48.699 "data_offset": 2048, 00:13:48.699 "data_size": 63488 00:13:48.699 }, 00:13:48.699 { 00:13:48.699 "name": "BaseBdev4", 00:13:48.699 "uuid": "69df1b46-7233-5fb6-b5b6-39ecbfd83e27", 00:13:48.699 "is_configured": true, 00:13:48.699 "data_offset": 2048, 00:13:48.699 "data_size": 63488 00:13:48.699 } 00:13:48.699 ] 00:13:48.699 }' 00:13:48.699 17:35:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:13:48.699 17:35:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:13:48.699 17:35:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:13:48.699 17:35:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:13:48.699 17:35:19 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@784 -- # killprocess 89599 00:13:48.699 17:35:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@950 -- # '[' -z 89599 ']' 00:13:48.699 17:35:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@954 -- # kill -0 89599 00:13:48.699 17:35:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@955 -- # uname 00:13:48.699 17:35:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:13:48.699 17:35:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 89599 00:13:48.699 17:35:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:13:48.699 killing process with pid 89599 00:13:48.699 Received shutdown signal, test time was about 17.451250 seconds 00:13:48.699 00:13:48.699 Latency(us) 00:13:48.699 [2024-11-27T17:35:19.891Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:13:48.699 [2024-11-27T17:35:19.891Z] =================================================================================================================== 00:13:48.699 [2024-11-27T17:35:19.891Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:13:48.699 17:35:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:13:48.699 17:35:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@968 -- # echo 'killing process with pid 89599' 00:13:48.699 17:35:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@969 -- # kill 89599 00:13:48.699 [2024-11-27 17:35:19.862260] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:13:48.699 [2024-11-27 17:35:19.862352] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:13:48.699 [2024-11-27 17:35:19.862414] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:13:48.699 [2024-11-27 17:35:19.862423] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state offline 00:13:48.699 17:35:19 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@974 -- # wait 89599 00:13:48.959 [2024-11-27 17:35:19.908134] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:13:48.959 17:35:20 bdev_raid.raid_rebuild_test_sb_io -- bdev/bdev_raid.sh@786 -- # return 0 00:13:48.959 00:13:48.959 real 0m19.476s 00:13:48.959 user 0m25.956s 00:13:48.959 sys 0m2.593s 00:13:48.959 17:35:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@1126 -- # xtrace_disable 00:13:48.959 17:35:20 bdev_raid.raid_rebuild_test_sb_io -- common/autotest_common.sh@10 -- # set +x 00:13:48.959 ************************************ 00:13:48.959 END TEST raid_rebuild_test_sb_io 00:13:48.959 ************************************ 00:13:49.219 17:35:20 bdev_raid -- bdev/bdev_raid.sh@985 -- # for n in {3..4} 00:13:49.219 17:35:20 bdev_raid -- bdev/bdev_raid.sh@986 -- # run_test raid5f_state_function_test raid_state_function_test raid5f 3 false 00:13:49.219 17:35:20 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:13:49.219 17:35:20 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:13:49.219 17:35:20 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:13:49.219 ************************************ 00:13:49.219 START TEST raid5f_state_function_test 00:13:49.219 ************************************ 00:13:49.219 17:35:20 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@1125 -- # raid_state_function_test raid5f 3 false 00:13:49.219 17:35:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=raid5f 00:13:49.219 17:35:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=3 00:13:49.219 17:35:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:13:49.219 17:35:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:13:49.219 17:35:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:13:49.219 17:35:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:13:49.219 17:35:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:13:49.219 17:35:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:13:49.219 17:35:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:13:49.219 17:35:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:13:49.219 17:35:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:13:49.219 17:35:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:13:49.219 17:35:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:13:49.219 17:35:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:13:49.219 17:35:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:13:49.219 17:35:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:13:49.219 17:35:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:13:49.219 17:35:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:13:49.219 17:35:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:13:49.219 17:35:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:13:49.219 17:35:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:13:49.219 17:35:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' raid5f '!=' raid1 ']' 00:13:49.219 17:35:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:13:49.219 17:35:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:13:49.219 17:35:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:13:49.219 17:35:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:13:49.219 17:35:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=90305 00:13:49.219 17:35:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:13:49.219 17:35:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 90305' 00:13:49.219 Process raid pid: 90305 00:13:49.219 17:35:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 90305 00:13:49.219 17:35:20 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@831 -- # '[' -z 90305 ']' 00:13:49.219 17:35:20 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:13:49.219 17:35:20 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:13:49.220 17:35:20 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:13:49.220 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:13:49.220 17:35:20 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:13:49.220 17:35:20 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:49.220 [2024-11-27 17:35:20.327512] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:13:49.220 [2024-11-27 17:35:20.327732] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:13:49.480 [2024-11-27 17:35:20.476747] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:13:49.480 [2024-11-27 17:35:20.522508] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:13:49.480 [2024-11-27 17:35:20.565078] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:13:49.480 [2024-11-27 17:35:20.565233] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:13:50.050 17:35:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:13:50.050 17:35:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@864 -- # return 0 00:13:50.050 17:35:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:13:50.050 17:35:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:50.050 17:35:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:50.050 [2024-11-27 17:35:21.131119] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:13:50.050 [2024-11-27 17:35:21.131180] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:13:50.050 [2024-11-27 17:35:21.131192] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:13:50.050 [2024-11-27 17:35:21.131201] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:13:50.050 [2024-11-27 17:35:21.131207] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:13:50.050 [2024-11-27 17:35:21.131219] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:13:50.050 17:35:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:50.050 17:35:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:50.050 17:35:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:50.050 17:35:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:50.050 17:35:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:50.050 17:35:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:50.050 17:35:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:50.050 17:35:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:50.050 17:35:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:50.050 17:35:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:50.050 17:35:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:50.050 17:35:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:50.050 17:35:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:50.050 17:35:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:50.050 17:35:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:50.050 17:35:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:50.050 17:35:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:50.050 "name": "Existed_Raid", 00:13:50.050 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:50.050 "strip_size_kb": 64, 00:13:50.050 "state": "configuring", 00:13:50.050 "raid_level": "raid5f", 00:13:50.050 "superblock": false, 00:13:50.050 "num_base_bdevs": 3, 00:13:50.050 "num_base_bdevs_discovered": 0, 00:13:50.050 "num_base_bdevs_operational": 3, 00:13:50.050 "base_bdevs_list": [ 00:13:50.050 { 00:13:50.050 "name": "BaseBdev1", 00:13:50.050 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:50.050 "is_configured": false, 00:13:50.050 "data_offset": 0, 00:13:50.050 "data_size": 0 00:13:50.050 }, 00:13:50.050 { 00:13:50.050 "name": "BaseBdev2", 00:13:50.050 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:50.050 "is_configured": false, 00:13:50.050 "data_offset": 0, 00:13:50.050 "data_size": 0 00:13:50.050 }, 00:13:50.050 { 00:13:50.050 "name": "BaseBdev3", 00:13:50.050 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:50.050 "is_configured": false, 00:13:50.050 "data_offset": 0, 00:13:50.050 "data_size": 0 00:13:50.050 } 00:13:50.050 ] 00:13:50.050 }' 00:13:50.050 17:35:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:50.050 17:35:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:50.620 17:35:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:13:50.620 17:35:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:50.620 17:35:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:50.620 [2024-11-27 17:35:21.538310] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:13:50.620 [2024-11-27 17:35:21.538385] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:13:50.620 17:35:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:50.620 17:35:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:13:50.620 17:35:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:50.620 17:35:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:50.620 [2024-11-27 17:35:21.550319] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:13:50.620 [2024-11-27 17:35:21.550390] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:13:50.620 [2024-11-27 17:35:21.550414] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:13:50.620 [2024-11-27 17:35:21.550435] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:13:50.620 [2024-11-27 17:35:21.550451] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:13:50.620 [2024-11-27 17:35:21.550472] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:13:50.620 17:35:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:50.620 17:35:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:13:50.620 17:35:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:50.620 17:35:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:50.620 [2024-11-27 17:35:21.571190] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:13:50.620 BaseBdev1 00:13:50.620 17:35:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:50.620 17:35:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:13:50.620 17:35:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:13:50.620 17:35:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:13:50.620 17:35:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@901 -- # local i 00:13:50.620 17:35:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:13:50.621 17:35:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:13:50.621 17:35:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:13:50.621 17:35:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:50.621 17:35:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:50.621 17:35:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:50.621 17:35:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:13:50.621 17:35:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:50.621 17:35:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:50.621 [ 00:13:50.621 { 00:13:50.621 "name": "BaseBdev1", 00:13:50.621 "aliases": [ 00:13:50.621 "d2979a6e-5b9e-4fca-a889-7abba68623fd" 00:13:50.621 ], 00:13:50.621 "product_name": "Malloc disk", 00:13:50.621 "block_size": 512, 00:13:50.621 "num_blocks": 65536, 00:13:50.621 "uuid": "d2979a6e-5b9e-4fca-a889-7abba68623fd", 00:13:50.621 "assigned_rate_limits": { 00:13:50.621 "rw_ios_per_sec": 0, 00:13:50.621 "rw_mbytes_per_sec": 0, 00:13:50.621 "r_mbytes_per_sec": 0, 00:13:50.621 "w_mbytes_per_sec": 0 00:13:50.621 }, 00:13:50.621 "claimed": true, 00:13:50.621 "claim_type": "exclusive_write", 00:13:50.621 "zoned": false, 00:13:50.621 "supported_io_types": { 00:13:50.621 "read": true, 00:13:50.621 "write": true, 00:13:50.621 "unmap": true, 00:13:50.621 "flush": true, 00:13:50.621 "reset": true, 00:13:50.621 "nvme_admin": false, 00:13:50.621 "nvme_io": false, 00:13:50.621 "nvme_io_md": false, 00:13:50.621 "write_zeroes": true, 00:13:50.621 "zcopy": true, 00:13:50.621 "get_zone_info": false, 00:13:50.621 "zone_management": false, 00:13:50.621 "zone_append": false, 00:13:50.621 "compare": false, 00:13:50.621 "compare_and_write": false, 00:13:50.621 "abort": true, 00:13:50.621 "seek_hole": false, 00:13:50.621 "seek_data": false, 00:13:50.621 "copy": true, 00:13:50.621 "nvme_iov_md": false 00:13:50.621 }, 00:13:50.621 "memory_domains": [ 00:13:50.621 { 00:13:50.621 "dma_device_id": "system", 00:13:50.621 "dma_device_type": 1 00:13:50.621 }, 00:13:50.621 { 00:13:50.621 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:50.621 "dma_device_type": 2 00:13:50.621 } 00:13:50.621 ], 00:13:50.621 "driver_specific": {} 00:13:50.621 } 00:13:50.621 ] 00:13:50.621 17:35:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:50.621 17:35:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:13:50.621 17:35:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:50.621 17:35:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:50.621 17:35:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:50.621 17:35:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:50.621 17:35:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:50.621 17:35:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:50.621 17:35:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:50.621 17:35:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:50.621 17:35:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:50.621 17:35:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:50.621 17:35:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:50.621 17:35:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:50.621 17:35:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:50.621 17:35:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:50.621 17:35:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:50.621 17:35:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:50.621 "name": "Existed_Raid", 00:13:50.621 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:50.621 "strip_size_kb": 64, 00:13:50.621 "state": "configuring", 00:13:50.621 "raid_level": "raid5f", 00:13:50.621 "superblock": false, 00:13:50.621 "num_base_bdevs": 3, 00:13:50.621 "num_base_bdevs_discovered": 1, 00:13:50.621 "num_base_bdevs_operational": 3, 00:13:50.621 "base_bdevs_list": [ 00:13:50.621 { 00:13:50.621 "name": "BaseBdev1", 00:13:50.621 "uuid": "d2979a6e-5b9e-4fca-a889-7abba68623fd", 00:13:50.621 "is_configured": true, 00:13:50.621 "data_offset": 0, 00:13:50.621 "data_size": 65536 00:13:50.621 }, 00:13:50.621 { 00:13:50.621 "name": "BaseBdev2", 00:13:50.621 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:50.621 "is_configured": false, 00:13:50.621 "data_offset": 0, 00:13:50.621 "data_size": 0 00:13:50.621 }, 00:13:50.621 { 00:13:50.621 "name": "BaseBdev3", 00:13:50.621 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:50.621 "is_configured": false, 00:13:50.621 "data_offset": 0, 00:13:50.621 "data_size": 0 00:13:50.621 } 00:13:50.621 ] 00:13:50.621 }' 00:13:50.621 17:35:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:50.621 17:35:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:50.882 17:35:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:13:50.882 17:35:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:50.882 17:35:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:50.882 [2024-11-27 17:35:22.058598] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:13:50.882 [2024-11-27 17:35:22.058639] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:13:50.882 17:35:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:50.882 17:35:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:13:50.882 17:35:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:50.882 17:35:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:50.882 [2024-11-27 17:35:22.070631] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:13:51.142 [2024-11-27 17:35:22.072409] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:13:51.142 [2024-11-27 17:35:22.072449] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:13:51.142 [2024-11-27 17:35:22.072457] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:13:51.142 [2024-11-27 17:35:22.072467] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:13:51.142 17:35:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:51.142 17:35:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:13:51.142 17:35:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:13:51.142 17:35:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:51.142 17:35:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:51.142 17:35:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:51.142 17:35:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:51.142 17:35:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:51.142 17:35:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:51.142 17:35:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:51.142 17:35:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:51.142 17:35:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:51.142 17:35:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:51.142 17:35:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:51.142 17:35:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:51.142 17:35:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:51.142 17:35:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:51.142 17:35:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:51.142 17:35:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:51.142 "name": "Existed_Raid", 00:13:51.142 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:51.142 "strip_size_kb": 64, 00:13:51.142 "state": "configuring", 00:13:51.142 "raid_level": "raid5f", 00:13:51.142 "superblock": false, 00:13:51.142 "num_base_bdevs": 3, 00:13:51.142 "num_base_bdevs_discovered": 1, 00:13:51.142 "num_base_bdevs_operational": 3, 00:13:51.142 "base_bdevs_list": [ 00:13:51.142 { 00:13:51.142 "name": "BaseBdev1", 00:13:51.142 "uuid": "d2979a6e-5b9e-4fca-a889-7abba68623fd", 00:13:51.142 "is_configured": true, 00:13:51.142 "data_offset": 0, 00:13:51.142 "data_size": 65536 00:13:51.142 }, 00:13:51.142 { 00:13:51.142 "name": "BaseBdev2", 00:13:51.142 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:51.142 "is_configured": false, 00:13:51.142 "data_offset": 0, 00:13:51.142 "data_size": 0 00:13:51.142 }, 00:13:51.142 { 00:13:51.142 "name": "BaseBdev3", 00:13:51.142 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:51.142 "is_configured": false, 00:13:51.142 "data_offset": 0, 00:13:51.142 "data_size": 0 00:13:51.142 } 00:13:51.142 ] 00:13:51.142 }' 00:13:51.142 17:35:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:51.142 17:35:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:51.402 17:35:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:13:51.402 17:35:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:51.402 17:35:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:51.402 [2024-11-27 17:35:22.570736] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:13:51.402 BaseBdev2 00:13:51.402 17:35:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:51.402 17:35:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:13:51.402 17:35:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:13:51.402 17:35:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:13:51.402 17:35:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@901 -- # local i 00:13:51.402 17:35:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:13:51.402 17:35:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:13:51.402 17:35:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:13:51.402 17:35:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:51.402 17:35:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:51.402 17:35:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:51.402 17:35:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:13:51.402 17:35:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:51.402 17:35:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:51.662 [ 00:13:51.662 { 00:13:51.662 "name": "BaseBdev2", 00:13:51.662 "aliases": [ 00:13:51.662 "b19c8055-6f61-42a2-8cd4-e54d87a7c6b9" 00:13:51.662 ], 00:13:51.662 "product_name": "Malloc disk", 00:13:51.662 "block_size": 512, 00:13:51.662 "num_blocks": 65536, 00:13:51.662 "uuid": "b19c8055-6f61-42a2-8cd4-e54d87a7c6b9", 00:13:51.662 "assigned_rate_limits": { 00:13:51.662 "rw_ios_per_sec": 0, 00:13:51.662 "rw_mbytes_per_sec": 0, 00:13:51.662 "r_mbytes_per_sec": 0, 00:13:51.662 "w_mbytes_per_sec": 0 00:13:51.662 }, 00:13:51.662 "claimed": true, 00:13:51.662 "claim_type": "exclusive_write", 00:13:51.662 "zoned": false, 00:13:51.662 "supported_io_types": { 00:13:51.662 "read": true, 00:13:51.662 "write": true, 00:13:51.662 "unmap": true, 00:13:51.662 "flush": true, 00:13:51.662 "reset": true, 00:13:51.662 "nvme_admin": false, 00:13:51.662 "nvme_io": false, 00:13:51.662 "nvme_io_md": false, 00:13:51.662 "write_zeroes": true, 00:13:51.662 "zcopy": true, 00:13:51.662 "get_zone_info": false, 00:13:51.662 "zone_management": false, 00:13:51.662 "zone_append": false, 00:13:51.662 "compare": false, 00:13:51.662 "compare_and_write": false, 00:13:51.662 "abort": true, 00:13:51.662 "seek_hole": false, 00:13:51.662 "seek_data": false, 00:13:51.662 "copy": true, 00:13:51.662 "nvme_iov_md": false 00:13:51.662 }, 00:13:51.662 "memory_domains": [ 00:13:51.662 { 00:13:51.662 "dma_device_id": "system", 00:13:51.662 "dma_device_type": 1 00:13:51.662 }, 00:13:51.662 { 00:13:51.662 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:51.662 "dma_device_type": 2 00:13:51.662 } 00:13:51.662 ], 00:13:51.662 "driver_specific": {} 00:13:51.662 } 00:13:51.662 ] 00:13:51.662 17:35:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:51.662 17:35:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:13:51.662 17:35:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:13:51.662 17:35:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:13:51.662 17:35:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:51.662 17:35:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:51.662 17:35:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:51.662 17:35:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:51.662 17:35:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:51.662 17:35:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:51.662 17:35:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:51.662 17:35:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:51.662 17:35:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:51.662 17:35:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:51.662 17:35:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:51.662 17:35:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:51.662 17:35:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:51.662 17:35:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:51.662 17:35:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:51.662 17:35:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:51.662 "name": "Existed_Raid", 00:13:51.662 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:51.662 "strip_size_kb": 64, 00:13:51.662 "state": "configuring", 00:13:51.662 "raid_level": "raid5f", 00:13:51.662 "superblock": false, 00:13:51.662 "num_base_bdevs": 3, 00:13:51.662 "num_base_bdevs_discovered": 2, 00:13:51.662 "num_base_bdevs_operational": 3, 00:13:51.662 "base_bdevs_list": [ 00:13:51.662 { 00:13:51.662 "name": "BaseBdev1", 00:13:51.662 "uuid": "d2979a6e-5b9e-4fca-a889-7abba68623fd", 00:13:51.662 "is_configured": true, 00:13:51.662 "data_offset": 0, 00:13:51.662 "data_size": 65536 00:13:51.662 }, 00:13:51.662 { 00:13:51.662 "name": "BaseBdev2", 00:13:51.662 "uuid": "b19c8055-6f61-42a2-8cd4-e54d87a7c6b9", 00:13:51.662 "is_configured": true, 00:13:51.662 "data_offset": 0, 00:13:51.662 "data_size": 65536 00:13:51.662 }, 00:13:51.662 { 00:13:51.662 "name": "BaseBdev3", 00:13:51.662 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:51.662 "is_configured": false, 00:13:51.662 "data_offset": 0, 00:13:51.662 "data_size": 0 00:13:51.662 } 00:13:51.662 ] 00:13:51.662 }' 00:13:51.662 17:35:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:51.662 17:35:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:51.923 17:35:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:13:51.923 17:35:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:51.923 17:35:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:51.923 [2024-11-27 17:35:23.072679] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:13:51.923 [2024-11-27 17:35:23.072786] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:13:51.923 [2024-11-27 17:35:23.072804] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 131072, blocklen 512 00:13:51.923 [2024-11-27 17:35:23.073058] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:13:51.923 [2024-11-27 17:35:23.073506] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:13:51.923 [2024-11-27 17:35:23.073534] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:13:51.923 [2024-11-27 17:35:23.073741] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:51.923 BaseBdev3 00:13:51.923 17:35:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:51.923 17:35:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:13:51.923 17:35:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:13:51.923 17:35:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:13:51.923 17:35:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@901 -- # local i 00:13:51.923 17:35:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:13:51.923 17:35:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:13:51.923 17:35:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:13:51.923 17:35:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:51.923 17:35:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:51.923 17:35:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:51.923 17:35:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:13:51.923 17:35:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:51.923 17:35:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:51.923 [ 00:13:51.923 { 00:13:51.923 "name": "BaseBdev3", 00:13:51.923 "aliases": [ 00:13:51.923 "9d89e122-371a-4286-85f0-3bdd4c80f01a" 00:13:51.923 ], 00:13:51.923 "product_name": "Malloc disk", 00:13:51.923 "block_size": 512, 00:13:51.923 "num_blocks": 65536, 00:13:51.923 "uuid": "9d89e122-371a-4286-85f0-3bdd4c80f01a", 00:13:51.923 "assigned_rate_limits": { 00:13:51.923 "rw_ios_per_sec": 0, 00:13:51.923 "rw_mbytes_per_sec": 0, 00:13:51.923 "r_mbytes_per_sec": 0, 00:13:51.923 "w_mbytes_per_sec": 0 00:13:51.923 }, 00:13:51.923 "claimed": true, 00:13:51.923 "claim_type": "exclusive_write", 00:13:51.923 "zoned": false, 00:13:51.923 "supported_io_types": { 00:13:51.923 "read": true, 00:13:51.923 "write": true, 00:13:51.923 "unmap": true, 00:13:51.923 "flush": true, 00:13:51.923 "reset": true, 00:13:51.923 "nvme_admin": false, 00:13:51.923 "nvme_io": false, 00:13:51.923 "nvme_io_md": false, 00:13:51.923 "write_zeroes": true, 00:13:51.923 "zcopy": true, 00:13:51.923 "get_zone_info": false, 00:13:51.923 "zone_management": false, 00:13:51.923 "zone_append": false, 00:13:51.923 "compare": false, 00:13:51.923 "compare_and_write": false, 00:13:51.923 "abort": true, 00:13:51.923 "seek_hole": false, 00:13:51.923 "seek_data": false, 00:13:51.923 "copy": true, 00:13:51.923 "nvme_iov_md": false 00:13:51.923 }, 00:13:51.923 "memory_domains": [ 00:13:51.923 { 00:13:51.923 "dma_device_id": "system", 00:13:51.923 "dma_device_type": 1 00:13:51.923 }, 00:13:51.923 { 00:13:51.923 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:51.923 "dma_device_type": 2 00:13:51.923 } 00:13:51.923 ], 00:13:51.923 "driver_specific": {} 00:13:51.923 } 00:13:51.923 ] 00:13:51.923 17:35:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:51.923 17:35:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:13:51.923 17:35:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:13:51.923 17:35:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:13:51.923 17:35:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 3 00:13:51.923 17:35:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:51.923 17:35:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:51.923 17:35:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:51.923 17:35:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:51.923 17:35:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:52.184 17:35:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:52.184 17:35:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:52.184 17:35:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:52.184 17:35:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:52.184 17:35:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:52.184 17:35:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:52.184 17:35:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:52.184 17:35:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:52.184 17:35:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:52.184 17:35:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:52.184 "name": "Existed_Raid", 00:13:52.184 "uuid": "4d93e1a9-bb43-40b0-8688-8e8bcbdee322", 00:13:52.184 "strip_size_kb": 64, 00:13:52.184 "state": "online", 00:13:52.184 "raid_level": "raid5f", 00:13:52.184 "superblock": false, 00:13:52.184 "num_base_bdevs": 3, 00:13:52.184 "num_base_bdevs_discovered": 3, 00:13:52.184 "num_base_bdevs_operational": 3, 00:13:52.184 "base_bdevs_list": [ 00:13:52.184 { 00:13:52.184 "name": "BaseBdev1", 00:13:52.184 "uuid": "d2979a6e-5b9e-4fca-a889-7abba68623fd", 00:13:52.184 "is_configured": true, 00:13:52.184 "data_offset": 0, 00:13:52.184 "data_size": 65536 00:13:52.184 }, 00:13:52.184 { 00:13:52.184 "name": "BaseBdev2", 00:13:52.184 "uuid": "b19c8055-6f61-42a2-8cd4-e54d87a7c6b9", 00:13:52.184 "is_configured": true, 00:13:52.184 "data_offset": 0, 00:13:52.184 "data_size": 65536 00:13:52.184 }, 00:13:52.184 { 00:13:52.184 "name": "BaseBdev3", 00:13:52.184 "uuid": "9d89e122-371a-4286-85f0-3bdd4c80f01a", 00:13:52.184 "is_configured": true, 00:13:52.184 "data_offset": 0, 00:13:52.184 "data_size": 65536 00:13:52.184 } 00:13:52.184 ] 00:13:52.184 }' 00:13:52.184 17:35:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:52.184 17:35:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:52.444 17:35:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:13:52.444 17:35:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:13:52.444 17:35:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:13:52.444 17:35:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:13:52.444 17:35:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:13:52.444 17:35:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:13:52.444 17:35:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:13:52.444 17:35:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:13:52.444 17:35:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:52.444 17:35:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:52.444 [2024-11-27 17:35:23.607984] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:13:52.444 17:35:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:52.703 17:35:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:13:52.703 "name": "Existed_Raid", 00:13:52.703 "aliases": [ 00:13:52.703 "4d93e1a9-bb43-40b0-8688-8e8bcbdee322" 00:13:52.703 ], 00:13:52.703 "product_name": "Raid Volume", 00:13:52.703 "block_size": 512, 00:13:52.703 "num_blocks": 131072, 00:13:52.703 "uuid": "4d93e1a9-bb43-40b0-8688-8e8bcbdee322", 00:13:52.703 "assigned_rate_limits": { 00:13:52.703 "rw_ios_per_sec": 0, 00:13:52.703 "rw_mbytes_per_sec": 0, 00:13:52.703 "r_mbytes_per_sec": 0, 00:13:52.703 "w_mbytes_per_sec": 0 00:13:52.703 }, 00:13:52.703 "claimed": false, 00:13:52.703 "zoned": false, 00:13:52.703 "supported_io_types": { 00:13:52.703 "read": true, 00:13:52.703 "write": true, 00:13:52.703 "unmap": false, 00:13:52.703 "flush": false, 00:13:52.703 "reset": true, 00:13:52.703 "nvme_admin": false, 00:13:52.703 "nvme_io": false, 00:13:52.703 "nvme_io_md": false, 00:13:52.703 "write_zeroes": true, 00:13:52.703 "zcopy": false, 00:13:52.703 "get_zone_info": false, 00:13:52.703 "zone_management": false, 00:13:52.703 "zone_append": false, 00:13:52.703 "compare": false, 00:13:52.703 "compare_and_write": false, 00:13:52.703 "abort": false, 00:13:52.703 "seek_hole": false, 00:13:52.703 "seek_data": false, 00:13:52.703 "copy": false, 00:13:52.703 "nvme_iov_md": false 00:13:52.703 }, 00:13:52.703 "driver_specific": { 00:13:52.703 "raid": { 00:13:52.703 "uuid": "4d93e1a9-bb43-40b0-8688-8e8bcbdee322", 00:13:52.703 "strip_size_kb": 64, 00:13:52.703 "state": "online", 00:13:52.703 "raid_level": "raid5f", 00:13:52.703 "superblock": false, 00:13:52.703 "num_base_bdevs": 3, 00:13:52.703 "num_base_bdevs_discovered": 3, 00:13:52.703 "num_base_bdevs_operational": 3, 00:13:52.703 "base_bdevs_list": [ 00:13:52.703 { 00:13:52.703 "name": "BaseBdev1", 00:13:52.703 "uuid": "d2979a6e-5b9e-4fca-a889-7abba68623fd", 00:13:52.703 "is_configured": true, 00:13:52.703 "data_offset": 0, 00:13:52.703 "data_size": 65536 00:13:52.703 }, 00:13:52.703 { 00:13:52.703 "name": "BaseBdev2", 00:13:52.703 "uuid": "b19c8055-6f61-42a2-8cd4-e54d87a7c6b9", 00:13:52.703 "is_configured": true, 00:13:52.703 "data_offset": 0, 00:13:52.703 "data_size": 65536 00:13:52.703 }, 00:13:52.703 { 00:13:52.703 "name": "BaseBdev3", 00:13:52.703 "uuid": "9d89e122-371a-4286-85f0-3bdd4c80f01a", 00:13:52.703 "is_configured": true, 00:13:52.703 "data_offset": 0, 00:13:52.703 "data_size": 65536 00:13:52.703 } 00:13:52.703 ] 00:13:52.703 } 00:13:52.703 } 00:13:52.703 }' 00:13:52.703 17:35:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:13:52.703 17:35:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:13:52.703 BaseBdev2 00:13:52.703 BaseBdev3' 00:13:52.703 17:35:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:52.703 17:35:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:13:52.703 17:35:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:52.703 17:35:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:13:52.703 17:35:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:52.703 17:35:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:52.703 17:35:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:52.703 17:35:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:52.703 17:35:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:52.703 17:35:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:52.703 17:35:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:52.703 17:35:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:13:52.703 17:35:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:52.703 17:35:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:52.703 17:35:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:52.704 17:35:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:52.704 17:35:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:52.704 17:35:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:52.704 17:35:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:52.704 17:35:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:13:52.704 17:35:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:52.704 17:35:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:52.704 17:35:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:52.704 17:35:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:52.963 17:35:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:52.963 17:35:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:52.963 17:35:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:13:52.963 17:35:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:52.963 17:35:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:52.963 [2024-11-27 17:35:23.903354] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:13:52.963 17:35:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:52.963 17:35:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:13:52.963 17:35:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy raid5f 00:13:52.963 17:35:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:13:52.963 17:35:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@199 -- # return 0 00:13:52.963 17:35:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:13:52.963 17:35:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 2 00:13:52.963 17:35:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:52.963 17:35:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:52.963 17:35:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:52.963 17:35:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:52.963 17:35:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:13:52.963 17:35:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:52.963 17:35:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:52.963 17:35:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:52.963 17:35:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:52.963 17:35:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:52.963 17:35:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:52.963 17:35:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:52.963 17:35:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:52.963 17:35:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:52.964 17:35:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:52.964 "name": "Existed_Raid", 00:13:52.964 "uuid": "4d93e1a9-bb43-40b0-8688-8e8bcbdee322", 00:13:52.964 "strip_size_kb": 64, 00:13:52.964 "state": "online", 00:13:52.964 "raid_level": "raid5f", 00:13:52.964 "superblock": false, 00:13:52.964 "num_base_bdevs": 3, 00:13:52.964 "num_base_bdevs_discovered": 2, 00:13:52.964 "num_base_bdevs_operational": 2, 00:13:52.964 "base_bdevs_list": [ 00:13:52.964 { 00:13:52.964 "name": null, 00:13:52.964 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:52.964 "is_configured": false, 00:13:52.964 "data_offset": 0, 00:13:52.964 "data_size": 65536 00:13:52.964 }, 00:13:52.964 { 00:13:52.964 "name": "BaseBdev2", 00:13:52.964 "uuid": "b19c8055-6f61-42a2-8cd4-e54d87a7c6b9", 00:13:52.964 "is_configured": true, 00:13:52.964 "data_offset": 0, 00:13:52.964 "data_size": 65536 00:13:52.964 }, 00:13:52.964 { 00:13:52.964 "name": "BaseBdev3", 00:13:52.964 "uuid": "9d89e122-371a-4286-85f0-3bdd4c80f01a", 00:13:52.964 "is_configured": true, 00:13:52.964 "data_offset": 0, 00:13:52.964 "data_size": 65536 00:13:52.964 } 00:13:52.964 ] 00:13:52.964 }' 00:13:52.964 17:35:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:52.964 17:35:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:53.223 17:35:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:13:53.223 17:35:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:13:53.223 17:35:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:53.223 17:35:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:53.223 17:35:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:53.223 17:35:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:13:53.223 17:35:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:53.483 17:35:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:13:53.483 17:35:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:13:53.483 17:35:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:13:53.483 17:35:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:53.483 17:35:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:53.483 [2024-11-27 17:35:24.421509] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:13:53.483 [2024-11-27 17:35:24.421642] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:13:53.483 [2024-11-27 17:35:24.432717] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:13:53.483 17:35:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:53.483 17:35:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:13:53.483 17:35:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:13:53.483 17:35:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:53.483 17:35:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:13:53.483 17:35:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:53.483 17:35:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:53.483 17:35:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:53.483 17:35:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:13:53.483 17:35:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:13:53.483 17:35:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:13:53.483 17:35:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:53.483 17:35:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:53.483 [2024-11-27 17:35:24.492636] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:13:53.483 [2024-11-27 17:35:24.492674] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:13:53.483 17:35:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:53.483 17:35:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:13:53.483 17:35:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:13:53.483 17:35:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:53.484 17:35:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:13:53.484 17:35:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:53.484 17:35:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:53.484 17:35:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:53.484 17:35:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:13:53.484 17:35:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:13:53.484 17:35:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 3 -gt 2 ']' 00:13:53.484 17:35:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:13:53.484 17:35:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:13:53.484 17:35:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:13:53.484 17:35:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:53.484 17:35:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:53.484 BaseBdev2 00:13:53.484 17:35:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:53.484 17:35:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:13:53.484 17:35:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:13:53.484 17:35:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:13:53.484 17:35:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@901 -- # local i 00:13:53.484 17:35:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:13:53.484 17:35:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:13:53.484 17:35:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:13:53.484 17:35:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:53.484 17:35:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:53.484 17:35:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:53.484 17:35:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:13:53.484 17:35:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:53.484 17:35:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:53.484 [ 00:13:53.484 { 00:13:53.484 "name": "BaseBdev2", 00:13:53.484 "aliases": [ 00:13:53.484 "19cd3e7f-392d-44b7-a277-cfaf16209363" 00:13:53.484 ], 00:13:53.484 "product_name": "Malloc disk", 00:13:53.484 "block_size": 512, 00:13:53.484 "num_blocks": 65536, 00:13:53.484 "uuid": "19cd3e7f-392d-44b7-a277-cfaf16209363", 00:13:53.484 "assigned_rate_limits": { 00:13:53.484 "rw_ios_per_sec": 0, 00:13:53.484 "rw_mbytes_per_sec": 0, 00:13:53.484 "r_mbytes_per_sec": 0, 00:13:53.484 "w_mbytes_per_sec": 0 00:13:53.484 }, 00:13:53.484 "claimed": false, 00:13:53.484 "zoned": false, 00:13:53.484 "supported_io_types": { 00:13:53.484 "read": true, 00:13:53.484 "write": true, 00:13:53.484 "unmap": true, 00:13:53.484 "flush": true, 00:13:53.484 "reset": true, 00:13:53.484 "nvme_admin": false, 00:13:53.484 "nvme_io": false, 00:13:53.484 "nvme_io_md": false, 00:13:53.484 "write_zeroes": true, 00:13:53.484 "zcopy": true, 00:13:53.484 "get_zone_info": false, 00:13:53.484 "zone_management": false, 00:13:53.484 "zone_append": false, 00:13:53.484 "compare": false, 00:13:53.484 "compare_and_write": false, 00:13:53.484 "abort": true, 00:13:53.484 "seek_hole": false, 00:13:53.484 "seek_data": false, 00:13:53.484 "copy": true, 00:13:53.484 "nvme_iov_md": false 00:13:53.484 }, 00:13:53.484 "memory_domains": [ 00:13:53.484 { 00:13:53.484 "dma_device_id": "system", 00:13:53.484 "dma_device_type": 1 00:13:53.484 }, 00:13:53.484 { 00:13:53.484 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:53.484 "dma_device_type": 2 00:13:53.484 } 00:13:53.484 ], 00:13:53.484 "driver_specific": {} 00:13:53.484 } 00:13:53.484 ] 00:13:53.484 17:35:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:53.484 17:35:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:13:53.484 17:35:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:13:53.484 17:35:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:13:53.484 17:35:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:13:53.484 17:35:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:53.484 17:35:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:53.484 BaseBdev3 00:13:53.484 17:35:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:53.484 17:35:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:13:53.484 17:35:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:13:53.484 17:35:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:13:53.484 17:35:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@901 -- # local i 00:13:53.484 17:35:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:13:53.484 17:35:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:13:53.484 17:35:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:13:53.484 17:35:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:53.484 17:35:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:53.484 17:35:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:53.484 17:35:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:13:53.484 17:35:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:53.484 17:35:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:53.484 [ 00:13:53.484 { 00:13:53.484 "name": "BaseBdev3", 00:13:53.484 "aliases": [ 00:13:53.484 "9e6fe33d-08d3-43ff-9d96-167c0268396d" 00:13:53.484 ], 00:13:53.484 "product_name": "Malloc disk", 00:13:53.484 "block_size": 512, 00:13:53.484 "num_blocks": 65536, 00:13:53.484 "uuid": "9e6fe33d-08d3-43ff-9d96-167c0268396d", 00:13:53.484 "assigned_rate_limits": { 00:13:53.484 "rw_ios_per_sec": 0, 00:13:53.484 "rw_mbytes_per_sec": 0, 00:13:53.484 "r_mbytes_per_sec": 0, 00:13:53.484 "w_mbytes_per_sec": 0 00:13:53.484 }, 00:13:53.484 "claimed": false, 00:13:53.484 "zoned": false, 00:13:53.484 "supported_io_types": { 00:13:53.484 "read": true, 00:13:53.484 "write": true, 00:13:53.484 "unmap": true, 00:13:53.484 "flush": true, 00:13:53.484 "reset": true, 00:13:53.484 "nvme_admin": false, 00:13:53.484 "nvme_io": false, 00:13:53.484 "nvme_io_md": false, 00:13:53.484 "write_zeroes": true, 00:13:53.484 "zcopy": true, 00:13:53.484 "get_zone_info": false, 00:13:53.484 "zone_management": false, 00:13:53.484 "zone_append": false, 00:13:53.484 "compare": false, 00:13:53.484 "compare_and_write": false, 00:13:53.484 "abort": true, 00:13:53.484 "seek_hole": false, 00:13:53.484 "seek_data": false, 00:13:53.484 "copy": true, 00:13:53.484 "nvme_iov_md": false 00:13:53.484 }, 00:13:53.484 "memory_domains": [ 00:13:53.484 { 00:13:53.484 "dma_device_id": "system", 00:13:53.484 "dma_device_type": 1 00:13:53.484 }, 00:13:53.484 { 00:13:53.484 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:53.484 "dma_device_type": 2 00:13:53.484 } 00:13:53.485 ], 00:13:53.485 "driver_specific": {} 00:13:53.485 } 00:13:53.485 ] 00:13:53.485 17:35:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:53.485 17:35:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:13:53.485 17:35:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:13:53.485 17:35:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:13:53.485 17:35:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:13:53.485 17:35:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:53.485 17:35:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:53.485 [2024-11-27 17:35:24.666812] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:13:53.485 [2024-11-27 17:35:24.666903] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:13:53.485 [2024-11-27 17:35:24.666942] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:13:53.485 [2024-11-27 17:35:24.668617] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:13:53.485 17:35:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:53.485 17:35:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:53.485 17:35:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:53.745 17:35:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:53.745 17:35:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:53.745 17:35:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:53.745 17:35:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:53.745 17:35:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:53.745 17:35:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:53.745 17:35:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:53.745 17:35:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:53.745 17:35:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:53.745 17:35:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:53.745 17:35:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:53.745 17:35:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:53.745 17:35:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:53.745 17:35:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:53.745 "name": "Existed_Raid", 00:13:53.745 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:53.745 "strip_size_kb": 64, 00:13:53.745 "state": "configuring", 00:13:53.745 "raid_level": "raid5f", 00:13:53.745 "superblock": false, 00:13:53.745 "num_base_bdevs": 3, 00:13:53.745 "num_base_bdevs_discovered": 2, 00:13:53.745 "num_base_bdevs_operational": 3, 00:13:53.745 "base_bdevs_list": [ 00:13:53.745 { 00:13:53.745 "name": "BaseBdev1", 00:13:53.745 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:53.745 "is_configured": false, 00:13:53.745 "data_offset": 0, 00:13:53.745 "data_size": 0 00:13:53.745 }, 00:13:53.745 { 00:13:53.745 "name": "BaseBdev2", 00:13:53.745 "uuid": "19cd3e7f-392d-44b7-a277-cfaf16209363", 00:13:53.745 "is_configured": true, 00:13:53.745 "data_offset": 0, 00:13:53.745 "data_size": 65536 00:13:53.745 }, 00:13:53.745 { 00:13:53.745 "name": "BaseBdev3", 00:13:53.745 "uuid": "9e6fe33d-08d3-43ff-9d96-167c0268396d", 00:13:53.745 "is_configured": true, 00:13:53.745 "data_offset": 0, 00:13:53.746 "data_size": 65536 00:13:53.746 } 00:13:53.746 ] 00:13:53.746 }' 00:13:53.746 17:35:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:53.746 17:35:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:54.005 17:35:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:13:54.006 17:35:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:54.006 17:35:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:54.006 [2024-11-27 17:35:25.122001] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:13:54.006 17:35:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:54.006 17:35:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:54.006 17:35:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:54.006 17:35:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:54.006 17:35:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:54.006 17:35:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:54.006 17:35:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:54.006 17:35:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:54.006 17:35:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:54.006 17:35:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:54.006 17:35:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:54.006 17:35:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:54.006 17:35:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:54.006 17:35:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:54.006 17:35:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:54.006 17:35:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:54.006 17:35:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:54.006 "name": "Existed_Raid", 00:13:54.006 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:54.006 "strip_size_kb": 64, 00:13:54.006 "state": "configuring", 00:13:54.006 "raid_level": "raid5f", 00:13:54.006 "superblock": false, 00:13:54.006 "num_base_bdevs": 3, 00:13:54.006 "num_base_bdevs_discovered": 1, 00:13:54.006 "num_base_bdevs_operational": 3, 00:13:54.006 "base_bdevs_list": [ 00:13:54.006 { 00:13:54.006 "name": "BaseBdev1", 00:13:54.006 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:54.006 "is_configured": false, 00:13:54.006 "data_offset": 0, 00:13:54.006 "data_size": 0 00:13:54.006 }, 00:13:54.006 { 00:13:54.006 "name": null, 00:13:54.006 "uuid": "19cd3e7f-392d-44b7-a277-cfaf16209363", 00:13:54.006 "is_configured": false, 00:13:54.006 "data_offset": 0, 00:13:54.006 "data_size": 65536 00:13:54.006 }, 00:13:54.006 { 00:13:54.006 "name": "BaseBdev3", 00:13:54.006 "uuid": "9e6fe33d-08d3-43ff-9d96-167c0268396d", 00:13:54.006 "is_configured": true, 00:13:54.006 "data_offset": 0, 00:13:54.006 "data_size": 65536 00:13:54.006 } 00:13:54.006 ] 00:13:54.006 }' 00:13:54.006 17:35:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:54.006 17:35:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:54.575 17:35:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:54.575 17:35:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:13:54.575 17:35:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:54.575 17:35:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:54.575 17:35:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:54.575 17:35:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:13:54.575 17:35:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:13:54.575 17:35:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:54.575 17:35:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:54.575 [2024-11-27 17:35:25.664125] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:13:54.575 BaseBdev1 00:13:54.575 17:35:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:54.575 17:35:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:13:54.575 17:35:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:13:54.575 17:35:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:13:54.575 17:35:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@901 -- # local i 00:13:54.575 17:35:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:13:54.575 17:35:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:13:54.575 17:35:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:13:54.575 17:35:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:54.575 17:35:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:54.575 17:35:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:54.575 17:35:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:13:54.575 17:35:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:54.575 17:35:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:54.575 [ 00:13:54.575 { 00:13:54.575 "name": "BaseBdev1", 00:13:54.575 "aliases": [ 00:13:54.575 "2c4ffa6b-40dd-4889-915e-6ef9b20a04e0" 00:13:54.575 ], 00:13:54.575 "product_name": "Malloc disk", 00:13:54.575 "block_size": 512, 00:13:54.575 "num_blocks": 65536, 00:13:54.575 "uuid": "2c4ffa6b-40dd-4889-915e-6ef9b20a04e0", 00:13:54.575 "assigned_rate_limits": { 00:13:54.575 "rw_ios_per_sec": 0, 00:13:54.575 "rw_mbytes_per_sec": 0, 00:13:54.575 "r_mbytes_per_sec": 0, 00:13:54.575 "w_mbytes_per_sec": 0 00:13:54.575 }, 00:13:54.575 "claimed": true, 00:13:54.575 "claim_type": "exclusive_write", 00:13:54.575 "zoned": false, 00:13:54.575 "supported_io_types": { 00:13:54.575 "read": true, 00:13:54.575 "write": true, 00:13:54.575 "unmap": true, 00:13:54.575 "flush": true, 00:13:54.575 "reset": true, 00:13:54.575 "nvme_admin": false, 00:13:54.575 "nvme_io": false, 00:13:54.575 "nvme_io_md": false, 00:13:54.575 "write_zeroes": true, 00:13:54.575 "zcopy": true, 00:13:54.575 "get_zone_info": false, 00:13:54.575 "zone_management": false, 00:13:54.575 "zone_append": false, 00:13:54.575 "compare": false, 00:13:54.575 "compare_and_write": false, 00:13:54.575 "abort": true, 00:13:54.575 "seek_hole": false, 00:13:54.575 "seek_data": false, 00:13:54.575 "copy": true, 00:13:54.575 "nvme_iov_md": false 00:13:54.575 }, 00:13:54.575 "memory_domains": [ 00:13:54.575 { 00:13:54.576 "dma_device_id": "system", 00:13:54.576 "dma_device_type": 1 00:13:54.576 }, 00:13:54.576 { 00:13:54.576 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:54.576 "dma_device_type": 2 00:13:54.576 } 00:13:54.576 ], 00:13:54.576 "driver_specific": {} 00:13:54.576 } 00:13:54.576 ] 00:13:54.576 17:35:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:54.576 17:35:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:13:54.576 17:35:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:54.576 17:35:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:54.576 17:35:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:54.576 17:35:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:54.576 17:35:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:54.576 17:35:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:54.576 17:35:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:54.576 17:35:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:54.576 17:35:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:54.576 17:35:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:54.576 17:35:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:54.576 17:35:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:54.576 17:35:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:54.576 17:35:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:54.576 17:35:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:54.576 17:35:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:54.576 "name": "Existed_Raid", 00:13:54.576 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:54.576 "strip_size_kb": 64, 00:13:54.576 "state": "configuring", 00:13:54.576 "raid_level": "raid5f", 00:13:54.576 "superblock": false, 00:13:54.576 "num_base_bdevs": 3, 00:13:54.576 "num_base_bdevs_discovered": 2, 00:13:54.576 "num_base_bdevs_operational": 3, 00:13:54.576 "base_bdevs_list": [ 00:13:54.576 { 00:13:54.576 "name": "BaseBdev1", 00:13:54.576 "uuid": "2c4ffa6b-40dd-4889-915e-6ef9b20a04e0", 00:13:54.576 "is_configured": true, 00:13:54.576 "data_offset": 0, 00:13:54.576 "data_size": 65536 00:13:54.576 }, 00:13:54.576 { 00:13:54.576 "name": null, 00:13:54.576 "uuid": "19cd3e7f-392d-44b7-a277-cfaf16209363", 00:13:54.576 "is_configured": false, 00:13:54.576 "data_offset": 0, 00:13:54.576 "data_size": 65536 00:13:54.576 }, 00:13:54.576 { 00:13:54.576 "name": "BaseBdev3", 00:13:54.576 "uuid": "9e6fe33d-08d3-43ff-9d96-167c0268396d", 00:13:54.576 "is_configured": true, 00:13:54.576 "data_offset": 0, 00:13:54.576 "data_size": 65536 00:13:54.576 } 00:13:54.576 ] 00:13:54.576 }' 00:13:54.576 17:35:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:54.576 17:35:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:55.145 17:35:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:13:55.145 17:35:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:55.145 17:35:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:55.145 17:35:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:55.145 17:35:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:55.145 17:35:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:13:55.145 17:35:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:13:55.145 17:35:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:55.146 17:35:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:55.146 [2024-11-27 17:35:26.191234] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:13:55.146 17:35:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:55.146 17:35:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:55.146 17:35:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:55.146 17:35:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:55.146 17:35:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:55.146 17:35:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:55.146 17:35:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:55.146 17:35:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:55.146 17:35:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:55.146 17:35:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:55.146 17:35:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:55.146 17:35:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:55.146 17:35:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:55.146 17:35:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:55.146 17:35:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:55.146 17:35:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:55.146 17:35:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:55.146 "name": "Existed_Raid", 00:13:55.146 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:55.146 "strip_size_kb": 64, 00:13:55.146 "state": "configuring", 00:13:55.146 "raid_level": "raid5f", 00:13:55.146 "superblock": false, 00:13:55.146 "num_base_bdevs": 3, 00:13:55.146 "num_base_bdevs_discovered": 1, 00:13:55.146 "num_base_bdevs_operational": 3, 00:13:55.146 "base_bdevs_list": [ 00:13:55.146 { 00:13:55.146 "name": "BaseBdev1", 00:13:55.146 "uuid": "2c4ffa6b-40dd-4889-915e-6ef9b20a04e0", 00:13:55.146 "is_configured": true, 00:13:55.146 "data_offset": 0, 00:13:55.146 "data_size": 65536 00:13:55.146 }, 00:13:55.146 { 00:13:55.146 "name": null, 00:13:55.146 "uuid": "19cd3e7f-392d-44b7-a277-cfaf16209363", 00:13:55.146 "is_configured": false, 00:13:55.146 "data_offset": 0, 00:13:55.146 "data_size": 65536 00:13:55.146 }, 00:13:55.146 { 00:13:55.146 "name": null, 00:13:55.146 "uuid": "9e6fe33d-08d3-43ff-9d96-167c0268396d", 00:13:55.146 "is_configured": false, 00:13:55.146 "data_offset": 0, 00:13:55.146 "data_size": 65536 00:13:55.146 } 00:13:55.146 ] 00:13:55.146 }' 00:13:55.146 17:35:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:55.146 17:35:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:55.721 17:35:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:55.721 17:35:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:13:55.721 17:35:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:55.721 17:35:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:55.721 17:35:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:55.721 17:35:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:13:55.721 17:35:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:13:55.721 17:35:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:55.721 17:35:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:55.721 [2024-11-27 17:35:26.710500] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:13:55.721 17:35:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:55.721 17:35:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:55.721 17:35:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:55.721 17:35:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:55.721 17:35:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:55.721 17:35:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:55.721 17:35:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:55.721 17:35:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:55.721 17:35:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:55.721 17:35:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:55.721 17:35:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:55.721 17:35:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:55.721 17:35:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:55.721 17:35:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:55.721 17:35:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:55.721 17:35:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:55.721 17:35:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:55.721 "name": "Existed_Raid", 00:13:55.721 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:55.721 "strip_size_kb": 64, 00:13:55.721 "state": "configuring", 00:13:55.721 "raid_level": "raid5f", 00:13:55.721 "superblock": false, 00:13:55.721 "num_base_bdevs": 3, 00:13:55.721 "num_base_bdevs_discovered": 2, 00:13:55.721 "num_base_bdevs_operational": 3, 00:13:55.721 "base_bdevs_list": [ 00:13:55.721 { 00:13:55.721 "name": "BaseBdev1", 00:13:55.721 "uuid": "2c4ffa6b-40dd-4889-915e-6ef9b20a04e0", 00:13:55.721 "is_configured": true, 00:13:55.721 "data_offset": 0, 00:13:55.721 "data_size": 65536 00:13:55.721 }, 00:13:55.721 { 00:13:55.721 "name": null, 00:13:55.721 "uuid": "19cd3e7f-392d-44b7-a277-cfaf16209363", 00:13:55.721 "is_configured": false, 00:13:55.721 "data_offset": 0, 00:13:55.721 "data_size": 65536 00:13:55.721 }, 00:13:55.721 { 00:13:55.721 "name": "BaseBdev3", 00:13:55.721 "uuid": "9e6fe33d-08d3-43ff-9d96-167c0268396d", 00:13:55.721 "is_configured": true, 00:13:55.721 "data_offset": 0, 00:13:55.721 "data_size": 65536 00:13:55.721 } 00:13:55.721 ] 00:13:55.721 }' 00:13:55.721 17:35:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:55.721 17:35:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:55.981 17:35:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:55.981 17:35:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:13:55.981 17:35:27 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:55.981 17:35:27 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:55.981 17:35:27 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:56.242 17:35:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:13:56.242 17:35:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:13:56.242 17:35:27 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:56.242 17:35:27 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:56.242 [2024-11-27 17:35:27.205730] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:13:56.242 17:35:27 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:56.242 17:35:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:56.242 17:35:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:56.242 17:35:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:56.242 17:35:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:56.242 17:35:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:56.242 17:35:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:56.242 17:35:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:56.242 17:35:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:56.242 17:35:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:56.242 17:35:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:56.242 17:35:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:56.242 17:35:27 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:56.242 17:35:27 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:56.242 17:35:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:56.242 17:35:27 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:56.242 17:35:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:56.242 "name": "Existed_Raid", 00:13:56.242 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:56.242 "strip_size_kb": 64, 00:13:56.242 "state": "configuring", 00:13:56.242 "raid_level": "raid5f", 00:13:56.242 "superblock": false, 00:13:56.242 "num_base_bdevs": 3, 00:13:56.242 "num_base_bdevs_discovered": 1, 00:13:56.242 "num_base_bdevs_operational": 3, 00:13:56.242 "base_bdevs_list": [ 00:13:56.242 { 00:13:56.242 "name": null, 00:13:56.242 "uuid": "2c4ffa6b-40dd-4889-915e-6ef9b20a04e0", 00:13:56.242 "is_configured": false, 00:13:56.242 "data_offset": 0, 00:13:56.242 "data_size": 65536 00:13:56.242 }, 00:13:56.242 { 00:13:56.242 "name": null, 00:13:56.242 "uuid": "19cd3e7f-392d-44b7-a277-cfaf16209363", 00:13:56.242 "is_configured": false, 00:13:56.242 "data_offset": 0, 00:13:56.242 "data_size": 65536 00:13:56.242 }, 00:13:56.242 { 00:13:56.242 "name": "BaseBdev3", 00:13:56.242 "uuid": "9e6fe33d-08d3-43ff-9d96-167c0268396d", 00:13:56.242 "is_configured": true, 00:13:56.242 "data_offset": 0, 00:13:56.242 "data_size": 65536 00:13:56.242 } 00:13:56.242 ] 00:13:56.242 }' 00:13:56.242 17:35:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:56.242 17:35:27 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:56.501 17:35:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:56.501 17:35:27 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:56.501 17:35:27 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:56.762 17:35:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:13:56.762 17:35:27 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:56.762 17:35:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:13:56.762 17:35:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:13:56.762 17:35:27 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:56.762 17:35:27 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:56.762 [2024-11-27 17:35:27.743019] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:13:56.762 17:35:27 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:56.762 17:35:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:56.762 17:35:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:56.762 17:35:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:56.762 17:35:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:56.762 17:35:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:56.762 17:35:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:56.762 17:35:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:56.762 17:35:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:56.762 17:35:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:56.762 17:35:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:56.762 17:35:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:56.762 17:35:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:56.762 17:35:27 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:56.762 17:35:27 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:56.762 17:35:27 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:56.762 17:35:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:56.762 "name": "Existed_Raid", 00:13:56.762 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:56.762 "strip_size_kb": 64, 00:13:56.762 "state": "configuring", 00:13:56.762 "raid_level": "raid5f", 00:13:56.762 "superblock": false, 00:13:56.762 "num_base_bdevs": 3, 00:13:56.762 "num_base_bdevs_discovered": 2, 00:13:56.762 "num_base_bdevs_operational": 3, 00:13:56.762 "base_bdevs_list": [ 00:13:56.762 { 00:13:56.762 "name": null, 00:13:56.762 "uuid": "2c4ffa6b-40dd-4889-915e-6ef9b20a04e0", 00:13:56.762 "is_configured": false, 00:13:56.762 "data_offset": 0, 00:13:56.762 "data_size": 65536 00:13:56.762 }, 00:13:56.762 { 00:13:56.762 "name": "BaseBdev2", 00:13:56.762 "uuid": "19cd3e7f-392d-44b7-a277-cfaf16209363", 00:13:56.762 "is_configured": true, 00:13:56.762 "data_offset": 0, 00:13:56.762 "data_size": 65536 00:13:56.762 }, 00:13:56.762 { 00:13:56.762 "name": "BaseBdev3", 00:13:56.762 "uuid": "9e6fe33d-08d3-43ff-9d96-167c0268396d", 00:13:56.762 "is_configured": true, 00:13:56.762 "data_offset": 0, 00:13:56.762 "data_size": 65536 00:13:56.762 } 00:13:56.762 ] 00:13:56.762 }' 00:13:56.762 17:35:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:56.762 17:35:27 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:57.333 17:35:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:57.333 17:35:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:13:57.333 17:35:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:57.333 17:35:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:57.333 17:35:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:57.333 17:35:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:13:57.333 17:35:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:57.333 17:35:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:13:57.333 17:35:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:57.333 17:35:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:57.333 17:35:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:57.333 17:35:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 2c4ffa6b-40dd-4889-915e-6ef9b20a04e0 00:13:57.333 17:35:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:57.333 17:35:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:57.333 [2024-11-27 17:35:28.312973] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:13:57.333 [2024-11-27 17:35:28.313012] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:13:57.333 [2024-11-27 17:35:28.313020] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 131072, blocklen 512 00:13:57.333 [2024-11-27 17:35:28.313248] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002870 00:13:57.333 [2024-11-27 17:35:28.313603] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:13:57.333 [2024-11-27 17:35:28.313628] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001c80 00:13:57.333 [2024-11-27 17:35:28.313794] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:13:57.333 NewBaseBdev 00:13:57.333 17:35:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:57.333 17:35:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:13:57.333 17:35:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=NewBaseBdev 00:13:57.333 17:35:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:13:57.333 17:35:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@901 -- # local i 00:13:57.333 17:35:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:13:57.333 17:35:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:13:57.333 17:35:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:13:57.333 17:35:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:57.333 17:35:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:57.333 17:35:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:57.333 17:35:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:13:57.333 17:35:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:57.333 17:35:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:57.333 [ 00:13:57.333 { 00:13:57.333 "name": "NewBaseBdev", 00:13:57.333 "aliases": [ 00:13:57.333 "2c4ffa6b-40dd-4889-915e-6ef9b20a04e0" 00:13:57.333 ], 00:13:57.333 "product_name": "Malloc disk", 00:13:57.333 "block_size": 512, 00:13:57.333 "num_blocks": 65536, 00:13:57.333 "uuid": "2c4ffa6b-40dd-4889-915e-6ef9b20a04e0", 00:13:57.333 "assigned_rate_limits": { 00:13:57.333 "rw_ios_per_sec": 0, 00:13:57.333 "rw_mbytes_per_sec": 0, 00:13:57.333 "r_mbytes_per_sec": 0, 00:13:57.333 "w_mbytes_per_sec": 0 00:13:57.333 }, 00:13:57.333 "claimed": true, 00:13:57.333 "claim_type": "exclusive_write", 00:13:57.333 "zoned": false, 00:13:57.333 "supported_io_types": { 00:13:57.333 "read": true, 00:13:57.333 "write": true, 00:13:57.333 "unmap": true, 00:13:57.333 "flush": true, 00:13:57.333 "reset": true, 00:13:57.333 "nvme_admin": false, 00:13:57.333 "nvme_io": false, 00:13:57.333 "nvme_io_md": false, 00:13:57.333 "write_zeroes": true, 00:13:57.333 "zcopy": true, 00:13:57.333 "get_zone_info": false, 00:13:57.333 "zone_management": false, 00:13:57.333 "zone_append": false, 00:13:57.333 "compare": false, 00:13:57.333 "compare_and_write": false, 00:13:57.333 "abort": true, 00:13:57.333 "seek_hole": false, 00:13:57.333 "seek_data": false, 00:13:57.333 "copy": true, 00:13:57.333 "nvme_iov_md": false 00:13:57.333 }, 00:13:57.333 "memory_domains": [ 00:13:57.333 { 00:13:57.333 "dma_device_id": "system", 00:13:57.333 "dma_device_type": 1 00:13:57.333 }, 00:13:57.333 { 00:13:57.333 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:57.333 "dma_device_type": 2 00:13:57.333 } 00:13:57.333 ], 00:13:57.333 "driver_specific": {} 00:13:57.333 } 00:13:57.333 ] 00:13:57.333 17:35:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:57.333 17:35:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:13:57.333 17:35:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 3 00:13:57.333 17:35:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:57.333 17:35:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:13:57.333 17:35:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:57.333 17:35:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:57.333 17:35:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:57.333 17:35:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:57.334 17:35:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:57.334 17:35:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:57.334 17:35:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:57.334 17:35:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:57.334 17:35:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:57.334 17:35:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:57.334 17:35:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:57.334 17:35:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:57.334 17:35:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:57.334 "name": "Existed_Raid", 00:13:57.334 "uuid": "f380a9c8-5cea-44ae-b1b7-c10592b2561a", 00:13:57.334 "strip_size_kb": 64, 00:13:57.334 "state": "online", 00:13:57.334 "raid_level": "raid5f", 00:13:57.334 "superblock": false, 00:13:57.334 "num_base_bdevs": 3, 00:13:57.334 "num_base_bdevs_discovered": 3, 00:13:57.334 "num_base_bdevs_operational": 3, 00:13:57.334 "base_bdevs_list": [ 00:13:57.334 { 00:13:57.334 "name": "NewBaseBdev", 00:13:57.334 "uuid": "2c4ffa6b-40dd-4889-915e-6ef9b20a04e0", 00:13:57.334 "is_configured": true, 00:13:57.334 "data_offset": 0, 00:13:57.334 "data_size": 65536 00:13:57.334 }, 00:13:57.334 { 00:13:57.334 "name": "BaseBdev2", 00:13:57.334 "uuid": "19cd3e7f-392d-44b7-a277-cfaf16209363", 00:13:57.334 "is_configured": true, 00:13:57.334 "data_offset": 0, 00:13:57.334 "data_size": 65536 00:13:57.334 }, 00:13:57.334 { 00:13:57.334 "name": "BaseBdev3", 00:13:57.334 "uuid": "9e6fe33d-08d3-43ff-9d96-167c0268396d", 00:13:57.334 "is_configured": true, 00:13:57.334 "data_offset": 0, 00:13:57.334 "data_size": 65536 00:13:57.334 } 00:13:57.334 ] 00:13:57.334 }' 00:13:57.334 17:35:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:57.334 17:35:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:57.594 17:35:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:13:57.594 17:35:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:13:57.594 17:35:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:13:57.594 17:35:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:13:57.594 17:35:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:13:57.594 17:35:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:13:57.594 17:35:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:13:57.594 17:35:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:13:57.594 17:35:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:57.594 17:35:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:57.594 [2024-11-27 17:35:28.760407] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:13:57.594 17:35:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:57.855 17:35:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:13:57.855 "name": "Existed_Raid", 00:13:57.855 "aliases": [ 00:13:57.855 "f380a9c8-5cea-44ae-b1b7-c10592b2561a" 00:13:57.855 ], 00:13:57.855 "product_name": "Raid Volume", 00:13:57.855 "block_size": 512, 00:13:57.855 "num_blocks": 131072, 00:13:57.855 "uuid": "f380a9c8-5cea-44ae-b1b7-c10592b2561a", 00:13:57.855 "assigned_rate_limits": { 00:13:57.855 "rw_ios_per_sec": 0, 00:13:57.855 "rw_mbytes_per_sec": 0, 00:13:57.855 "r_mbytes_per_sec": 0, 00:13:57.855 "w_mbytes_per_sec": 0 00:13:57.855 }, 00:13:57.855 "claimed": false, 00:13:57.855 "zoned": false, 00:13:57.855 "supported_io_types": { 00:13:57.855 "read": true, 00:13:57.855 "write": true, 00:13:57.855 "unmap": false, 00:13:57.855 "flush": false, 00:13:57.855 "reset": true, 00:13:57.855 "nvme_admin": false, 00:13:57.855 "nvme_io": false, 00:13:57.855 "nvme_io_md": false, 00:13:57.855 "write_zeroes": true, 00:13:57.855 "zcopy": false, 00:13:57.855 "get_zone_info": false, 00:13:57.855 "zone_management": false, 00:13:57.855 "zone_append": false, 00:13:57.855 "compare": false, 00:13:57.855 "compare_and_write": false, 00:13:57.855 "abort": false, 00:13:57.855 "seek_hole": false, 00:13:57.855 "seek_data": false, 00:13:57.855 "copy": false, 00:13:57.855 "nvme_iov_md": false 00:13:57.855 }, 00:13:57.855 "driver_specific": { 00:13:57.855 "raid": { 00:13:57.855 "uuid": "f380a9c8-5cea-44ae-b1b7-c10592b2561a", 00:13:57.855 "strip_size_kb": 64, 00:13:57.855 "state": "online", 00:13:57.855 "raid_level": "raid5f", 00:13:57.855 "superblock": false, 00:13:57.855 "num_base_bdevs": 3, 00:13:57.855 "num_base_bdevs_discovered": 3, 00:13:57.855 "num_base_bdevs_operational": 3, 00:13:57.855 "base_bdevs_list": [ 00:13:57.855 { 00:13:57.855 "name": "NewBaseBdev", 00:13:57.855 "uuid": "2c4ffa6b-40dd-4889-915e-6ef9b20a04e0", 00:13:57.855 "is_configured": true, 00:13:57.855 "data_offset": 0, 00:13:57.855 "data_size": 65536 00:13:57.855 }, 00:13:57.855 { 00:13:57.855 "name": "BaseBdev2", 00:13:57.855 "uuid": "19cd3e7f-392d-44b7-a277-cfaf16209363", 00:13:57.855 "is_configured": true, 00:13:57.855 "data_offset": 0, 00:13:57.855 "data_size": 65536 00:13:57.855 }, 00:13:57.855 { 00:13:57.855 "name": "BaseBdev3", 00:13:57.855 "uuid": "9e6fe33d-08d3-43ff-9d96-167c0268396d", 00:13:57.855 "is_configured": true, 00:13:57.855 "data_offset": 0, 00:13:57.855 "data_size": 65536 00:13:57.855 } 00:13:57.855 ] 00:13:57.855 } 00:13:57.855 } 00:13:57.855 }' 00:13:57.855 17:35:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:13:57.855 17:35:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:13:57.855 BaseBdev2 00:13:57.855 BaseBdev3' 00:13:57.855 17:35:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:57.855 17:35:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:13:57.855 17:35:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:57.855 17:35:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:13:57.855 17:35:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:57.855 17:35:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:57.855 17:35:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:57.855 17:35:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:57.855 17:35:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:57.855 17:35:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:57.855 17:35:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:57.855 17:35:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:13:57.855 17:35:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:57.855 17:35:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:57.855 17:35:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:57.855 17:35:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:57.855 17:35:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:57.855 17:35:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:57.855 17:35:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:13:57.855 17:35:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:13:57.855 17:35:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:57.855 17:35:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:57.855 17:35:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:13:57.855 17:35:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:57.855 17:35:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:13:57.855 17:35:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:13:57.855 17:35:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:13:57.855 17:35:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:57.855 17:35:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:57.855 [2024-11-27 17:35:29.023827] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:13:57.855 [2024-11-27 17:35:29.023848] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:13:57.855 [2024-11-27 17:35:29.023905] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:13:57.855 [2024-11-27 17:35:29.024123] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:13:57.855 [2024-11-27 17:35:29.024135] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name Existed_Raid, state offline 00:13:57.855 17:35:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:57.855 17:35:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 90305 00:13:57.855 17:35:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@950 -- # '[' -z 90305 ']' 00:13:57.855 17:35:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@954 -- # kill -0 90305 00:13:57.855 17:35:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@955 -- # uname 00:13:57.855 17:35:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:13:57.855 17:35:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 90305 00:13:58.122 killing process with pid 90305 00:13:58.122 17:35:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:13:58.122 17:35:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:13:58.122 17:35:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 90305' 00:13:58.122 17:35:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@969 -- # kill 90305 00:13:58.122 [2024-11-27 17:35:29.071038] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:13:58.122 17:35:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@974 -- # wait 90305 00:13:58.122 [2024-11-27 17:35:29.101786] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:13:58.382 17:35:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:13:58.382 00:13:58.382 real 0m9.124s 00:13:58.382 user 0m15.547s 00:13:58.382 sys 0m1.930s 00:13:58.382 ************************************ 00:13:58.382 END TEST raid5f_state_function_test 00:13:58.382 ************************************ 00:13:58.382 17:35:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:13:58.382 17:35:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:13:58.382 17:35:29 bdev_raid -- bdev/bdev_raid.sh@987 -- # run_test raid5f_state_function_test_sb raid_state_function_test raid5f 3 true 00:13:58.382 17:35:29 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:13:58.382 17:35:29 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:13:58.382 17:35:29 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:13:58.382 ************************************ 00:13:58.382 START TEST raid5f_state_function_test_sb 00:13:58.382 ************************************ 00:13:58.382 17:35:29 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@1125 -- # raid_state_function_test raid5f 3 true 00:13:58.382 17:35:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=raid5f 00:13:58.382 17:35:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=3 00:13:58.382 17:35:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:13:58.382 17:35:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:13:58.382 17:35:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:13:58.382 17:35:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:13:58.382 17:35:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:13:58.382 17:35:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:13:58.382 17:35:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:13:58.382 17:35:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:13:58.382 17:35:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:13:58.382 17:35:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:13:58.382 17:35:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:13:58.382 17:35:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:13:58.382 17:35:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:13:58.382 17:35:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:13:58.382 17:35:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:13:58.382 17:35:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:13:58.382 17:35:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:13:58.382 17:35:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:13:58.382 17:35:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:13:58.382 17:35:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' raid5f '!=' raid1 ']' 00:13:58.382 17:35:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:13:58.382 17:35:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:13:58.382 17:35:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:13:58.382 17:35:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:13:58.382 Process raid pid: 90910 00:13:58.382 17:35:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=90910 00:13:58.382 17:35:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 90910' 00:13:58.382 17:35:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 90910 00:13:58.382 17:35:29 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@831 -- # '[' -z 90910 ']' 00:13:58.382 17:35:29 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:13:58.382 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:13:58.382 17:35:29 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:13:58.382 17:35:29 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@836 -- # local max_retries=100 00:13:58.382 17:35:29 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:13:58.382 17:35:29 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@840 -- # xtrace_disable 00:13:58.382 17:35:29 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:58.382 [2024-11-27 17:35:29.515985] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:13:58.382 [2024-11-27 17:35:29.516101] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:13:58.643 [2024-11-27 17:35:29.663847] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:13:58.643 [2024-11-27 17:35:29.709520] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:13:58.643 [2024-11-27 17:35:29.751997] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:13:58.643 [2024-11-27 17:35:29.752034] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:13:59.215 17:35:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:13:59.215 17:35:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@864 -- # return 0 00:13:59.215 17:35:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:13:59.215 17:35:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:59.215 17:35:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:59.215 [2024-11-27 17:35:30.337547] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:13:59.215 [2024-11-27 17:35:30.337591] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:13:59.215 [2024-11-27 17:35:30.337602] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:13:59.215 [2024-11-27 17:35:30.337611] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:13:59.215 [2024-11-27 17:35:30.337616] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:13:59.215 [2024-11-27 17:35:30.337628] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:13:59.215 17:35:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:59.215 17:35:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:59.215 17:35:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:59.215 17:35:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:59.215 17:35:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:59.215 17:35:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:59.215 17:35:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:59.215 17:35:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:59.215 17:35:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:59.215 17:35:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:59.215 17:35:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:59.215 17:35:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:59.215 17:35:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:59.215 17:35:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:59.215 17:35:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:59.215 17:35:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:59.215 17:35:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:59.215 "name": "Existed_Raid", 00:13:59.215 "uuid": "34a337de-8f7a-4034-a423-25572de44686", 00:13:59.215 "strip_size_kb": 64, 00:13:59.215 "state": "configuring", 00:13:59.215 "raid_level": "raid5f", 00:13:59.215 "superblock": true, 00:13:59.215 "num_base_bdevs": 3, 00:13:59.215 "num_base_bdevs_discovered": 0, 00:13:59.215 "num_base_bdevs_operational": 3, 00:13:59.215 "base_bdevs_list": [ 00:13:59.215 { 00:13:59.215 "name": "BaseBdev1", 00:13:59.215 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:59.215 "is_configured": false, 00:13:59.215 "data_offset": 0, 00:13:59.215 "data_size": 0 00:13:59.215 }, 00:13:59.215 { 00:13:59.215 "name": "BaseBdev2", 00:13:59.215 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:59.215 "is_configured": false, 00:13:59.215 "data_offset": 0, 00:13:59.215 "data_size": 0 00:13:59.215 }, 00:13:59.215 { 00:13:59.215 "name": "BaseBdev3", 00:13:59.215 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:59.215 "is_configured": false, 00:13:59.215 "data_offset": 0, 00:13:59.215 "data_size": 0 00:13:59.215 } 00:13:59.215 ] 00:13:59.215 }' 00:13:59.215 17:35:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:59.215 17:35:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:59.784 17:35:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:13:59.784 17:35:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:59.784 17:35:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:59.784 [2024-11-27 17:35:30.780663] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:13:59.784 [2024-11-27 17:35:30.780752] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:13:59.784 17:35:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:59.784 17:35:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:13:59.784 17:35:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:59.784 17:35:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:59.784 [2024-11-27 17:35:30.792669] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:13:59.784 [2024-11-27 17:35:30.792739] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:13:59.784 [2024-11-27 17:35:30.792762] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:13:59.784 [2024-11-27 17:35:30.792784] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:13:59.784 [2024-11-27 17:35:30.792800] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:13:59.785 [2024-11-27 17:35:30.792819] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:13:59.785 17:35:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:59.785 17:35:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:13:59.785 17:35:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:59.785 17:35:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:59.785 [2024-11-27 17:35:30.813374] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:13:59.785 BaseBdev1 00:13:59.785 17:35:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:59.785 17:35:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:13:59.785 17:35:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:13:59.785 17:35:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:13:59.785 17:35:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:13:59.785 17:35:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:13:59.785 17:35:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:13:59.785 17:35:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:13:59.785 17:35:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:59.785 17:35:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:59.785 17:35:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:59.785 17:35:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:13:59.785 17:35:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:59.785 17:35:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:59.785 [ 00:13:59.785 { 00:13:59.785 "name": "BaseBdev1", 00:13:59.785 "aliases": [ 00:13:59.785 "85f5d588-707e-4e4d-83d2-f117c95a7f3b" 00:13:59.785 ], 00:13:59.785 "product_name": "Malloc disk", 00:13:59.785 "block_size": 512, 00:13:59.785 "num_blocks": 65536, 00:13:59.785 "uuid": "85f5d588-707e-4e4d-83d2-f117c95a7f3b", 00:13:59.785 "assigned_rate_limits": { 00:13:59.785 "rw_ios_per_sec": 0, 00:13:59.785 "rw_mbytes_per_sec": 0, 00:13:59.785 "r_mbytes_per_sec": 0, 00:13:59.785 "w_mbytes_per_sec": 0 00:13:59.785 }, 00:13:59.785 "claimed": true, 00:13:59.785 "claim_type": "exclusive_write", 00:13:59.785 "zoned": false, 00:13:59.785 "supported_io_types": { 00:13:59.785 "read": true, 00:13:59.785 "write": true, 00:13:59.785 "unmap": true, 00:13:59.785 "flush": true, 00:13:59.785 "reset": true, 00:13:59.785 "nvme_admin": false, 00:13:59.785 "nvme_io": false, 00:13:59.785 "nvme_io_md": false, 00:13:59.785 "write_zeroes": true, 00:13:59.785 "zcopy": true, 00:13:59.785 "get_zone_info": false, 00:13:59.785 "zone_management": false, 00:13:59.785 "zone_append": false, 00:13:59.785 "compare": false, 00:13:59.785 "compare_and_write": false, 00:13:59.785 "abort": true, 00:13:59.785 "seek_hole": false, 00:13:59.785 "seek_data": false, 00:13:59.785 "copy": true, 00:13:59.785 "nvme_iov_md": false 00:13:59.785 }, 00:13:59.785 "memory_domains": [ 00:13:59.785 { 00:13:59.785 "dma_device_id": "system", 00:13:59.785 "dma_device_type": 1 00:13:59.785 }, 00:13:59.785 { 00:13:59.785 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:13:59.785 "dma_device_type": 2 00:13:59.785 } 00:13:59.785 ], 00:13:59.785 "driver_specific": {} 00:13:59.785 } 00:13:59.785 ] 00:13:59.785 17:35:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:59.785 17:35:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:13:59.785 17:35:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:13:59.785 17:35:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:13:59.785 17:35:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:13:59.785 17:35:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:13:59.785 17:35:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:13:59.785 17:35:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:13:59.785 17:35:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:13:59.785 17:35:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:13:59.785 17:35:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:13:59.785 17:35:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:13:59.785 17:35:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:13:59.785 17:35:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:13:59.785 17:35:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:59.785 17:35:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:13:59.785 17:35:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:59.785 17:35:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:13:59.785 "name": "Existed_Raid", 00:13:59.785 "uuid": "d5063325-8d18-4d36-8f33-ecf27a06a104", 00:13:59.785 "strip_size_kb": 64, 00:13:59.785 "state": "configuring", 00:13:59.785 "raid_level": "raid5f", 00:13:59.785 "superblock": true, 00:13:59.785 "num_base_bdevs": 3, 00:13:59.785 "num_base_bdevs_discovered": 1, 00:13:59.785 "num_base_bdevs_operational": 3, 00:13:59.785 "base_bdevs_list": [ 00:13:59.785 { 00:13:59.785 "name": "BaseBdev1", 00:13:59.785 "uuid": "85f5d588-707e-4e4d-83d2-f117c95a7f3b", 00:13:59.785 "is_configured": true, 00:13:59.785 "data_offset": 2048, 00:13:59.785 "data_size": 63488 00:13:59.785 }, 00:13:59.785 { 00:13:59.785 "name": "BaseBdev2", 00:13:59.785 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:59.785 "is_configured": false, 00:13:59.785 "data_offset": 0, 00:13:59.785 "data_size": 0 00:13:59.785 }, 00:13:59.785 { 00:13:59.785 "name": "BaseBdev3", 00:13:59.785 "uuid": "00000000-0000-0000-0000-000000000000", 00:13:59.785 "is_configured": false, 00:13:59.785 "data_offset": 0, 00:13:59.785 "data_size": 0 00:13:59.785 } 00:13:59.785 ] 00:13:59.785 }' 00:13:59.785 17:35:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:13:59.785 17:35:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:00.044 17:35:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:14:00.044 17:35:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:00.044 17:35:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:00.044 [2024-11-27 17:35:31.228663] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:14:00.044 [2024-11-27 17:35:31.228743] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:14:00.044 17:35:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:00.044 17:35:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:14:00.303 17:35:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:00.303 17:35:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:00.303 [2024-11-27 17:35:31.236708] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:14:00.303 [2024-11-27 17:35:31.238462] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:14:00.303 [2024-11-27 17:35:31.238500] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:14:00.303 [2024-11-27 17:35:31.238509] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:14:00.303 [2024-11-27 17:35:31.238519] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:14:00.303 17:35:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:00.303 17:35:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:14:00.303 17:35:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:14:00.303 17:35:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:14:00.303 17:35:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:00.303 17:35:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:00.303 17:35:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:00.303 17:35:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:00.303 17:35:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:00.304 17:35:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:00.304 17:35:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:00.304 17:35:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:00.304 17:35:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:00.304 17:35:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:00.304 17:35:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:00.304 17:35:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:00.304 17:35:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:00.304 17:35:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:00.304 17:35:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:00.304 "name": "Existed_Raid", 00:14:00.304 "uuid": "a2a6a3de-2832-4814-a11e-43c5de71d499", 00:14:00.304 "strip_size_kb": 64, 00:14:00.304 "state": "configuring", 00:14:00.304 "raid_level": "raid5f", 00:14:00.304 "superblock": true, 00:14:00.304 "num_base_bdevs": 3, 00:14:00.304 "num_base_bdevs_discovered": 1, 00:14:00.304 "num_base_bdevs_operational": 3, 00:14:00.304 "base_bdevs_list": [ 00:14:00.304 { 00:14:00.304 "name": "BaseBdev1", 00:14:00.304 "uuid": "85f5d588-707e-4e4d-83d2-f117c95a7f3b", 00:14:00.304 "is_configured": true, 00:14:00.304 "data_offset": 2048, 00:14:00.304 "data_size": 63488 00:14:00.304 }, 00:14:00.304 { 00:14:00.304 "name": "BaseBdev2", 00:14:00.304 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:00.304 "is_configured": false, 00:14:00.304 "data_offset": 0, 00:14:00.304 "data_size": 0 00:14:00.304 }, 00:14:00.304 { 00:14:00.304 "name": "BaseBdev3", 00:14:00.304 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:00.304 "is_configured": false, 00:14:00.304 "data_offset": 0, 00:14:00.304 "data_size": 0 00:14:00.304 } 00:14:00.304 ] 00:14:00.304 }' 00:14:00.304 17:35:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:00.304 17:35:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:00.563 17:35:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:14:00.563 17:35:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:00.563 17:35:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:00.563 [2024-11-27 17:35:31.715907] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:14:00.563 BaseBdev2 00:14:00.563 17:35:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:00.563 17:35:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:14:00.563 17:35:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:14:00.563 17:35:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:14:00.563 17:35:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:14:00.563 17:35:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:14:00.563 17:35:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:14:00.563 17:35:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:14:00.563 17:35:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:00.564 17:35:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:00.564 17:35:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:00.564 17:35:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:14:00.564 17:35:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:00.564 17:35:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:00.564 [ 00:14:00.564 { 00:14:00.564 "name": "BaseBdev2", 00:14:00.564 "aliases": [ 00:14:00.564 "9a069dca-aee4-4442-9180-4813dcba767c" 00:14:00.564 ], 00:14:00.564 "product_name": "Malloc disk", 00:14:00.564 "block_size": 512, 00:14:00.564 "num_blocks": 65536, 00:14:00.564 "uuid": "9a069dca-aee4-4442-9180-4813dcba767c", 00:14:00.564 "assigned_rate_limits": { 00:14:00.564 "rw_ios_per_sec": 0, 00:14:00.564 "rw_mbytes_per_sec": 0, 00:14:00.564 "r_mbytes_per_sec": 0, 00:14:00.564 "w_mbytes_per_sec": 0 00:14:00.564 }, 00:14:00.564 "claimed": true, 00:14:00.564 "claim_type": "exclusive_write", 00:14:00.564 "zoned": false, 00:14:00.564 "supported_io_types": { 00:14:00.564 "read": true, 00:14:00.564 "write": true, 00:14:00.564 "unmap": true, 00:14:00.564 "flush": true, 00:14:00.564 "reset": true, 00:14:00.564 "nvme_admin": false, 00:14:00.564 "nvme_io": false, 00:14:00.564 "nvme_io_md": false, 00:14:00.564 "write_zeroes": true, 00:14:00.564 "zcopy": true, 00:14:00.564 "get_zone_info": false, 00:14:00.564 "zone_management": false, 00:14:00.564 "zone_append": false, 00:14:00.564 "compare": false, 00:14:00.564 "compare_and_write": false, 00:14:00.564 "abort": true, 00:14:00.564 "seek_hole": false, 00:14:00.823 "seek_data": false, 00:14:00.823 "copy": true, 00:14:00.823 "nvme_iov_md": false 00:14:00.823 }, 00:14:00.823 "memory_domains": [ 00:14:00.823 { 00:14:00.823 "dma_device_id": "system", 00:14:00.823 "dma_device_type": 1 00:14:00.823 }, 00:14:00.823 { 00:14:00.823 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:00.823 "dma_device_type": 2 00:14:00.823 } 00:14:00.823 ], 00:14:00.823 "driver_specific": {} 00:14:00.823 } 00:14:00.823 ] 00:14:00.823 17:35:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:00.823 17:35:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:14:00.823 17:35:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:14:00.823 17:35:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:14:00.823 17:35:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:14:00.823 17:35:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:00.823 17:35:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:00.823 17:35:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:00.823 17:35:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:00.823 17:35:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:00.823 17:35:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:00.823 17:35:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:00.823 17:35:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:00.823 17:35:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:00.823 17:35:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:00.823 17:35:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:00.823 17:35:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:00.823 17:35:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:00.823 17:35:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:00.823 17:35:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:00.823 "name": "Existed_Raid", 00:14:00.823 "uuid": "a2a6a3de-2832-4814-a11e-43c5de71d499", 00:14:00.823 "strip_size_kb": 64, 00:14:00.823 "state": "configuring", 00:14:00.823 "raid_level": "raid5f", 00:14:00.823 "superblock": true, 00:14:00.823 "num_base_bdevs": 3, 00:14:00.823 "num_base_bdevs_discovered": 2, 00:14:00.823 "num_base_bdevs_operational": 3, 00:14:00.823 "base_bdevs_list": [ 00:14:00.823 { 00:14:00.823 "name": "BaseBdev1", 00:14:00.823 "uuid": "85f5d588-707e-4e4d-83d2-f117c95a7f3b", 00:14:00.823 "is_configured": true, 00:14:00.823 "data_offset": 2048, 00:14:00.823 "data_size": 63488 00:14:00.823 }, 00:14:00.823 { 00:14:00.823 "name": "BaseBdev2", 00:14:00.823 "uuid": "9a069dca-aee4-4442-9180-4813dcba767c", 00:14:00.823 "is_configured": true, 00:14:00.823 "data_offset": 2048, 00:14:00.823 "data_size": 63488 00:14:00.823 }, 00:14:00.823 { 00:14:00.823 "name": "BaseBdev3", 00:14:00.824 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:00.824 "is_configured": false, 00:14:00.824 "data_offset": 0, 00:14:00.824 "data_size": 0 00:14:00.824 } 00:14:00.824 ] 00:14:00.824 }' 00:14:00.824 17:35:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:00.824 17:35:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:01.083 17:35:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:14:01.083 17:35:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:01.083 17:35:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:01.083 [2024-11-27 17:35:32.217925] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:14:01.083 [2024-11-27 17:35:32.218104] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:14:01.083 [2024-11-27 17:35:32.218120] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:14:01.083 [2024-11-27 17:35:32.218406] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:14:01.083 BaseBdev3 00:14:01.083 [2024-11-27 17:35:32.218826] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:14:01.083 [2024-11-27 17:35:32.218847] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:14:01.083 [2024-11-27 17:35:32.218979] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:01.083 17:35:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:01.083 17:35:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:14:01.083 17:35:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:14:01.083 17:35:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:14:01.083 17:35:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:14:01.083 17:35:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:14:01.083 17:35:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:14:01.083 17:35:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:14:01.083 17:35:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:01.083 17:35:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:01.083 17:35:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:01.083 17:35:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:14:01.083 17:35:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:01.083 17:35:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:01.083 [ 00:14:01.083 { 00:14:01.083 "name": "BaseBdev3", 00:14:01.083 "aliases": [ 00:14:01.083 "d69ed9bc-0e9b-4b1e-9844-eee5f37e1b7d" 00:14:01.083 ], 00:14:01.083 "product_name": "Malloc disk", 00:14:01.083 "block_size": 512, 00:14:01.083 "num_blocks": 65536, 00:14:01.083 "uuid": "d69ed9bc-0e9b-4b1e-9844-eee5f37e1b7d", 00:14:01.083 "assigned_rate_limits": { 00:14:01.083 "rw_ios_per_sec": 0, 00:14:01.083 "rw_mbytes_per_sec": 0, 00:14:01.083 "r_mbytes_per_sec": 0, 00:14:01.083 "w_mbytes_per_sec": 0 00:14:01.083 }, 00:14:01.083 "claimed": true, 00:14:01.083 "claim_type": "exclusive_write", 00:14:01.083 "zoned": false, 00:14:01.083 "supported_io_types": { 00:14:01.083 "read": true, 00:14:01.083 "write": true, 00:14:01.083 "unmap": true, 00:14:01.083 "flush": true, 00:14:01.083 "reset": true, 00:14:01.083 "nvme_admin": false, 00:14:01.083 "nvme_io": false, 00:14:01.083 "nvme_io_md": false, 00:14:01.083 "write_zeroes": true, 00:14:01.083 "zcopy": true, 00:14:01.083 "get_zone_info": false, 00:14:01.083 "zone_management": false, 00:14:01.083 "zone_append": false, 00:14:01.083 "compare": false, 00:14:01.083 "compare_and_write": false, 00:14:01.083 "abort": true, 00:14:01.083 "seek_hole": false, 00:14:01.083 "seek_data": false, 00:14:01.083 "copy": true, 00:14:01.083 "nvme_iov_md": false 00:14:01.083 }, 00:14:01.083 "memory_domains": [ 00:14:01.083 { 00:14:01.083 "dma_device_id": "system", 00:14:01.083 "dma_device_type": 1 00:14:01.083 }, 00:14:01.083 { 00:14:01.083 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:01.083 "dma_device_type": 2 00:14:01.083 } 00:14:01.083 ], 00:14:01.083 "driver_specific": {} 00:14:01.083 } 00:14:01.083 ] 00:14:01.084 17:35:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:01.084 17:35:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:14:01.084 17:35:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:14:01.084 17:35:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:14:01.084 17:35:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 3 00:14:01.084 17:35:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:01.084 17:35:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:01.084 17:35:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:01.084 17:35:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:01.084 17:35:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:01.084 17:35:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:01.084 17:35:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:01.084 17:35:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:01.084 17:35:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:01.084 17:35:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:01.084 17:35:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:01.084 17:35:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:01.084 17:35:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:01.343 17:35:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:01.343 17:35:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:01.343 "name": "Existed_Raid", 00:14:01.343 "uuid": "a2a6a3de-2832-4814-a11e-43c5de71d499", 00:14:01.343 "strip_size_kb": 64, 00:14:01.343 "state": "online", 00:14:01.343 "raid_level": "raid5f", 00:14:01.343 "superblock": true, 00:14:01.343 "num_base_bdevs": 3, 00:14:01.343 "num_base_bdevs_discovered": 3, 00:14:01.343 "num_base_bdevs_operational": 3, 00:14:01.343 "base_bdevs_list": [ 00:14:01.343 { 00:14:01.343 "name": "BaseBdev1", 00:14:01.343 "uuid": "85f5d588-707e-4e4d-83d2-f117c95a7f3b", 00:14:01.343 "is_configured": true, 00:14:01.343 "data_offset": 2048, 00:14:01.343 "data_size": 63488 00:14:01.343 }, 00:14:01.343 { 00:14:01.343 "name": "BaseBdev2", 00:14:01.343 "uuid": "9a069dca-aee4-4442-9180-4813dcba767c", 00:14:01.343 "is_configured": true, 00:14:01.343 "data_offset": 2048, 00:14:01.343 "data_size": 63488 00:14:01.343 }, 00:14:01.343 { 00:14:01.343 "name": "BaseBdev3", 00:14:01.343 "uuid": "d69ed9bc-0e9b-4b1e-9844-eee5f37e1b7d", 00:14:01.343 "is_configured": true, 00:14:01.343 "data_offset": 2048, 00:14:01.343 "data_size": 63488 00:14:01.343 } 00:14:01.343 ] 00:14:01.343 }' 00:14:01.343 17:35:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:01.343 17:35:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:01.601 17:35:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:14:01.601 17:35:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:14:01.601 17:35:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:14:01.601 17:35:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:14:01.601 17:35:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:14:01.601 17:35:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:14:01.601 17:35:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:14:01.602 17:35:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:14:01.602 17:35:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:01.602 17:35:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:01.602 [2024-11-27 17:35:32.693335] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:01.602 17:35:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:01.602 17:35:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:14:01.602 "name": "Existed_Raid", 00:14:01.602 "aliases": [ 00:14:01.602 "a2a6a3de-2832-4814-a11e-43c5de71d499" 00:14:01.602 ], 00:14:01.602 "product_name": "Raid Volume", 00:14:01.602 "block_size": 512, 00:14:01.602 "num_blocks": 126976, 00:14:01.602 "uuid": "a2a6a3de-2832-4814-a11e-43c5de71d499", 00:14:01.602 "assigned_rate_limits": { 00:14:01.602 "rw_ios_per_sec": 0, 00:14:01.602 "rw_mbytes_per_sec": 0, 00:14:01.602 "r_mbytes_per_sec": 0, 00:14:01.602 "w_mbytes_per_sec": 0 00:14:01.602 }, 00:14:01.602 "claimed": false, 00:14:01.602 "zoned": false, 00:14:01.602 "supported_io_types": { 00:14:01.602 "read": true, 00:14:01.602 "write": true, 00:14:01.602 "unmap": false, 00:14:01.602 "flush": false, 00:14:01.602 "reset": true, 00:14:01.602 "nvme_admin": false, 00:14:01.602 "nvme_io": false, 00:14:01.602 "nvme_io_md": false, 00:14:01.602 "write_zeroes": true, 00:14:01.602 "zcopy": false, 00:14:01.602 "get_zone_info": false, 00:14:01.602 "zone_management": false, 00:14:01.602 "zone_append": false, 00:14:01.602 "compare": false, 00:14:01.602 "compare_and_write": false, 00:14:01.602 "abort": false, 00:14:01.602 "seek_hole": false, 00:14:01.602 "seek_data": false, 00:14:01.602 "copy": false, 00:14:01.602 "nvme_iov_md": false 00:14:01.602 }, 00:14:01.602 "driver_specific": { 00:14:01.602 "raid": { 00:14:01.602 "uuid": "a2a6a3de-2832-4814-a11e-43c5de71d499", 00:14:01.602 "strip_size_kb": 64, 00:14:01.602 "state": "online", 00:14:01.602 "raid_level": "raid5f", 00:14:01.602 "superblock": true, 00:14:01.602 "num_base_bdevs": 3, 00:14:01.602 "num_base_bdevs_discovered": 3, 00:14:01.602 "num_base_bdevs_operational": 3, 00:14:01.602 "base_bdevs_list": [ 00:14:01.602 { 00:14:01.602 "name": "BaseBdev1", 00:14:01.602 "uuid": "85f5d588-707e-4e4d-83d2-f117c95a7f3b", 00:14:01.602 "is_configured": true, 00:14:01.602 "data_offset": 2048, 00:14:01.602 "data_size": 63488 00:14:01.602 }, 00:14:01.602 { 00:14:01.602 "name": "BaseBdev2", 00:14:01.602 "uuid": "9a069dca-aee4-4442-9180-4813dcba767c", 00:14:01.602 "is_configured": true, 00:14:01.602 "data_offset": 2048, 00:14:01.602 "data_size": 63488 00:14:01.602 }, 00:14:01.602 { 00:14:01.602 "name": "BaseBdev3", 00:14:01.602 "uuid": "d69ed9bc-0e9b-4b1e-9844-eee5f37e1b7d", 00:14:01.602 "is_configured": true, 00:14:01.602 "data_offset": 2048, 00:14:01.602 "data_size": 63488 00:14:01.602 } 00:14:01.602 ] 00:14:01.602 } 00:14:01.602 } 00:14:01.602 }' 00:14:01.602 17:35:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:14:01.602 17:35:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:14:01.602 BaseBdev2 00:14:01.602 BaseBdev3' 00:14:01.602 17:35:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:01.860 17:35:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:14:01.860 17:35:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:01.860 17:35:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:14:01.860 17:35:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:01.860 17:35:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:01.860 17:35:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:01.861 17:35:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:01.861 17:35:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:01.861 17:35:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:01.861 17:35:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:01.861 17:35:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:01.861 17:35:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:14:01.861 17:35:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:01.861 17:35:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:01.861 17:35:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:01.861 17:35:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:01.861 17:35:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:01.861 17:35:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:01.861 17:35:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:14:01.861 17:35:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:01.861 17:35:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:01.861 17:35:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:01.861 17:35:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:01.861 17:35:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:01.861 17:35:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:01.861 17:35:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:14:01.861 17:35:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:01.861 17:35:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:01.861 [2024-11-27 17:35:32.976708] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:14:01.861 17:35:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:01.861 17:35:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:14:01.861 17:35:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy raid5f 00:14:01.861 17:35:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:14:01.861 17:35:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@199 -- # return 0 00:14:01.861 17:35:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:14:01.861 17:35:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 2 00:14:01.861 17:35:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:01.861 17:35:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:01.861 17:35:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:01.861 17:35:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:01.861 17:35:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:14:01.861 17:35:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:01.861 17:35:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:01.861 17:35:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:01.861 17:35:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:01.861 17:35:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:01.861 17:35:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:01.861 17:35:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:01.861 17:35:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:01.861 17:35:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:01.861 17:35:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:01.861 "name": "Existed_Raid", 00:14:01.861 "uuid": "a2a6a3de-2832-4814-a11e-43c5de71d499", 00:14:01.861 "strip_size_kb": 64, 00:14:01.861 "state": "online", 00:14:01.861 "raid_level": "raid5f", 00:14:01.861 "superblock": true, 00:14:01.861 "num_base_bdevs": 3, 00:14:01.861 "num_base_bdevs_discovered": 2, 00:14:01.861 "num_base_bdevs_operational": 2, 00:14:01.861 "base_bdevs_list": [ 00:14:01.861 { 00:14:01.861 "name": null, 00:14:01.861 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:01.861 "is_configured": false, 00:14:01.861 "data_offset": 0, 00:14:01.861 "data_size": 63488 00:14:01.861 }, 00:14:01.861 { 00:14:01.861 "name": "BaseBdev2", 00:14:01.861 "uuid": "9a069dca-aee4-4442-9180-4813dcba767c", 00:14:01.861 "is_configured": true, 00:14:01.861 "data_offset": 2048, 00:14:01.861 "data_size": 63488 00:14:01.861 }, 00:14:01.861 { 00:14:01.861 "name": "BaseBdev3", 00:14:01.861 "uuid": "d69ed9bc-0e9b-4b1e-9844-eee5f37e1b7d", 00:14:01.861 "is_configured": true, 00:14:01.861 "data_offset": 2048, 00:14:01.861 "data_size": 63488 00:14:01.861 } 00:14:01.861 ] 00:14:01.861 }' 00:14:01.861 17:35:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:01.861 17:35:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:02.428 17:35:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:14:02.428 17:35:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:14:02.428 17:35:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:14:02.428 17:35:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:02.429 17:35:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:02.429 17:35:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:02.429 17:35:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:02.429 17:35:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:14:02.429 17:35:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:14:02.429 17:35:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:14:02.429 17:35:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:02.429 17:35:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:02.429 [2024-11-27 17:35:33.471271] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:14:02.429 [2024-11-27 17:35:33.471432] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:14:02.429 [2024-11-27 17:35:33.482382] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:02.429 17:35:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:02.429 17:35:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:14:02.429 17:35:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:14:02.429 17:35:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:14:02.429 17:35:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:02.429 17:35:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:02.429 17:35:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:02.429 17:35:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:02.429 17:35:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:14:02.429 17:35:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:14:02.429 17:35:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:14:02.429 17:35:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:02.429 17:35:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:02.429 [2024-11-27 17:35:33.526374] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:14:02.429 [2024-11-27 17:35:33.526413] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:14:02.429 17:35:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:02.429 17:35:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:14:02.429 17:35:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:14:02.429 17:35:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:02.429 17:35:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:02.429 17:35:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:14:02.429 17:35:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:02.429 17:35:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:02.429 17:35:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:14:02.429 17:35:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:14:02.429 17:35:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 3 -gt 2 ']' 00:14:02.429 17:35:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:14:02.429 17:35:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:14:02.429 17:35:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:14:02.429 17:35:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:02.429 17:35:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:02.429 BaseBdev2 00:14:02.429 17:35:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:02.429 17:35:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:14:02.429 17:35:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:14:02.429 17:35:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:14:02.429 17:35:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:14:02.429 17:35:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:14:02.429 17:35:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:14:02.429 17:35:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:14:02.429 17:35:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:02.429 17:35:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:02.429 17:35:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:02.429 17:35:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:14:02.429 17:35:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:02.429 17:35:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:02.689 [ 00:14:02.689 { 00:14:02.689 "name": "BaseBdev2", 00:14:02.689 "aliases": [ 00:14:02.689 "62f68267-a872-41d1-adef-cf167e8b5f59" 00:14:02.689 ], 00:14:02.689 "product_name": "Malloc disk", 00:14:02.689 "block_size": 512, 00:14:02.689 "num_blocks": 65536, 00:14:02.689 "uuid": "62f68267-a872-41d1-adef-cf167e8b5f59", 00:14:02.689 "assigned_rate_limits": { 00:14:02.689 "rw_ios_per_sec": 0, 00:14:02.689 "rw_mbytes_per_sec": 0, 00:14:02.689 "r_mbytes_per_sec": 0, 00:14:02.689 "w_mbytes_per_sec": 0 00:14:02.689 }, 00:14:02.689 "claimed": false, 00:14:02.689 "zoned": false, 00:14:02.689 "supported_io_types": { 00:14:02.689 "read": true, 00:14:02.689 "write": true, 00:14:02.689 "unmap": true, 00:14:02.689 "flush": true, 00:14:02.689 "reset": true, 00:14:02.689 "nvme_admin": false, 00:14:02.689 "nvme_io": false, 00:14:02.689 "nvme_io_md": false, 00:14:02.689 "write_zeroes": true, 00:14:02.689 "zcopy": true, 00:14:02.689 "get_zone_info": false, 00:14:02.689 "zone_management": false, 00:14:02.689 "zone_append": false, 00:14:02.689 "compare": false, 00:14:02.689 "compare_and_write": false, 00:14:02.689 "abort": true, 00:14:02.689 "seek_hole": false, 00:14:02.689 "seek_data": false, 00:14:02.689 "copy": true, 00:14:02.689 "nvme_iov_md": false 00:14:02.689 }, 00:14:02.689 "memory_domains": [ 00:14:02.689 { 00:14:02.689 "dma_device_id": "system", 00:14:02.689 "dma_device_type": 1 00:14:02.689 }, 00:14:02.689 { 00:14:02.689 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:02.689 "dma_device_type": 2 00:14:02.689 } 00:14:02.689 ], 00:14:02.689 "driver_specific": {} 00:14:02.689 } 00:14:02.689 ] 00:14:02.689 17:35:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:02.689 17:35:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:14:02.689 17:35:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:14:02.689 17:35:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:14:02.689 17:35:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:14:02.689 17:35:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:02.689 17:35:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:02.689 BaseBdev3 00:14:02.689 17:35:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:02.689 17:35:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:14:02.689 17:35:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:14:02.689 17:35:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:14:02.689 17:35:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:14:02.689 17:35:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:14:02.689 17:35:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:14:02.689 17:35:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:14:02.689 17:35:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:02.689 17:35:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:02.689 17:35:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:02.689 17:35:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:14:02.689 17:35:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:02.689 17:35:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:02.689 [ 00:14:02.689 { 00:14:02.689 "name": "BaseBdev3", 00:14:02.689 "aliases": [ 00:14:02.689 "da68fb38-9430-4f11-8cab-43e5ab9673bd" 00:14:02.689 ], 00:14:02.689 "product_name": "Malloc disk", 00:14:02.689 "block_size": 512, 00:14:02.689 "num_blocks": 65536, 00:14:02.689 "uuid": "da68fb38-9430-4f11-8cab-43e5ab9673bd", 00:14:02.689 "assigned_rate_limits": { 00:14:02.689 "rw_ios_per_sec": 0, 00:14:02.689 "rw_mbytes_per_sec": 0, 00:14:02.689 "r_mbytes_per_sec": 0, 00:14:02.689 "w_mbytes_per_sec": 0 00:14:02.689 }, 00:14:02.689 "claimed": false, 00:14:02.689 "zoned": false, 00:14:02.689 "supported_io_types": { 00:14:02.689 "read": true, 00:14:02.689 "write": true, 00:14:02.689 "unmap": true, 00:14:02.689 "flush": true, 00:14:02.689 "reset": true, 00:14:02.689 "nvme_admin": false, 00:14:02.689 "nvme_io": false, 00:14:02.689 "nvme_io_md": false, 00:14:02.689 "write_zeroes": true, 00:14:02.689 "zcopy": true, 00:14:02.689 "get_zone_info": false, 00:14:02.689 "zone_management": false, 00:14:02.689 "zone_append": false, 00:14:02.689 "compare": false, 00:14:02.689 "compare_and_write": false, 00:14:02.689 "abort": true, 00:14:02.689 "seek_hole": false, 00:14:02.689 "seek_data": false, 00:14:02.689 "copy": true, 00:14:02.689 "nvme_iov_md": false 00:14:02.689 }, 00:14:02.689 "memory_domains": [ 00:14:02.690 { 00:14:02.690 "dma_device_id": "system", 00:14:02.690 "dma_device_type": 1 00:14:02.690 }, 00:14:02.690 { 00:14:02.690 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:02.690 "dma_device_type": 2 00:14:02.690 } 00:14:02.690 ], 00:14:02.690 "driver_specific": {} 00:14:02.690 } 00:14:02.690 ] 00:14:02.690 17:35:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:02.690 17:35:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:14:02.690 17:35:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:14:02.690 17:35:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:14:02.690 17:35:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n Existed_Raid 00:14:02.690 17:35:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:02.690 17:35:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:02.690 [2024-11-27 17:35:33.703752] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:14:02.690 [2024-11-27 17:35:33.703834] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:14:02.690 [2024-11-27 17:35:33.703871] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:14:02.690 [2024-11-27 17:35:33.705582] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:14:02.690 17:35:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:02.690 17:35:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:14:02.690 17:35:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:02.690 17:35:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:02.690 17:35:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:02.690 17:35:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:02.690 17:35:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:02.690 17:35:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:02.690 17:35:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:02.690 17:35:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:02.690 17:35:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:02.690 17:35:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:02.690 17:35:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:02.690 17:35:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:02.690 17:35:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:02.690 17:35:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:02.690 17:35:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:02.690 "name": "Existed_Raid", 00:14:02.690 "uuid": "0148c0f3-cacd-4c33-a619-29885e43851c", 00:14:02.690 "strip_size_kb": 64, 00:14:02.690 "state": "configuring", 00:14:02.690 "raid_level": "raid5f", 00:14:02.690 "superblock": true, 00:14:02.690 "num_base_bdevs": 3, 00:14:02.690 "num_base_bdevs_discovered": 2, 00:14:02.690 "num_base_bdevs_operational": 3, 00:14:02.690 "base_bdevs_list": [ 00:14:02.690 { 00:14:02.690 "name": "BaseBdev1", 00:14:02.690 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:02.690 "is_configured": false, 00:14:02.690 "data_offset": 0, 00:14:02.690 "data_size": 0 00:14:02.690 }, 00:14:02.690 { 00:14:02.690 "name": "BaseBdev2", 00:14:02.690 "uuid": "62f68267-a872-41d1-adef-cf167e8b5f59", 00:14:02.690 "is_configured": true, 00:14:02.690 "data_offset": 2048, 00:14:02.690 "data_size": 63488 00:14:02.690 }, 00:14:02.690 { 00:14:02.690 "name": "BaseBdev3", 00:14:02.690 "uuid": "da68fb38-9430-4f11-8cab-43e5ab9673bd", 00:14:02.690 "is_configured": true, 00:14:02.690 "data_offset": 2048, 00:14:02.690 "data_size": 63488 00:14:02.690 } 00:14:02.690 ] 00:14:02.690 }' 00:14:02.690 17:35:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:02.690 17:35:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:03.259 17:35:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:14:03.259 17:35:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:03.259 17:35:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:03.259 [2024-11-27 17:35:34.171109] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:14:03.259 17:35:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:03.259 17:35:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:14:03.259 17:35:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:03.259 17:35:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:03.259 17:35:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:03.259 17:35:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:03.259 17:35:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:03.259 17:35:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:03.259 17:35:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:03.259 17:35:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:03.259 17:35:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:03.259 17:35:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:03.259 17:35:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:03.259 17:35:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:03.259 17:35:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:03.259 17:35:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:03.259 17:35:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:03.259 "name": "Existed_Raid", 00:14:03.259 "uuid": "0148c0f3-cacd-4c33-a619-29885e43851c", 00:14:03.259 "strip_size_kb": 64, 00:14:03.259 "state": "configuring", 00:14:03.259 "raid_level": "raid5f", 00:14:03.259 "superblock": true, 00:14:03.259 "num_base_bdevs": 3, 00:14:03.259 "num_base_bdevs_discovered": 1, 00:14:03.259 "num_base_bdevs_operational": 3, 00:14:03.259 "base_bdevs_list": [ 00:14:03.259 { 00:14:03.259 "name": "BaseBdev1", 00:14:03.259 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:03.259 "is_configured": false, 00:14:03.259 "data_offset": 0, 00:14:03.259 "data_size": 0 00:14:03.259 }, 00:14:03.259 { 00:14:03.259 "name": null, 00:14:03.259 "uuid": "62f68267-a872-41d1-adef-cf167e8b5f59", 00:14:03.259 "is_configured": false, 00:14:03.259 "data_offset": 0, 00:14:03.259 "data_size": 63488 00:14:03.259 }, 00:14:03.259 { 00:14:03.259 "name": "BaseBdev3", 00:14:03.259 "uuid": "da68fb38-9430-4f11-8cab-43e5ab9673bd", 00:14:03.259 "is_configured": true, 00:14:03.259 "data_offset": 2048, 00:14:03.259 "data_size": 63488 00:14:03.259 } 00:14:03.259 ] 00:14:03.259 }' 00:14:03.259 17:35:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:03.259 17:35:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:03.519 17:35:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:03.519 17:35:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:14:03.519 17:35:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:03.519 17:35:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:03.519 17:35:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:03.519 17:35:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:14:03.519 17:35:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:14:03.519 17:35:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:03.519 17:35:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:03.519 [2024-11-27 17:35:34.673062] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:14:03.519 BaseBdev1 00:14:03.519 17:35:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:03.519 17:35:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:14:03.519 17:35:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:14:03.519 17:35:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:14:03.519 17:35:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:14:03.519 17:35:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:14:03.519 17:35:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:14:03.519 17:35:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:14:03.519 17:35:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:03.519 17:35:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:03.519 17:35:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:03.519 17:35:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:14:03.519 17:35:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:03.519 17:35:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:03.519 [ 00:14:03.519 { 00:14:03.519 "name": "BaseBdev1", 00:14:03.519 "aliases": [ 00:14:03.519 "08030365-66d3-4aae-92ce-01ca812a7a47" 00:14:03.519 ], 00:14:03.519 "product_name": "Malloc disk", 00:14:03.519 "block_size": 512, 00:14:03.519 "num_blocks": 65536, 00:14:03.519 "uuid": "08030365-66d3-4aae-92ce-01ca812a7a47", 00:14:03.519 "assigned_rate_limits": { 00:14:03.519 "rw_ios_per_sec": 0, 00:14:03.519 "rw_mbytes_per_sec": 0, 00:14:03.519 "r_mbytes_per_sec": 0, 00:14:03.519 "w_mbytes_per_sec": 0 00:14:03.519 }, 00:14:03.519 "claimed": true, 00:14:03.519 "claim_type": "exclusive_write", 00:14:03.519 "zoned": false, 00:14:03.519 "supported_io_types": { 00:14:03.519 "read": true, 00:14:03.519 "write": true, 00:14:03.519 "unmap": true, 00:14:03.519 "flush": true, 00:14:03.519 "reset": true, 00:14:03.519 "nvme_admin": false, 00:14:03.519 "nvme_io": false, 00:14:03.519 "nvme_io_md": false, 00:14:03.519 "write_zeroes": true, 00:14:03.519 "zcopy": true, 00:14:03.519 "get_zone_info": false, 00:14:03.519 "zone_management": false, 00:14:03.519 "zone_append": false, 00:14:03.519 "compare": false, 00:14:03.519 "compare_and_write": false, 00:14:03.519 "abort": true, 00:14:03.519 "seek_hole": false, 00:14:03.519 "seek_data": false, 00:14:03.519 "copy": true, 00:14:03.519 "nvme_iov_md": false 00:14:03.519 }, 00:14:03.780 "memory_domains": [ 00:14:03.780 { 00:14:03.780 "dma_device_id": "system", 00:14:03.780 "dma_device_type": 1 00:14:03.780 }, 00:14:03.780 { 00:14:03.780 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:03.780 "dma_device_type": 2 00:14:03.780 } 00:14:03.780 ], 00:14:03.780 "driver_specific": {} 00:14:03.780 } 00:14:03.780 ] 00:14:03.780 17:35:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:03.780 17:35:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:14:03.780 17:35:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:14:03.780 17:35:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:03.780 17:35:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:03.780 17:35:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:03.780 17:35:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:03.780 17:35:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:03.780 17:35:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:03.780 17:35:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:03.780 17:35:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:03.780 17:35:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:03.780 17:35:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:03.780 17:35:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:03.780 17:35:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:03.780 17:35:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:03.780 17:35:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:03.780 17:35:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:03.780 "name": "Existed_Raid", 00:14:03.780 "uuid": "0148c0f3-cacd-4c33-a619-29885e43851c", 00:14:03.780 "strip_size_kb": 64, 00:14:03.780 "state": "configuring", 00:14:03.780 "raid_level": "raid5f", 00:14:03.780 "superblock": true, 00:14:03.780 "num_base_bdevs": 3, 00:14:03.780 "num_base_bdevs_discovered": 2, 00:14:03.780 "num_base_bdevs_operational": 3, 00:14:03.780 "base_bdevs_list": [ 00:14:03.780 { 00:14:03.780 "name": "BaseBdev1", 00:14:03.780 "uuid": "08030365-66d3-4aae-92ce-01ca812a7a47", 00:14:03.780 "is_configured": true, 00:14:03.780 "data_offset": 2048, 00:14:03.780 "data_size": 63488 00:14:03.780 }, 00:14:03.780 { 00:14:03.780 "name": null, 00:14:03.780 "uuid": "62f68267-a872-41d1-adef-cf167e8b5f59", 00:14:03.780 "is_configured": false, 00:14:03.780 "data_offset": 0, 00:14:03.780 "data_size": 63488 00:14:03.780 }, 00:14:03.780 { 00:14:03.780 "name": "BaseBdev3", 00:14:03.780 "uuid": "da68fb38-9430-4f11-8cab-43e5ab9673bd", 00:14:03.780 "is_configured": true, 00:14:03.780 "data_offset": 2048, 00:14:03.780 "data_size": 63488 00:14:03.780 } 00:14:03.780 ] 00:14:03.780 }' 00:14:03.780 17:35:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:03.780 17:35:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:04.039 17:35:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:04.039 17:35:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:04.039 17:35:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:04.039 17:35:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:14:04.040 17:35:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:04.040 17:35:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:14:04.040 17:35:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:14:04.040 17:35:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:04.040 17:35:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:04.040 [2024-11-27 17:35:35.196231] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:14:04.040 17:35:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:04.040 17:35:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:14:04.040 17:35:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:04.040 17:35:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:04.040 17:35:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:04.040 17:35:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:04.040 17:35:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:04.040 17:35:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:04.040 17:35:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:04.040 17:35:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:04.040 17:35:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:04.040 17:35:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:04.040 17:35:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:04.040 17:35:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:04.040 17:35:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:04.040 17:35:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:04.300 17:35:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:04.300 "name": "Existed_Raid", 00:14:04.300 "uuid": "0148c0f3-cacd-4c33-a619-29885e43851c", 00:14:04.300 "strip_size_kb": 64, 00:14:04.300 "state": "configuring", 00:14:04.300 "raid_level": "raid5f", 00:14:04.300 "superblock": true, 00:14:04.300 "num_base_bdevs": 3, 00:14:04.300 "num_base_bdevs_discovered": 1, 00:14:04.300 "num_base_bdevs_operational": 3, 00:14:04.300 "base_bdevs_list": [ 00:14:04.300 { 00:14:04.300 "name": "BaseBdev1", 00:14:04.300 "uuid": "08030365-66d3-4aae-92ce-01ca812a7a47", 00:14:04.300 "is_configured": true, 00:14:04.300 "data_offset": 2048, 00:14:04.300 "data_size": 63488 00:14:04.300 }, 00:14:04.300 { 00:14:04.300 "name": null, 00:14:04.300 "uuid": "62f68267-a872-41d1-adef-cf167e8b5f59", 00:14:04.300 "is_configured": false, 00:14:04.300 "data_offset": 0, 00:14:04.300 "data_size": 63488 00:14:04.300 }, 00:14:04.300 { 00:14:04.300 "name": null, 00:14:04.300 "uuid": "da68fb38-9430-4f11-8cab-43e5ab9673bd", 00:14:04.300 "is_configured": false, 00:14:04.300 "data_offset": 0, 00:14:04.300 "data_size": 63488 00:14:04.300 } 00:14:04.300 ] 00:14:04.300 }' 00:14:04.300 17:35:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:04.300 17:35:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:04.560 17:35:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:04.560 17:35:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:04.560 17:35:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:14:04.560 17:35:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:04.560 17:35:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:04.560 17:35:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:14:04.560 17:35:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:14:04.560 17:35:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:04.560 17:35:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:04.560 [2024-11-27 17:35:35.683371] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:14:04.560 17:35:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:04.560 17:35:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:14:04.560 17:35:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:04.560 17:35:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:04.560 17:35:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:04.560 17:35:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:04.560 17:35:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:04.560 17:35:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:04.560 17:35:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:04.560 17:35:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:04.561 17:35:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:04.561 17:35:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:04.561 17:35:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:04.561 17:35:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:04.561 17:35:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:04.561 17:35:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:04.561 17:35:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:04.561 "name": "Existed_Raid", 00:14:04.561 "uuid": "0148c0f3-cacd-4c33-a619-29885e43851c", 00:14:04.561 "strip_size_kb": 64, 00:14:04.561 "state": "configuring", 00:14:04.561 "raid_level": "raid5f", 00:14:04.561 "superblock": true, 00:14:04.561 "num_base_bdevs": 3, 00:14:04.561 "num_base_bdevs_discovered": 2, 00:14:04.561 "num_base_bdevs_operational": 3, 00:14:04.561 "base_bdevs_list": [ 00:14:04.561 { 00:14:04.561 "name": "BaseBdev1", 00:14:04.561 "uuid": "08030365-66d3-4aae-92ce-01ca812a7a47", 00:14:04.561 "is_configured": true, 00:14:04.561 "data_offset": 2048, 00:14:04.561 "data_size": 63488 00:14:04.561 }, 00:14:04.561 { 00:14:04.561 "name": null, 00:14:04.561 "uuid": "62f68267-a872-41d1-adef-cf167e8b5f59", 00:14:04.561 "is_configured": false, 00:14:04.561 "data_offset": 0, 00:14:04.561 "data_size": 63488 00:14:04.561 }, 00:14:04.561 { 00:14:04.561 "name": "BaseBdev3", 00:14:04.561 "uuid": "da68fb38-9430-4f11-8cab-43e5ab9673bd", 00:14:04.561 "is_configured": true, 00:14:04.561 "data_offset": 2048, 00:14:04.561 "data_size": 63488 00:14:04.561 } 00:14:04.561 ] 00:14:04.561 }' 00:14:04.561 17:35:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:04.561 17:35:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:05.131 17:35:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:05.131 17:35:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:05.131 17:35:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:05.131 17:35:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:14:05.131 17:35:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:05.131 17:35:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:14:05.131 17:35:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:14:05.131 17:35:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:05.131 17:35:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:05.131 [2024-11-27 17:35:36.171004] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:14:05.131 17:35:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:05.131 17:35:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:14:05.131 17:35:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:05.131 17:35:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:05.131 17:35:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:05.131 17:35:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:05.131 17:35:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:05.131 17:35:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:05.131 17:35:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:05.131 17:35:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:05.131 17:35:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:05.131 17:35:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:05.132 17:35:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:05.132 17:35:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:05.132 17:35:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:05.132 17:35:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:05.132 17:35:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:05.132 "name": "Existed_Raid", 00:14:05.132 "uuid": "0148c0f3-cacd-4c33-a619-29885e43851c", 00:14:05.132 "strip_size_kb": 64, 00:14:05.132 "state": "configuring", 00:14:05.132 "raid_level": "raid5f", 00:14:05.132 "superblock": true, 00:14:05.132 "num_base_bdevs": 3, 00:14:05.132 "num_base_bdevs_discovered": 1, 00:14:05.132 "num_base_bdevs_operational": 3, 00:14:05.132 "base_bdevs_list": [ 00:14:05.132 { 00:14:05.132 "name": null, 00:14:05.132 "uuid": "08030365-66d3-4aae-92ce-01ca812a7a47", 00:14:05.132 "is_configured": false, 00:14:05.132 "data_offset": 0, 00:14:05.132 "data_size": 63488 00:14:05.132 }, 00:14:05.132 { 00:14:05.132 "name": null, 00:14:05.132 "uuid": "62f68267-a872-41d1-adef-cf167e8b5f59", 00:14:05.132 "is_configured": false, 00:14:05.132 "data_offset": 0, 00:14:05.132 "data_size": 63488 00:14:05.132 }, 00:14:05.132 { 00:14:05.132 "name": "BaseBdev3", 00:14:05.132 "uuid": "da68fb38-9430-4f11-8cab-43e5ab9673bd", 00:14:05.132 "is_configured": true, 00:14:05.132 "data_offset": 2048, 00:14:05.132 "data_size": 63488 00:14:05.132 } 00:14:05.132 ] 00:14:05.132 }' 00:14:05.132 17:35:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:05.132 17:35:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:05.701 17:35:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:05.701 17:35:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:05.701 17:35:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:05.701 17:35:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:14:05.701 17:35:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:05.701 17:35:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:14:05.701 17:35:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:14:05.701 17:35:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:05.701 17:35:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:05.701 [2024-11-27 17:35:36.708625] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:14:05.701 17:35:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:05.701 17:35:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 3 00:14:05.701 17:35:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:05.701 17:35:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:05.701 17:35:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:05.702 17:35:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:05.702 17:35:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:05.702 17:35:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:05.702 17:35:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:05.702 17:35:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:05.702 17:35:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:05.702 17:35:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:05.702 17:35:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:05.702 17:35:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:05.702 17:35:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:05.702 17:35:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:05.702 17:35:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:05.702 "name": "Existed_Raid", 00:14:05.702 "uuid": "0148c0f3-cacd-4c33-a619-29885e43851c", 00:14:05.702 "strip_size_kb": 64, 00:14:05.702 "state": "configuring", 00:14:05.702 "raid_level": "raid5f", 00:14:05.702 "superblock": true, 00:14:05.702 "num_base_bdevs": 3, 00:14:05.702 "num_base_bdevs_discovered": 2, 00:14:05.702 "num_base_bdevs_operational": 3, 00:14:05.702 "base_bdevs_list": [ 00:14:05.702 { 00:14:05.702 "name": null, 00:14:05.702 "uuid": "08030365-66d3-4aae-92ce-01ca812a7a47", 00:14:05.702 "is_configured": false, 00:14:05.702 "data_offset": 0, 00:14:05.702 "data_size": 63488 00:14:05.702 }, 00:14:05.702 { 00:14:05.702 "name": "BaseBdev2", 00:14:05.702 "uuid": "62f68267-a872-41d1-adef-cf167e8b5f59", 00:14:05.702 "is_configured": true, 00:14:05.702 "data_offset": 2048, 00:14:05.702 "data_size": 63488 00:14:05.702 }, 00:14:05.702 { 00:14:05.702 "name": "BaseBdev3", 00:14:05.702 "uuid": "da68fb38-9430-4f11-8cab-43e5ab9673bd", 00:14:05.702 "is_configured": true, 00:14:05.702 "data_offset": 2048, 00:14:05.702 "data_size": 63488 00:14:05.702 } 00:14:05.702 ] 00:14:05.702 }' 00:14:05.702 17:35:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:05.702 17:35:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:06.271 17:35:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:14:06.271 17:35:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:06.271 17:35:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:06.271 17:35:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:06.271 17:35:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:06.272 17:35:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:14:06.272 17:35:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:06.272 17:35:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:14:06.272 17:35:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:06.272 17:35:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:06.272 17:35:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:06.272 17:35:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 08030365-66d3-4aae-92ce-01ca812a7a47 00:14:06.272 17:35:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:06.272 17:35:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:06.272 [2024-11-27 17:35:37.273605] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:14:06.272 [2024-11-27 17:35:37.273762] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:14:06.272 [2024-11-27 17:35:37.273777] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:14:06.272 [2024-11-27 17:35:37.274004] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002870 00:14:06.272 NewBaseBdev 00:14:06.272 [2024-11-27 17:35:37.274381] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:14:06.272 [2024-11-27 17:35:37.274404] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001c80 00:14:06.272 [2024-11-27 17:35:37.274508] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:06.272 17:35:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:06.272 17:35:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:14:06.272 17:35:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=NewBaseBdev 00:14:06.272 17:35:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:14:06.272 17:35:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:14:06.272 17:35:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:14:06.272 17:35:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:14:06.272 17:35:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:14:06.272 17:35:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:06.272 17:35:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:06.272 17:35:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:06.272 17:35:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:14:06.272 17:35:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:06.272 17:35:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:06.272 [ 00:14:06.272 { 00:14:06.272 "name": "NewBaseBdev", 00:14:06.272 "aliases": [ 00:14:06.272 "08030365-66d3-4aae-92ce-01ca812a7a47" 00:14:06.272 ], 00:14:06.272 "product_name": "Malloc disk", 00:14:06.272 "block_size": 512, 00:14:06.272 "num_blocks": 65536, 00:14:06.272 "uuid": "08030365-66d3-4aae-92ce-01ca812a7a47", 00:14:06.272 "assigned_rate_limits": { 00:14:06.272 "rw_ios_per_sec": 0, 00:14:06.272 "rw_mbytes_per_sec": 0, 00:14:06.272 "r_mbytes_per_sec": 0, 00:14:06.272 "w_mbytes_per_sec": 0 00:14:06.272 }, 00:14:06.272 "claimed": true, 00:14:06.272 "claim_type": "exclusive_write", 00:14:06.272 "zoned": false, 00:14:06.272 "supported_io_types": { 00:14:06.272 "read": true, 00:14:06.272 "write": true, 00:14:06.272 "unmap": true, 00:14:06.272 "flush": true, 00:14:06.272 "reset": true, 00:14:06.272 "nvme_admin": false, 00:14:06.272 "nvme_io": false, 00:14:06.272 "nvme_io_md": false, 00:14:06.272 "write_zeroes": true, 00:14:06.272 "zcopy": true, 00:14:06.272 "get_zone_info": false, 00:14:06.272 "zone_management": false, 00:14:06.272 "zone_append": false, 00:14:06.272 "compare": false, 00:14:06.272 "compare_and_write": false, 00:14:06.272 "abort": true, 00:14:06.272 "seek_hole": false, 00:14:06.272 "seek_data": false, 00:14:06.272 "copy": true, 00:14:06.272 "nvme_iov_md": false 00:14:06.272 }, 00:14:06.272 "memory_domains": [ 00:14:06.272 { 00:14:06.272 "dma_device_id": "system", 00:14:06.272 "dma_device_type": 1 00:14:06.272 }, 00:14:06.272 { 00:14:06.272 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:06.272 "dma_device_type": 2 00:14:06.272 } 00:14:06.272 ], 00:14:06.272 "driver_specific": {} 00:14:06.272 } 00:14:06.272 ] 00:14:06.272 17:35:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:06.272 17:35:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:14:06.272 17:35:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 3 00:14:06.272 17:35:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:06.272 17:35:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:06.272 17:35:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:06.272 17:35:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:06.272 17:35:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:06.272 17:35:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:06.272 17:35:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:06.272 17:35:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:06.272 17:35:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:06.272 17:35:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:06.272 17:35:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:06.272 17:35:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:06.272 17:35:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:06.272 17:35:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:06.272 17:35:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:06.272 "name": "Existed_Raid", 00:14:06.272 "uuid": "0148c0f3-cacd-4c33-a619-29885e43851c", 00:14:06.272 "strip_size_kb": 64, 00:14:06.272 "state": "online", 00:14:06.272 "raid_level": "raid5f", 00:14:06.272 "superblock": true, 00:14:06.272 "num_base_bdevs": 3, 00:14:06.272 "num_base_bdevs_discovered": 3, 00:14:06.272 "num_base_bdevs_operational": 3, 00:14:06.272 "base_bdevs_list": [ 00:14:06.272 { 00:14:06.272 "name": "NewBaseBdev", 00:14:06.272 "uuid": "08030365-66d3-4aae-92ce-01ca812a7a47", 00:14:06.272 "is_configured": true, 00:14:06.272 "data_offset": 2048, 00:14:06.272 "data_size": 63488 00:14:06.272 }, 00:14:06.272 { 00:14:06.273 "name": "BaseBdev2", 00:14:06.273 "uuid": "62f68267-a872-41d1-adef-cf167e8b5f59", 00:14:06.273 "is_configured": true, 00:14:06.273 "data_offset": 2048, 00:14:06.273 "data_size": 63488 00:14:06.273 }, 00:14:06.273 { 00:14:06.273 "name": "BaseBdev3", 00:14:06.273 "uuid": "da68fb38-9430-4f11-8cab-43e5ab9673bd", 00:14:06.273 "is_configured": true, 00:14:06.273 "data_offset": 2048, 00:14:06.273 "data_size": 63488 00:14:06.273 } 00:14:06.273 ] 00:14:06.273 }' 00:14:06.273 17:35:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:06.273 17:35:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:06.842 17:35:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:14:06.842 17:35:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:14:06.842 17:35:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:14:06.842 17:35:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:14:06.842 17:35:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:14:06.842 17:35:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:14:06.842 17:35:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:14:06.842 17:35:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:14:06.842 17:35:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:06.842 17:35:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:06.842 [2024-11-27 17:35:37.756972] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:06.842 17:35:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:06.842 17:35:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:14:06.842 "name": "Existed_Raid", 00:14:06.842 "aliases": [ 00:14:06.842 "0148c0f3-cacd-4c33-a619-29885e43851c" 00:14:06.842 ], 00:14:06.842 "product_name": "Raid Volume", 00:14:06.842 "block_size": 512, 00:14:06.842 "num_blocks": 126976, 00:14:06.842 "uuid": "0148c0f3-cacd-4c33-a619-29885e43851c", 00:14:06.842 "assigned_rate_limits": { 00:14:06.842 "rw_ios_per_sec": 0, 00:14:06.842 "rw_mbytes_per_sec": 0, 00:14:06.842 "r_mbytes_per_sec": 0, 00:14:06.842 "w_mbytes_per_sec": 0 00:14:06.842 }, 00:14:06.842 "claimed": false, 00:14:06.842 "zoned": false, 00:14:06.842 "supported_io_types": { 00:14:06.842 "read": true, 00:14:06.842 "write": true, 00:14:06.842 "unmap": false, 00:14:06.842 "flush": false, 00:14:06.842 "reset": true, 00:14:06.842 "nvme_admin": false, 00:14:06.842 "nvme_io": false, 00:14:06.842 "nvme_io_md": false, 00:14:06.842 "write_zeroes": true, 00:14:06.842 "zcopy": false, 00:14:06.842 "get_zone_info": false, 00:14:06.842 "zone_management": false, 00:14:06.842 "zone_append": false, 00:14:06.842 "compare": false, 00:14:06.842 "compare_and_write": false, 00:14:06.842 "abort": false, 00:14:06.842 "seek_hole": false, 00:14:06.842 "seek_data": false, 00:14:06.842 "copy": false, 00:14:06.842 "nvme_iov_md": false 00:14:06.842 }, 00:14:06.842 "driver_specific": { 00:14:06.842 "raid": { 00:14:06.842 "uuid": "0148c0f3-cacd-4c33-a619-29885e43851c", 00:14:06.842 "strip_size_kb": 64, 00:14:06.842 "state": "online", 00:14:06.842 "raid_level": "raid5f", 00:14:06.842 "superblock": true, 00:14:06.842 "num_base_bdevs": 3, 00:14:06.842 "num_base_bdevs_discovered": 3, 00:14:06.842 "num_base_bdevs_operational": 3, 00:14:06.842 "base_bdevs_list": [ 00:14:06.842 { 00:14:06.842 "name": "NewBaseBdev", 00:14:06.842 "uuid": "08030365-66d3-4aae-92ce-01ca812a7a47", 00:14:06.842 "is_configured": true, 00:14:06.842 "data_offset": 2048, 00:14:06.842 "data_size": 63488 00:14:06.842 }, 00:14:06.842 { 00:14:06.842 "name": "BaseBdev2", 00:14:06.842 "uuid": "62f68267-a872-41d1-adef-cf167e8b5f59", 00:14:06.842 "is_configured": true, 00:14:06.842 "data_offset": 2048, 00:14:06.842 "data_size": 63488 00:14:06.842 }, 00:14:06.842 { 00:14:06.842 "name": "BaseBdev3", 00:14:06.842 "uuid": "da68fb38-9430-4f11-8cab-43e5ab9673bd", 00:14:06.842 "is_configured": true, 00:14:06.842 "data_offset": 2048, 00:14:06.842 "data_size": 63488 00:14:06.842 } 00:14:06.842 ] 00:14:06.842 } 00:14:06.842 } 00:14:06.842 }' 00:14:06.842 17:35:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:14:06.842 17:35:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:14:06.842 BaseBdev2 00:14:06.842 BaseBdev3' 00:14:06.842 17:35:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:06.842 17:35:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:14:06.842 17:35:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:06.842 17:35:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:14:06.842 17:35:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:06.842 17:35:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:06.842 17:35:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:06.842 17:35:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:06.842 17:35:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:06.842 17:35:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:06.842 17:35:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:06.842 17:35:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:14:06.842 17:35:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:06.842 17:35:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:06.842 17:35:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:06.842 17:35:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:06.842 17:35:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:06.842 17:35:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:06.842 17:35:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:06.842 17:35:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:06.842 17:35:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:14:06.842 17:35:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:06.842 17:35:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:06.842 17:35:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:06.842 17:35:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:06.842 17:35:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:06.842 17:35:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:14:06.842 17:35:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:06.842 17:35:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:06.842 [2024-11-27 17:35:38.000360] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:14:06.842 [2024-11-27 17:35:38.000382] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:14:06.842 [2024-11-27 17:35:38.000435] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:06.842 [2024-11-27 17:35:38.000657] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:14:06.842 [2024-11-27 17:35:38.000670] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name Existed_Raid, state offline 00:14:06.842 17:35:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:06.842 17:35:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 90910 00:14:06.842 17:35:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@950 -- # '[' -z 90910 ']' 00:14:06.842 17:35:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@954 -- # kill -0 90910 00:14:06.842 17:35:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@955 -- # uname 00:14:06.842 17:35:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:14:06.842 17:35:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 90910 00:14:07.102 17:35:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:14:07.102 17:35:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:14:07.102 killing process with pid 90910 00:14:07.102 17:35:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@968 -- # echo 'killing process with pid 90910' 00:14:07.102 17:35:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@969 -- # kill 90910 00:14:07.102 [2024-11-27 17:35:38.050718] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:14:07.102 17:35:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@974 -- # wait 90910 00:14:07.102 [2024-11-27 17:35:38.081998] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:14:07.362 17:35:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:14:07.362 00:14:07.362 real 0m8.907s 00:14:07.362 user 0m15.142s 00:14:07.362 sys 0m1.941s 00:14:07.362 17:35:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@1126 -- # xtrace_disable 00:14:07.362 17:35:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:07.362 ************************************ 00:14:07.362 END TEST raid5f_state_function_test_sb 00:14:07.362 ************************************ 00:14:07.362 17:35:38 bdev_raid -- bdev/bdev_raid.sh@988 -- # run_test raid5f_superblock_test raid_superblock_test raid5f 3 00:14:07.362 17:35:38 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:14:07.362 17:35:38 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:14:07.362 17:35:38 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:14:07.362 ************************************ 00:14:07.362 START TEST raid5f_superblock_test 00:14:07.362 ************************************ 00:14:07.362 17:35:38 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@1125 -- # raid_superblock_test raid5f 3 00:14:07.362 17:35:38 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=raid5f 00:14:07.362 17:35:38 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=3 00:14:07.362 17:35:38 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:14:07.362 17:35:38 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:14:07.362 17:35:38 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:14:07.362 17:35:38 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:14:07.362 17:35:38 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:14:07.362 17:35:38 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:14:07.362 17:35:38 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:14:07.362 17:35:38 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:14:07.362 17:35:38 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:14:07.362 17:35:38 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:14:07.362 17:35:38 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:14:07.362 17:35:38 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' raid5f '!=' raid1 ']' 00:14:07.362 17:35:38 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@405 -- # strip_size=64 00:14:07.362 17:35:38 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@406 -- # strip_size_create_arg='-z 64' 00:14:07.362 17:35:38 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=91514 00:14:07.362 17:35:38 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:14:07.362 17:35:38 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 91514 00:14:07.362 17:35:38 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@831 -- # '[' -z 91514 ']' 00:14:07.362 17:35:38 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:14:07.362 17:35:38 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:14:07.362 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:14:07.362 17:35:38 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:14:07.362 17:35:38 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:14:07.362 17:35:38 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:07.362 [2024-11-27 17:35:38.508733] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:14:07.362 [2024-11-27 17:35:38.508899] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid91514 ] 00:14:07.623 [2024-11-27 17:35:38.653879] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:14:07.623 [2024-11-27 17:35:38.697627] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:14:07.623 [2024-11-27 17:35:38.739116] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:14:07.623 [2024-11-27 17:35:38.739160] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:14:08.202 17:35:39 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:14:08.202 17:35:39 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@864 -- # return 0 00:14:08.202 17:35:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:14:08.202 17:35:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:14:08.202 17:35:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:14:08.202 17:35:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:14:08.202 17:35:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:14:08.202 17:35:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:14:08.202 17:35:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:14:08.202 17:35:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:14:08.202 17:35:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:14:08.202 17:35:39 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:08.202 17:35:39 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:08.202 malloc1 00:14:08.202 17:35:39 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:08.202 17:35:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:14:08.202 17:35:39 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:08.202 17:35:39 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:08.202 [2024-11-27 17:35:39.341234] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:14:08.202 [2024-11-27 17:35:39.341341] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:08.202 [2024-11-27 17:35:39.341380] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:14:08.202 [2024-11-27 17:35:39.341416] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:08.202 [2024-11-27 17:35:39.343395] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:08.202 [2024-11-27 17:35:39.343468] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:14:08.202 pt1 00:14:08.202 17:35:39 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:08.202 17:35:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:14:08.202 17:35:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:14:08.202 17:35:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:14:08.202 17:35:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:14:08.202 17:35:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:14:08.202 17:35:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:14:08.202 17:35:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:14:08.202 17:35:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:14:08.202 17:35:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:14:08.202 17:35:39 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:08.202 17:35:39 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:08.202 malloc2 00:14:08.202 17:35:39 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:08.202 17:35:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:14:08.202 17:35:39 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:08.202 17:35:39 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:08.202 [2024-11-27 17:35:39.389218] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:14:08.202 [2024-11-27 17:35:39.389396] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:08.202 [2024-11-27 17:35:39.389437] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:14:08.202 [2024-11-27 17:35:39.389462] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:08.464 [2024-11-27 17:35:39.393909] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:08.464 [2024-11-27 17:35:39.393982] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:14:08.464 pt2 00:14:08.464 17:35:39 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:08.464 17:35:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:14:08.464 17:35:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:14:08.464 17:35:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc3 00:14:08.464 17:35:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt3 00:14:08.464 17:35:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000003 00:14:08.464 17:35:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:14:08.464 17:35:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:14:08.464 17:35:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:14:08.464 17:35:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc3 00:14:08.464 17:35:39 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:08.464 17:35:39 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:08.464 malloc3 00:14:08.464 17:35:39 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:08.464 17:35:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:14:08.464 17:35:39 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:08.464 17:35:39 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:08.464 [2024-11-27 17:35:39.423616] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:14:08.464 [2024-11-27 17:35:39.423707] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:08.464 [2024-11-27 17:35:39.423737] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:14:08.464 [2024-11-27 17:35:39.423767] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:08.464 [2024-11-27 17:35:39.425699] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:08.464 [2024-11-27 17:35:39.425772] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:14:08.464 pt3 00:14:08.464 17:35:39 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:08.464 17:35:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:14:08.464 17:35:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:14:08.464 17:35:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''pt1 pt2 pt3'\''' -n raid_bdev1 -s 00:14:08.464 17:35:39 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:08.464 17:35:39 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:08.464 [2024-11-27 17:35:39.435680] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:14:08.464 [2024-11-27 17:35:39.437442] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:14:08.464 [2024-11-27 17:35:39.437529] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:14:08.464 [2024-11-27 17:35:39.437712] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:14:08.464 [2024-11-27 17:35:39.437764] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:14:08.464 [2024-11-27 17:35:39.438023] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:14:08.464 [2024-11-27 17:35:39.438461] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:14:08.464 [2024-11-27 17:35:39.438511] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:14:08.464 [2024-11-27 17:35:39.438659] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:08.464 17:35:39 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:08.464 17:35:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:14:08.464 17:35:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:08.465 17:35:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:08.465 17:35:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:08.465 17:35:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:08.465 17:35:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:08.465 17:35:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:08.465 17:35:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:08.465 17:35:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:08.465 17:35:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:08.465 17:35:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:08.465 17:35:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:08.465 17:35:39 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:08.465 17:35:39 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:08.465 17:35:39 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:08.465 17:35:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:08.465 "name": "raid_bdev1", 00:14:08.465 "uuid": "042f6175-b4c1-4ffc-8cb6-5e77f974aeb4", 00:14:08.465 "strip_size_kb": 64, 00:14:08.465 "state": "online", 00:14:08.465 "raid_level": "raid5f", 00:14:08.465 "superblock": true, 00:14:08.465 "num_base_bdevs": 3, 00:14:08.465 "num_base_bdevs_discovered": 3, 00:14:08.465 "num_base_bdevs_operational": 3, 00:14:08.465 "base_bdevs_list": [ 00:14:08.465 { 00:14:08.465 "name": "pt1", 00:14:08.465 "uuid": "00000000-0000-0000-0000-000000000001", 00:14:08.465 "is_configured": true, 00:14:08.465 "data_offset": 2048, 00:14:08.465 "data_size": 63488 00:14:08.465 }, 00:14:08.465 { 00:14:08.465 "name": "pt2", 00:14:08.465 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:08.465 "is_configured": true, 00:14:08.465 "data_offset": 2048, 00:14:08.465 "data_size": 63488 00:14:08.465 }, 00:14:08.465 { 00:14:08.465 "name": "pt3", 00:14:08.465 "uuid": "00000000-0000-0000-0000-000000000003", 00:14:08.465 "is_configured": true, 00:14:08.465 "data_offset": 2048, 00:14:08.465 "data_size": 63488 00:14:08.465 } 00:14:08.465 ] 00:14:08.465 }' 00:14:08.465 17:35:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:08.465 17:35:39 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:09.035 17:35:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:14:09.035 17:35:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:14:09.035 17:35:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:14:09.035 17:35:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:14:09.035 17:35:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:14:09.035 17:35:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:14:09.035 17:35:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:14:09.035 17:35:39 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:09.035 17:35:39 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:09.035 17:35:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:14:09.035 [2024-11-27 17:35:39.931413] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:09.035 17:35:39 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:09.035 17:35:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:14:09.035 "name": "raid_bdev1", 00:14:09.035 "aliases": [ 00:14:09.035 "042f6175-b4c1-4ffc-8cb6-5e77f974aeb4" 00:14:09.035 ], 00:14:09.035 "product_name": "Raid Volume", 00:14:09.035 "block_size": 512, 00:14:09.035 "num_blocks": 126976, 00:14:09.035 "uuid": "042f6175-b4c1-4ffc-8cb6-5e77f974aeb4", 00:14:09.035 "assigned_rate_limits": { 00:14:09.035 "rw_ios_per_sec": 0, 00:14:09.035 "rw_mbytes_per_sec": 0, 00:14:09.035 "r_mbytes_per_sec": 0, 00:14:09.035 "w_mbytes_per_sec": 0 00:14:09.035 }, 00:14:09.035 "claimed": false, 00:14:09.035 "zoned": false, 00:14:09.035 "supported_io_types": { 00:14:09.035 "read": true, 00:14:09.035 "write": true, 00:14:09.035 "unmap": false, 00:14:09.035 "flush": false, 00:14:09.035 "reset": true, 00:14:09.035 "nvme_admin": false, 00:14:09.035 "nvme_io": false, 00:14:09.035 "nvme_io_md": false, 00:14:09.035 "write_zeroes": true, 00:14:09.035 "zcopy": false, 00:14:09.035 "get_zone_info": false, 00:14:09.035 "zone_management": false, 00:14:09.035 "zone_append": false, 00:14:09.035 "compare": false, 00:14:09.035 "compare_and_write": false, 00:14:09.035 "abort": false, 00:14:09.035 "seek_hole": false, 00:14:09.035 "seek_data": false, 00:14:09.035 "copy": false, 00:14:09.035 "nvme_iov_md": false 00:14:09.035 }, 00:14:09.035 "driver_specific": { 00:14:09.035 "raid": { 00:14:09.035 "uuid": "042f6175-b4c1-4ffc-8cb6-5e77f974aeb4", 00:14:09.035 "strip_size_kb": 64, 00:14:09.035 "state": "online", 00:14:09.035 "raid_level": "raid5f", 00:14:09.035 "superblock": true, 00:14:09.035 "num_base_bdevs": 3, 00:14:09.035 "num_base_bdevs_discovered": 3, 00:14:09.035 "num_base_bdevs_operational": 3, 00:14:09.035 "base_bdevs_list": [ 00:14:09.035 { 00:14:09.035 "name": "pt1", 00:14:09.035 "uuid": "00000000-0000-0000-0000-000000000001", 00:14:09.035 "is_configured": true, 00:14:09.035 "data_offset": 2048, 00:14:09.035 "data_size": 63488 00:14:09.035 }, 00:14:09.035 { 00:14:09.035 "name": "pt2", 00:14:09.035 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:09.035 "is_configured": true, 00:14:09.035 "data_offset": 2048, 00:14:09.035 "data_size": 63488 00:14:09.035 }, 00:14:09.035 { 00:14:09.035 "name": "pt3", 00:14:09.035 "uuid": "00000000-0000-0000-0000-000000000003", 00:14:09.035 "is_configured": true, 00:14:09.035 "data_offset": 2048, 00:14:09.035 "data_size": 63488 00:14:09.035 } 00:14:09.035 ] 00:14:09.035 } 00:14:09.035 } 00:14:09.035 }' 00:14:09.035 17:35:39 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:14:09.035 17:35:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:14:09.035 pt2 00:14:09.035 pt3' 00:14:09.035 17:35:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:09.035 17:35:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:14:09.035 17:35:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:09.035 17:35:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:14:09.035 17:35:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:09.035 17:35:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:09.035 17:35:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:09.035 17:35:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:09.035 17:35:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:09.035 17:35:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:09.035 17:35:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:09.035 17:35:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:14:09.035 17:35:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:09.035 17:35:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:09.035 17:35:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:09.035 17:35:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:09.035 17:35:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:09.035 17:35:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:09.035 17:35:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:09.035 17:35:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:14:09.035 17:35:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:09.035 17:35:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:09.035 17:35:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:09.035 17:35:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:09.295 17:35:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:09.295 17:35:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:09.295 17:35:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:14:09.295 17:35:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:09.295 17:35:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:09.295 17:35:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:14:09.295 [2024-11-27 17:35:40.235330] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:09.295 17:35:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:09.295 17:35:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=042f6175-b4c1-4ffc-8cb6-5e77f974aeb4 00:14:09.295 17:35:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z 042f6175-b4c1-4ffc-8cb6-5e77f974aeb4 ']' 00:14:09.295 17:35:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:14:09.295 17:35:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:09.295 17:35:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:09.295 [2024-11-27 17:35:40.283085] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:14:09.295 [2024-11-27 17:35:40.283106] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:14:09.295 [2024-11-27 17:35:40.283178] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:09.295 [2024-11-27 17:35:40.283238] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:14:09.295 [2024-11-27 17:35:40.283249] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:14:09.295 17:35:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:09.295 17:35:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:09.295 17:35:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:14:09.295 17:35:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:09.295 17:35:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:09.295 17:35:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:09.295 17:35:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:14:09.295 17:35:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:14:09.295 17:35:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:14:09.295 17:35:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:14:09.295 17:35:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:09.295 17:35:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:09.295 17:35:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:09.295 17:35:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:14:09.295 17:35:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:14:09.295 17:35:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:09.295 17:35:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:09.295 17:35:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:09.295 17:35:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:14:09.295 17:35:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt3 00:14:09.295 17:35:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:09.295 17:35:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:09.295 17:35:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:09.295 17:35:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:14:09.295 17:35:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:14:09.295 17:35:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:09.295 17:35:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:09.295 17:35:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:09.295 17:35:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:14:09.295 17:35:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:14:09.295 17:35:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@650 -- # local es=0 00:14:09.295 17:35:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:14:09.295 17:35:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:14:09.295 17:35:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:14:09.295 17:35:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:14:09.295 17:35:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:14:09.295 17:35:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''malloc1 malloc2 malloc3'\''' -n raid_bdev1 00:14:09.295 17:35:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:09.295 17:35:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:09.295 [2024-11-27 17:35:40.439018] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:14:09.295 [2024-11-27 17:35:40.440899] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:14:09.295 [2024-11-27 17:35:40.440942] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc3 is claimed 00:14:09.295 [2024-11-27 17:35:40.440982] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:14:09.295 [2024-11-27 17:35:40.441018] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:14:09.295 [2024-11-27 17:35:40.441036] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc3 00:14:09.295 [2024-11-27 17:35:40.441049] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:14:09.295 [2024-11-27 17:35:40.441060] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state configuring 00:14:09.295 request: 00:14:09.295 { 00:14:09.295 "name": "raid_bdev1", 00:14:09.295 "raid_level": "raid5f", 00:14:09.295 "base_bdevs": [ 00:14:09.295 "malloc1", 00:14:09.295 "malloc2", 00:14:09.295 "malloc3" 00:14:09.295 ], 00:14:09.295 "strip_size_kb": 64, 00:14:09.295 "superblock": false, 00:14:09.295 "method": "bdev_raid_create", 00:14:09.295 "req_id": 1 00:14:09.295 } 00:14:09.295 Got JSON-RPC error response 00:14:09.295 response: 00:14:09.295 { 00:14:09.295 "code": -17, 00:14:09.295 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:14:09.296 } 00:14:09.296 17:35:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:14:09.296 17:35:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@653 -- # es=1 00:14:09.296 17:35:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:14:09.296 17:35:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:14:09.296 17:35:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:14:09.296 17:35:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:09.296 17:35:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:14:09.296 17:35:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:09.296 17:35:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:09.296 17:35:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:09.555 17:35:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:14:09.555 17:35:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:14:09.555 17:35:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:14:09.555 17:35:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:09.555 17:35:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:09.555 [2024-11-27 17:35:40.506988] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:14:09.555 [2024-11-27 17:35:40.507074] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:09.555 [2024-11-27 17:35:40.507102] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:14:09.555 [2024-11-27 17:35:40.507135] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:09.555 [2024-11-27 17:35:40.509156] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:09.555 [2024-11-27 17:35:40.509225] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:14:09.555 [2024-11-27 17:35:40.509303] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:14:09.555 [2024-11-27 17:35:40.509359] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:14:09.555 pt1 00:14:09.555 17:35:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:09.555 17:35:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 3 00:14:09.555 17:35:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:09.555 17:35:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:09.555 17:35:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:09.555 17:35:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:09.555 17:35:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:09.555 17:35:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:09.555 17:35:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:09.555 17:35:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:09.555 17:35:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:09.555 17:35:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:09.555 17:35:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:09.555 17:35:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:09.555 17:35:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:09.555 17:35:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:09.555 17:35:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:09.555 "name": "raid_bdev1", 00:14:09.555 "uuid": "042f6175-b4c1-4ffc-8cb6-5e77f974aeb4", 00:14:09.555 "strip_size_kb": 64, 00:14:09.555 "state": "configuring", 00:14:09.555 "raid_level": "raid5f", 00:14:09.555 "superblock": true, 00:14:09.555 "num_base_bdevs": 3, 00:14:09.555 "num_base_bdevs_discovered": 1, 00:14:09.555 "num_base_bdevs_operational": 3, 00:14:09.555 "base_bdevs_list": [ 00:14:09.555 { 00:14:09.555 "name": "pt1", 00:14:09.555 "uuid": "00000000-0000-0000-0000-000000000001", 00:14:09.555 "is_configured": true, 00:14:09.555 "data_offset": 2048, 00:14:09.555 "data_size": 63488 00:14:09.555 }, 00:14:09.555 { 00:14:09.555 "name": null, 00:14:09.555 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:09.555 "is_configured": false, 00:14:09.555 "data_offset": 2048, 00:14:09.555 "data_size": 63488 00:14:09.555 }, 00:14:09.555 { 00:14:09.555 "name": null, 00:14:09.555 "uuid": "00000000-0000-0000-0000-000000000003", 00:14:09.555 "is_configured": false, 00:14:09.555 "data_offset": 2048, 00:14:09.555 "data_size": 63488 00:14:09.555 } 00:14:09.555 ] 00:14:09.555 }' 00:14:09.555 17:35:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:09.555 17:35:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:09.815 17:35:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 3 -gt 2 ']' 00:14:09.815 17:35:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@472 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:14:09.815 17:35:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:09.815 17:35:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:09.815 [2024-11-27 17:35:40.978233] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:14:09.815 [2024-11-27 17:35:40.978287] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:09.815 [2024-11-27 17:35:40.978304] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:14:09.815 [2024-11-27 17:35:40.978315] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:09.815 [2024-11-27 17:35:40.978616] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:09.815 [2024-11-27 17:35:40.978634] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:14:09.815 [2024-11-27 17:35:40.978683] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:14:09.815 [2024-11-27 17:35:40.978702] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:14:09.815 pt2 00:14:09.815 17:35:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:09.815 17:35:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@473 -- # rpc_cmd bdev_passthru_delete pt2 00:14:09.815 17:35:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:09.815 17:35:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:09.815 [2024-11-27 17:35:40.990214] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: pt2 00:14:09.815 17:35:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:09.815 17:35:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@474 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 3 00:14:09.815 17:35:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:09.815 17:35:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:09.815 17:35:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:09.815 17:35:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:09.815 17:35:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:09.816 17:35:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:09.816 17:35:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:09.816 17:35:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:09.816 17:35:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:09.816 17:35:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:09.816 17:35:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:09.816 17:35:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:09.816 17:35:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:10.075 17:35:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:10.075 17:35:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:10.075 "name": "raid_bdev1", 00:14:10.075 "uuid": "042f6175-b4c1-4ffc-8cb6-5e77f974aeb4", 00:14:10.075 "strip_size_kb": 64, 00:14:10.075 "state": "configuring", 00:14:10.075 "raid_level": "raid5f", 00:14:10.075 "superblock": true, 00:14:10.075 "num_base_bdevs": 3, 00:14:10.075 "num_base_bdevs_discovered": 1, 00:14:10.075 "num_base_bdevs_operational": 3, 00:14:10.075 "base_bdevs_list": [ 00:14:10.075 { 00:14:10.075 "name": "pt1", 00:14:10.075 "uuid": "00000000-0000-0000-0000-000000000001", 00:14:10.075 "is_configured": true, 00:14:10.075 "data_offset": 2048, 00:14:10.075 "data_size": 63488 00:14:10.075 }, 00:14:10.075 { 00:14:10.075 "name": null, 00:14:10.075 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:10.075 "is_configured": false, 00:14:10.075 "data_offset": 0, 00:14:10.075 "data_size": 63488 00:14:10.075 }, 00:14:10.075 { 00:14:10.075 "name": null, 00:14:10.075 "uuid": "00000000-0000-0000-0000-000000000003", 00:14:10.075 "is_configured": false, 00:14:10.075 "data_offset": 2048, 00:14:10.075 "data_size": 63488 00:14:10.075 } 00:14:10.075 ] 00:14:10.075 }' 00:14:10.075 17:35:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:10.075 17:35:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:10.336 17:35:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:14:10.336 17:35:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:14:10.336 17:35:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:14:10.336 17:35:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:10.336 17:35:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:10.336 [2024-11-27 17:35:41.417440] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:14:10.336 [2024-11-27 17:35:41.417522] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:10.336 [2024-11-27 17:35:41.417550] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:14:10.336 [2024-11-27 17:35:41.417576] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:10.336 [2024-11-27 17:35:41.417857] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:10.336 [2024-11-27 17:35:41.417915] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:14:10.336 [2024-11-27 17:35:41.417988] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:14:10.336 [2024-11-27 17:35:41.418028] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:14:10.336 pt2 00:14:10.336 17:35:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:10.336 17:35:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:14:10.336 17:35:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:14:10.336 17:35:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:14:10.336 17:35:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:10.336 17:35:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:10.336 [2024-11-27 17:35:41.429430] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:14:10.336 [2024-11-27 17:35:41.429504] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:10.336 [2024-11-27 17:35:41.429535] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:14:10.336 [2024-11-27 17:35:41.429561] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:10.336 [2024-11-27 17:35:41.429846] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:10.336 [2024-11-27 17:35:41.429898] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:14:10.336 [2024-11-27 17:35:41.429969] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:14:10.336 [2024-11-27 17:35:41.430026] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:14:10.336 [2024-11-27 17:35:41.430136] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:14:10.336 [2024-11-27 17:35:41.430189] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:14:10.336 [2024-11-27 17:35:41.430401] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:14:10.336 [2024-11-27 17:35:41.430776] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:14:10.336 [2024-11-27 17:35:41.430826] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:14:10.336 [2024-11-27 17:35:41.430952] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:10.336 pt3 00:14:10.336 17:35:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:10.336 17:35:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:14:10.336 17:35:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:14:10.336 17:35:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:14:10.336 17:35:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:10.336 17:35:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:10.336 17:35:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:10.336 17:35:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:10.336 17:35:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:10.336 17:35:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:10.336 17:35:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:10.336 17:35:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:10.336 17:35:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:10.336 17:35:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:10.336 17:35:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:10.336 17:35:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:10.336 17:35:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:10.336 17:35:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:10.336 17:35:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:10.336 "name": "raid_bdev1", 00:14:10.336 "uuid": "042f6175-b4c1-4ffc-8cb6-5e77f974aeb4", 00:14:10.336 "strip_size_kb": 64, 00:14:10.336 "state": "online", 00:14:10.336 "raid_level": "raid5f", 00:14:10.336 "superblock": true, 00:14:10.336 "num_base_bdevs": 3, 00:14:10.336 "num_base_bdevs_discovered": 3, 00:14:10.336 "num_base_bdevs_operational": 3, 00:14:10.336 "base_bdevs_list": [ 00:14:10.336 { 00:14:10.336 "name": "pt1", 00:14:10.336 "uuid": "00000000-0000-0000-0000-000000000001", 00:14:10.336 "is_configured": true, 00:14:10.336 "data_offset": 2048, 00:14:10.336 "data_size": 63488 00:14:10.336 }, 00:14:10.336 { 00:14:10.336 "name": "pt2", 00:14:10.336 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:10.336 "is_configured": true, 00:14:10.336 "data_offset": 2048, 00:14:10.336 "data_size": 63488 00:14:10.336 }, 00:14:10.336 { 00:14:10.336 "name": "pt3", 00:14:10.336 "uuid": "00000000-0000-0000-0000-000000000003", 00:14:10.336 "is_configured": true, 00:14:10.336 "data_offset": 2048, 00:14:10.336 "data_size": 63488 00:14:10.336 } 00:14:10.336 ] 00:14:10.336 }' 00:14:10.336 17:35:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:10.336 17:35:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:10.906 17:35:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:14:10.906 17:35:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:14:10.906 17:35:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:14:10.906 17:35:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:14:10.906 17:35:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:14:10.906 17:35:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:14:10.906 17:35:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:14:10.906 17:35:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:14:10.906 17:35:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:10.906 17:35:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:10.906 [2024-11-27 17:35:41.848940] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:10.906 17:35:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:10.906 17:35:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:14:10.906 "name": "raid_bdev1", 00:14:10.906 "aliases": [ 00:14:10.906 "042f6175-b4c1-4ffc-8cb6-5e77f974aeb4" 00:14:10.906 ], 00:14:10.906 "product_name": "Raid Volume", 00:14:10.906 "block_size": 512, 00:14:10.906 "num_blocks": 126976, 00:14:10.906 "uuid": "042f6175-b4c1-4ffc-8cb6-5e77f974aeb4", 00:14:10.906 "assigned_rate_limits": { 00:14:10.906 "rw_ios_per_sec": 0, 00:14:10.906 "rw_mbytes_per_sec": 0, 00:14:10.906 "r_mbytes_per_sec": 0, 00:14:10.906 "w_mbytes_per_sec": 0 00:14:10.906 }, 00:14:10.906 "claimed": false, 00:14:10.906 "zoned": false, 00:14:10.906 "supported_io_types": { 00:14:10.906 "read": true, 00:14:10.906 "write": true, 00:14:10.906 "unmap": false, 00:14:10.907 "flush": false, 00:14:10.907 "reset": true, 00:14:10.907 "nvme_admin": false, 00:14:10.907 "nvme_io": false, 00:14:10.907 "nvme_io_md": false, 00:14:10.907 "write_zeroes": true, 00:14:10.907 "zcopy": false, 00:14:10.907 "get_zone_info": false, 00:14:10.907 "zone_management": false, 00:14:10.907 "zone_append": false, 00:14:10.907 "compare": false, 00:14:10.907 "compare_and_write": false, 00:14:10.907 "abort": false, 00:14:10.907 "seek_hole": false, 00:14:10.907 "seek_data": false, 00:14:10.907 "copy": false, 00:14:10.907 "nvme_iov_md": false 00:14:10.907 }, 00:14:10.907 "driver_specific": { 00:14:10.907 "raid": { 00:14:10.907 "uuid": "042f6175-b4c1-4ffc-8cb6-5e77f974aeb4", 00:14:10.907 "strip_size_kb": 64, 00:14:10.907 "state": "online", 00:14:10.907 "raid_level": "raid5f", 00:14:10.907 "superblock": true, 00:14:10.907 "num_base_bdevs": 3, 00:14:10.907 "num_base_bdevs_discovered": 3, 00:14:10.907 "num_base_bdevs_operational": 3, 00:14:10.907 "base_bdevs_list": [ 00:14:10.907 { 00:14:10.907 "name": "pt1", 00:14:10.907 "uuid": "00000000-0000-0000-0000-000000000001", 00:14:10.907 "is_configured": true, 00:14:10.907 "data_offset": 2048, 00:14:10.907 "data_size": 63488 00:14:10.907 }, 00:14:10.907 { 00:14:10.907 "name": "pt2", 00:14:10.907 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:10.907 "is_configured": true, 00:14:10.907 "data_offset": 2048, 00:14:10.907 "data_size": 63488 00:14:10.907 }, 00:14:10.907 { 00:14:10.907 "name": "pt3", 00:14:10.907 "uuid": "00000000-0000-0000-0000-000000000003", 00:14:10.907 "is_configured": true, 00:14:10.907 "data_offset": 2048, 00:14:10.907 "data_size": 63488 00:14:10.907 } 00:14:10.907 ] 00:14:10.907 } 00:14:10.907 } 00:14:10.907 }' 00:14:10.907 17:35:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:14:10.907 17:35:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:14:10.907 pt2 00:14:10.907 pt3' 00:14:10.907 17:35:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:10.907 17:35:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:14:10.907 17:35:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:10.907 17:35:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:14:10.907 17:35:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:10.907 17:35:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:10.907 17:35:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:10.907 17:35:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:10.907 17:35:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:10.907 17:35:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:10.907 17:35:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:10.907 17:35:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:14:10.907 17:35:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:10.907 17:35:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:10.907 17:35:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:10.907 17:35:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:10.907 17:35:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:10.907 17:35:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:10.907 17:35:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:10.907 17:35:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:14:10.907 17:35:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:10.907 17:35:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:10.907 17:35:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:10.907 17:35:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:11.167 17:35:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:11.167 17:35:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:11.167 17:35:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:14:11.167 17:35:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:14:11.167 17:35:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:11.167 17:35:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:11.167 [2024-11-27 17:35:42.116433] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:11.167 17:35:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:11.167 17:35:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' 042f6175-b4c1-4ffc-8cb6-5e77f974aeb4 '!=' 042f6175-b4c1-4ffc-8cb6-5e77f974aeb4 ']' 00:14:11.167 17:35:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy raid5f 00:14:11.167 17:35:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:14:11.167 17:35:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@199 -- # return 0 00:14:11.167 17:35:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@493 -- # rpc_cmd bdev_passthru_delete pt1 00:14:11.167 17:35:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:11.167 17:35:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:11.167 [2024-11-27 17:35:42.156261] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: pt1 00:14:11.167 17:35:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:11.167 17:35:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@496 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:14:11.167 17:35:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:11.167 17:35:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:11.167 17:35:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:11.167 17:35:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:11.167 17:35:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:14:11.167 17:35:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:11.167 17:35:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:11.167 17:35:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:11.167 17:35:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:11.167 17:35:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:11.167 17:35:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:11.167 17:35:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:11.167 17:35:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:11.167 17:35:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:11.167 17:35:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:11.167 "name": "raid_bdev1", 00:14:11.167 "uuid": "042f6175-b4c1-4ffc-8cb6-5e77f974aeb4", 00:14:11.167 "strip_size_kb": 64, 00:14:11.167 "state": "online", 00:14:11.167 "raid_level": "raid5f", 00:14:11.167 "superblock": true, 00:14:11.167 "num_base_bdevs": 3, 00:14:11.167 "num_base_bdevs_discovered": 2, 00:14:11.167 "num_base_bdevs_operational": 2, 00:14:11.167 "base_bdevs_list": [ 00:14:11.167 { 00:14:11.167 "name": null, 00:14:11.167 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:11.167 "is_configured": false, 00:14:11.167 "data_offset": 0, 00:14:11.167 "data_size": 63488 00:14:11.167 }, 00:14:11.167 { 00:14:11.167 "name": "pt2", 00:14:11.167 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:11.167 "is_configured": true, 00:14:11.167 "data_offset": 2048, 00:14:11.167 "data_size": 63488 00:14:11.167 }, 00:14:11.167 { 00:14:11.167 "name": "pt3", 00:14:11.167 "uuid": "00000000-0000-0000-0000-000000000003", 00:14:11.167 "is_configured": true, 00:14:11.167 "data_offset": 2048, 00:14:11.167 "data_size": 63488 00:14:11.167 } 00:14:11.167 ] 00:14:11.167 }' 00:14:11.167 17:35:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:11.167 17:35:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:11.427 17:35:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@499 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:14:11.427 17:35:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:11.427 17:35:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:11.427 [2024-11-27 17:35:42.615548] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:14:11.427 [2024-11-27 17:35:42.615572] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:14:11.427 [2024-11-27 17:35:42.615612] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:11.427 [2024-11-27 17:35:42.615652] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:14:11.427 [2024-11-27 17:35:42.615659] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:14:11.688 17:35:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:11.688 17:35:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@500 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:11.688 17:35:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:11.688 17:35:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@500 -- # jq -r '.[]' 00:14:11.688 17:35:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:11.688 17:35:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:11.688 17:35:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@500 -- # raid_bdev= 00:14:11.688 17:35:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@501 -- # '[' -n '' ']' 00:14:11.688 17:35:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i = 1 )) 00:14:11.688 17:35:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:14:11.688 17:35:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt2 00:14:11.688 17:35:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:11.688 17:35:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:11.688 17:35:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:11.688 17:35:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:14:11.688 17:35:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:14:11.688 17:35:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt3 00:14:11.688 17:35:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:11.688 17:35:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:11.688 17:35:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:11.688 17:35:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:14:11.688 17:35:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:14:11.688 17:35:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i = 1 )) 00:14:11.688 17:35:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:14:11.688 17:35:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@512 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:14:11.688 17:35:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:11.688 17:35:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:11.688 [2024-11-27 17:35:42.703390] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:14:11.688 [2024-11-27 17:35:42.703436] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:11.688 [2024-11-27 17:35:42.703452] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009980 00:14:11.688 [2024-11-27 17:35:42.703459] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:11.688 [2024-11-27 17:35:42.705370] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:11.688 [2024-11-27 17:35:42.705443] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:14:11.688 [2024-11-27 17:35:42.705498] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:14:11.688 [2024-11-27 17:35:42.705536] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:14:11.688 pt2 00:14:11.688 17:35:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:11.688 17:35:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@515 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 2 00:14:11.688 17:35:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:11.688 17:35:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:11.688 17:35:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:11.688 17:35:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:11.688 17:35:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:14:11.688 17:35:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:11.688 17:35:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:11.688 17:35:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:11.689 17:35:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:11.689 17:35:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:11.689 17:35:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:11.689 17:35:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:11.689 17:35:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:11.689 17:35:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:11.689 17:35:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:11.689 "name": "raid_bdev1", 00:14:11.689 "uuid": "042f6175-b4c1-4ffc-8cb6-5e77f974aeb4", 00:14:11.689 "strip_size_kb": 64, 00:14:11.689 "state": "configuring", 00:14:11.689 "raid_level": "raid5f", 00:14:11.689 "superblock": true, 00:14:11.689 "num_base_bdevs": 3, 00:14:11.689 "num_base_bdevs_discovered": 1, 00:14:11.689 "num_base_bdevs_operational": 2, 00:14:11.689 "base_bdevs_list": [ 00:14:11.689 { 00:14:11.689 "name": null, 00:14:11.689 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:11.689 "is_configured": false, 00:14:11.689 "data_offset": 2048, 00:14:11.689 "data_size": 63488 00:14:11.689 }, 00:14:11.689 { 00:14:11.689 "name": "pt2", 00:14:11.689 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:11.689 "is_configured": true, 00:14:11.689 "data_offset": 2048, 00:14:11.689 "data_size": 63488 00:14:11.689 }, 00:14:11.689 { 00:14:11.689 "name": null, 00:14:11.689 "uuid": "00000000-0000-0000-0000-000000000003", 00:14:11.689 "is_configured": false, 00:14:11.689 "data_offset": 2048, 00:14:11.689 "data_size": 63488 00:14:11.689 } 00:14:11.689 ] 00:14:11.689 }' 00:14:11.689 17:35:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:11.689 17:35:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:12.260 17:35:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i++ )) 00:14:12.260 17:35:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:14:12.260 17:35:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@519 -- # i=2 00:14:12.260 17:35:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@520 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:14:12.260 17:35:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:12.260 17:35:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:12.260 [2024-11-27 17:35:43.166700] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:14:12.260 [2024-11-27 17:35:43.166780] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:12.260 [2024-11-27 17:35:43.166810] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009f80 00:14:12.260 [2024-11-27 17:35:43.166842] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:12.260 [2024-11-27 17:35:43.167131] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:12.260 [2024-11-27 17:35:43.167194] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:14:12.260 [2024-11-27 17:35:43.167265] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:14:12.260 [2024-11-27 17:35:43.167307] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:14:12.260 [2024-11-27 17:35:43.167390] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:14:12.260 [2024-11-27 17:35:43.167424] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:14:12.260 [2024-11-27 17:35:43.167640] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:14:12.260 [2024-11-27 17:35:43.168088] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:14:12.260 [2024-11-27 17:35:43.168147] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001c80 00:14:12.260 [2024-11-27 17:35:43.168393] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:12.260 pt3 00:14:12.260 17:35:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:12.260 17:35:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@523 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:14:12.260 17:35:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:12.260 17:35:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:12.260 17:35:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:12.260 17:35:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:12.260 17:35:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:14:12.260 17:35:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:12.260 17:35:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:12.260 17:35:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:12.260 17:35:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:12.260 17:35:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:12.260 17:35:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:12.260 17:35:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:12.260 17:35:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:12.260 17:35:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:12.260 17:35:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:12.260 "name": "raid_bdev1", 00:14:12.260 "uuid": "042f6175-b4c1-4ffc-8cb6-5e77f974aeb4", 00:14:12.260 "strip_size_kb": 64, 00:14:12.260 "state": "online", 00:14:12.260 "raid_level": "raid5f", 00:14:12.260 "superblock": true, 00:14:12.260 "num_base_bdevs": 3, 00:14:12.260 "num_base_bdevs_discovered": 2, 00:14:12.260 "num_base_bdevs_operational": 2, 00:14:12.260 "base_bdevs_list": [ 00:14:12.260 { 00:14:12.260 "name": null, 00:14:12.260 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:12.260 "is_configured": false, 00:14:12.260 "data_offset": 2048, 00:14:12.260 "data_size": 63488 00:14:12.260 }, 00:14:12.260 { 00:14:12.260 "name": "pt2", 00:14:12.260 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:12.260 "is_configured": true, 00:14:12.260 "data_offset": 2048, 00:14:12.260 "data_size": 63488 00:14:12.260 }, 00:14:12.260 { 00:14:12.260 "name": "pt3", 00:14:12.260 "uuid": "00000000-0000-0000-0000-000000000003", 00:14:12.260 "is_configured": true, 00:14:12.260 "data_offset": 2048, 00:14:12.260 "data_size": 63488 00:14:12.260 } 00:14:12.260 ] 00:14:12.260 }' 00:14:12.260 17:35:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:12.260 17:35:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:12.521 17:35:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@526 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:14:12.521 17:35:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:12.521 17:35:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:12.521 [2024-11-27 17:35:43.618077] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:14:12.521 [2024-11-27 17:35:43.618099] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:14:12.521 [2024-11-27 17:35:43.618154] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:12.521 [2024-11-27 17:35:43.618195] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:14:12.521 [2024-11-27 17:35:43.618223] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name raid_bdev1, state offline 00:14:12.521 17:35:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:12.521 17:35:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@527 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:12.521 17:35:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:12.521 17:35:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:12.521 17:35:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@527 -- # jq -r '.[]' 00:14:12.521 17:35:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:12.521 17:35:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@527 -- # raid_bdev= 00:14:12.521 17:35:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@528 -- # '[' -n '' ']' 00:14:12.521 17:35:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@532 -- # '[' 3 -gt 2 ']' 00:14:12.521 17:35:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@534 -- # i=2 00:14:12.521 17:35:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@535 -- # rpc_cmd bdev_passthru_delete pt3 00:14:12.521 17:35:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:12.521 17:35:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:12.521 17:35:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:12.521 17:35:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@540 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:14:12.521 17:35:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:12.521 17:35:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:12.521 [2024-11-27 17:35:43.693936] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:14:12.521 [2024-11-27 17:35:43.693981] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:12.521 [2024-11-27 17:35:43.693993] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:14:12.521 [2024-11-27 17:35:43.694001] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:12.521 [2024-11-27 17:35:43.695948] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:12.521 [2024-11-27 17:35:43.695987] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:14:12.521 [2024-11-27 17:35:43.696036] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:14:12.521 [2024-11-27 17:35:43.696071] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:14:12.521 [2024-11-27 17:35:43.696170] bdev_raid.c:3675:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev pt2 (4) greater than existing raid bdev raid_bdev1 (2) 00:14:12.521 [2024-11-27 17:35:43.696185] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:14:12.521 [2024-11-27 17:35:43.696205] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002000 name raid_bdev1, state configuring 00:14:12.521 [2024-11-27 17:35:43.696246] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:14:12.521 pt1 00:14:12.521 17:35:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:12.521 17:35:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@542 -- # '[' 3 -gt 2 ']' 00:14:12.521 17:35:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@545 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 2 00:14:12.521 17:35:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:12.522 17:35:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:12.522 17:35:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:12.522 17:35:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:12.522 17:35:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:14:12.522 17:35:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:12.522 17:35:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:12.522 17:35:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:12.522 17:35:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:12.522 17:35:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:12.522 17:35:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:12.522 17:35:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:12.522 17:35:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:12.782 17:35:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:12.782 17:35:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:12.782 "name": "raid_bdev1", 00:14:12.782 "uuid": "042f6175-b4c1-4ffc-8cb6-5e77f974aeb4", 00:14:12.782 "strip_size_kb": 64, 00:14:12.782 "state": "configuring", 00:14:12.782 "raid_level": "raid5f", 00:14:12.782 "superblock": true, 00:14:12.782 "num_base_bdevs": 3, 00:14:12.782 "num_base_bdevs_discovered": 1, 00:14:12.782 "num_base_bdevs_operational": 2, 00:14:12.782 "base_bdevs_list": [ 00:14:12.782 { 00:14:12.782 "name": null, 00:14:12.782 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:12.782 "is_configured": false, 00:14:12.782 "data_offset": 2048, 00:14:12.782 "data_size": 63488 00:14:12.782 }, 00:14:12.782 { 00:14:12.782 "name": "pt2", 00:14:12.782 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:12.782 "is_configured": true, 00:14:12.782 "data_offset": 2048, 00:14:12.782 "data_size": 63488 00:14:12.782 }, 00:14:12.782 { 00:14:12.782 "name": null, 00:14:12.782 "uuid": "00000000-0000-0000-0000-000000000003", 00:14:12.782 "is_configured": false, 00:14:12.782 "data_offset": 2048, 00:14:12.782 "data_size": 63488 00:14:12.782 } 00:14:12.782 ] 00:14:12.782 }' 00:14:12.782 17:35:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:12.782 17:35:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:13.042 17:35:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@546 -- # rpc_cmd bdev_raid_get_bdevs configuring 00:14:13.042 17:35:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:13.042 17:35:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:13.042 17:35:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@546 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:14:13.042 17:35:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:13.042 17:35:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@546 -- # [[ false == \f\a\l\s\e ]] 00:14:13.042 17:35:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@549 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:14:13.042 17:35:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:13.042 17:35:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:13.042 [2024-11-27 17:35:44.217034] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:14:13.042 [2024-11-27 17:35:44.217124] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:13.042 [2024-11-27 17:35:44.217170] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a880 00:14:13.042 [2024-11-27 17:35:44.217202] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:13.042 [2024-11-27 17:35:44.217515] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:13.042 [2024-11-27 17:35:44.217577] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:14:13.042 [2024-11-27 17:35:44.217653] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:14:13.042 [2024-11-27 17:35:44.217710] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:14:13.042 [2024-11-27 17:35:44.217803] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000002380 00:14:13.042 [2024-11-27 17:35:44.217843] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:14:13.042 [2024-11-27 17:35:44.218063] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000026d0 00:14:13.042 [2024-11-27 17:35:44.218500] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000002380 00:14:13.042 [2024-11-27 17:35:44.218548] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000002380 00:14:13.042 [2024-11-27 17:35:44.218723] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:13.042 pt3 00:14:13.042 17:35:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:13.042 17:35:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@554 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:14:13.042 17:35:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:13.042 17:35:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:13.042 17:35:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:13.042 17:35:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:13.042 17:35:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:14:13.042 17:35:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:13.042 17:35:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:13.042 17:35:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:13.042 17:35:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:13.042 17:35:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:13.042 17:35:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:13.302 17:35:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:13.302 17:35:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:13.302 17:35:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:13.302 17:35:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:13.302 "name": "raid_bdev1", 00:14:13.302 "uuid": "042f6175-b4c1-4ffc-8cb6-5e77f974aeb4", 00:14:13.302 "strip_size_kb": 64, 00:14:13.302 "state": "online", 00:14:13.302 "raid_level": "raid5f", 00:14:13.302 "superblock": true, 00:14:13.302 "num_base_bdevs": 3, 00:14:13.302 "num_base_bdevs_discovered": 2, 00:14:13.302 "num_base_bdevs_operational": 2, 00:14:13.302 "base_bdevs_list": [ 00:14:13.302 { 00:14:13.302 "name": null, 00:14:13.302 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:13.302 "is_configured": false, 00:14:13.302 "data_offset": 2048, 00:14:13.302 "data_size": 63488 00:14:13.302 }, 00:14:13.302 { 00:14:13.302 "name": "pt2", 00:14:13.302 "uuid": "00000000-0000-0000-0000-000000000002", 00:14:13.302 "is_configured": true, 00:14:13.302 "data_offset": 2048, 00:14:13.302 "data_size": 63488 00:14:13.302 }, 00:14:13.302 { 00:14:13.302 "name": "pt3", 00:14:13.302 "uuid": "00000000-0000-0000-0000-000000000003", 00:14:13.302 "is_configured": true, 00:14:13.302 "data_offset": 2048, 00:14:13.302 "data_size": 63488 00:14:13.302 } 00:14:13.302 ] 00:14:13.302 }' 00:14:13.302 17:35:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:13.302 17:35:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:13.562 17:35:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@555 -- # rpc_cmd bdev_raid_get_bdevs online 00:14:13.562 17:35:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:13.562 17:35:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:13.562 17:35:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@555 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:14:13.562 17:35:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:13.562 17:35:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@555 -- # [[ false == \f\a\l\s\e ]] 00:14:13.562 17:35:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@558 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:14:13.562 17:35:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:13.562 17:35:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:13.562 17:35:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@558 -- # jq -r '.[] | .uuid' 00:14:13.562 [2024-11-27 17:35:44.676426] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:13.562 17:35:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:13.562 17:35:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@558 -- # '[' 042f6175-b4c1-4ffc-8cb6-5e77f974aeb4 '!=' 042f6175-b4c1-4ffc-8cb6-5e77f974aeb4 ']' 00:14:13.562 17:35:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 91514 00:14:13.562 17:35:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@950 -- # '[' -z 91514 ']' 00:14:13.562 17:35:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@954 -- # kill -0 91514 00:14:13.562 17:35:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@955 -- # uname 00:14:13.562 17:35:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:14:13.562 17:35:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 91514 00:14:13.562 17:35:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:14:13.562 17:35:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:14:13.562 17:35:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 91514' 00:14:13.562 killing process with pid 91514 00:14:13.562 17:35:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@969 -- # kill 91514 00:14:13.562 [2024-11-27 17:35:44.748050] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:14:13.562 [2024-11-27 17:35:44.748115] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:13.562 [2024-11-27 17:35:44.748187] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:14:13.562 [2024-11-27 17:35:44.748196] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002380 name raid_bdev1, state offline 00:14:13.562 17:35:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@974 -- # wait 91514 00:14:13.823 [2024-11-27 17:35:44.780961] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:14:14.084 17:35:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:14:14.084 00:14:14.084 real 0m6.615s 00:14:14.084 user 0m11.065s 00:14:14.084 sys 0m1.416s 00:14:14.084 17:35:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:14:14.084 ************************************ 00:14:14.084 END TEST raid5f_superblock_test 00:14:14.084 ************************************ 00:14:14.084 17:35:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:14:14.084 17:35:45 bdev_raid -- bdev/bdev_raid.sh@989 -- # '[' true = true ']' 00:14:14.084 17:35:45 bdev_raid -- bdev/bdev_raid.sh@990 -- # run_test raid5f_rebuild_test raid_rebuild_test raid5f 3 false false true 00:14:14.084 17:35:45 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 7 -le 1 ']' 00:14:14.084 17:35:45 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:14:14.084 17:35:45 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:14:14.084 ************************************ 00:14:14.084 START TEST raid5f_rebuild_test 00:14:14.084 ************************************ 00:14:14.084 17:35:45 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@1125 -- # raid_rebuild_test raid5f 3 false false true 00:14:14.084 17:35:45 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@569 -- # local raid_level=raid5f 00:14:14.084 17:35:45 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=3 00:14:14.084 17:35:45 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@571 -- # local superblock=false 00:14:14.084 17:35:45 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:14:14.084 17:35:45 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@573 -- # local verify=true 00:14:14.084 17:35:45 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:14:14.084 17:35:45 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:14:14.084 17:35:45 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:14:14.084 17:35:45 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:14:14.084 17:35:45 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:14:14.084 17:35:45 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:14:14.084 17:35:45 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:14:14.084 17:35:45 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:14:14.084 17:35:45 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev3 00:14:14.084 17:35:45 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:14:14.084 17:35:45 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:14:14.084 17:35:45 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:14:14.084 17:35:45 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:14:14.084 17:35:45 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:14:14.084 17:35:45 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # local strip_size 00:14:14.084 17:35:45 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@577 -- # local create_arg 00:14:14.084 17:35:45 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:14:14.084 17:35:45 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@579 -- # local data_offset 00:14:14.084 17:35:45 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@581 -- # '[' raid5f '!=' raid1 ']' 00:14:14.084 17:35:45 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@582 -- # '[' false = true ']' 00:14:14.084 17:35:45 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@586 -- # strip_size=64 00:14:14.084 17:35:45 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@587 -- # create_arg+=' -z 64' 00:14:14.084 17:35:45 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@592 -- # '[' false = true ']' 00:14:14.084 17:35:45 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@597 -- # raid_pid=91941 00:14:14.084 17:35:45 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:14:14.084 17:35:45 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@598 -- # waitforlisten 91941 00:14:14.084 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:14:14.084 17:35:45 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@831 -- # '[' -z 91941 ']' 00:14:14.084 17:35:45 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:14:14.085 17:35:45 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:14:14.085 17:35:45 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:14:14.085 17:35:45 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:14:14.085 17:35:45 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:14.085 I/O size of 3145728 is greater than zero copy threshold (65536). 00:14:14.085 Zero copy mechanism will not be used. 00:14:14.085 [2024-11-27 17:35:45.215685] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:14:14.085 [2024-11-27 17:35:45.215837] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid91941 ] 00:14:14.344 [2024-11-27 17:35:45.361184] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:14:14.345 [2024-11-27 17:35:45.404809] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:14:14.345 [2024-11-27 17:35:45.447410] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:14:14.345 [2024-11-27 17:35:45.447549] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:14:14.916 17:35:46 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:14:14.916 17:35:46 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@864 -- # return 0 00:14:14.916 17:35:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:14:14.916 17:35:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:14:14.916 17:35:46 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:14.916 17:35:46 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:14.916 BaseBdev1_malloc 00:14:14.916 17:35:46 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:14.916 17:35:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:14:14.916 17:35:46 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:14.916 17:35:46 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:14.916 [2024-11-27 17:35:46.030169] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:14:14.916 [2024-11-27 17:35:46.030270] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:14.916 [2024-11-27 17:35:46.030314] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:14:14.916 [2024-11-27 17:35:46.030348] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:14.916 [2024-11-27 17:35:46.032330] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:14.916 [2024-11-27 17:35:46.032396] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:14:14.916 BaseBdev1 00:14:14.916 17:35:46 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:14.916 17:35:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:14:14.916 17:35:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:14:14.916 17:35:46 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:14.916 17:35:46 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:14.916 BaseBdev2_malloc 00:14:14.916 17:35:46 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:14.916 17:35:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:14:14.916 17:35:46 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:14.916 17:35:46 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:14.916 [2024-11-27 17:35:46.074559] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:14:14.916 [2024-11-27 17:35:46.074745] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:14.916 [2024-11-27 17:35:46.074836] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:14:14.916 [2024-11-27 17:35:46.074929] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:14.916 [2024-11-27 17:35:46.079786] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:14.916 [2024-11-27 17:35:46.079931] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:14:14.916 BaseBdev2 00:14:14.916 17:35:46 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:14.916 17:35:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:14:14.916 17:35:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:14:14.916 17:35:46 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:14.916 17:35:46 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:14.916 BaseBdev3_malloc 00:14:14.916 17:35:46 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:14.916 17:35:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev3_malloc -p BaseBdev3 00:14:14.916 17:35:46 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:14.916 17:35:46 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:15.178 [2024-11-27 17:35:46.106171] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev3_malloc 00:14:15.178 [2024-11-27 17:35:46.106259] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:15.178 [2024-11-27 17:35:46.106301] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:14:15.178 [2024-11-27 17:35:46.106331] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:15.178 [2024-11-27 17:35:46.108332] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:15.178 [2024-11-27 17:35:46.108403] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:14:15.178 BaseBdev3 00:14:15.178 17:35:46 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:15.178 17:35:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:14:15.178 17:35:46 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:15.178 17:35:46 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:15.178 spare_malloc 00:14:15.178 17:35:46 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:15.178 17:35:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:14:15.178 17:35:46 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:15.178 17:35:46 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:15.178 spare_delay 00:14:15.178 17:35:46 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:15.178 17:35:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:14:15.178 17:35:46 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:15.178 17:35:46 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:15.178 [2024-11-27 17:35:46.146772] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:14:15.178 [2024-11-27 17:35:46.146824] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:15.178 [2024-11-27 17:35:46.146848] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:14:15.178 [2024-11-27 17:35:46.146857] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:15.178 [2024-11-27 17:35:46.148788] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:15.178 [2024-11-27 17:35:46.148822] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:14:15.178 spare 00:14:15.178 17:35:46 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:15.178 17:35:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n raid_bdev1 00:14:15.178 17:35:46 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:15.178 17:35:46 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:15.178 [2024-11-27 17:35:46.158822] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:14:15.178 [2024-11-27 17:35:46.160490] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:14:15.178 [2024-11-27 17:35:46.160545] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:14:15.178 [2024-11-27 17:35:46.160613] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:14:15.178 [2024-11-27 17:35:46.160623] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 131072, blocklen 512 00:14:15.178 [2024-11-27 17:35:46.160858] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:14:15.178 [2024-11-27 17:35:46.161261] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:14:15.178 [2024-11-27 17:35:46.161273] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:14:15.178 [2024-11-27 17:35:46.161385] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:15.178 17:35:46 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:15.178 17:35:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:14:15.178 17:35:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:15.178 17:35:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:15.178 17:35:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:15.178 17:35:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:15.178 17:35:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:15.178 17:35:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:15.178 17:35:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:15.178 17:35:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:15.178 17:35:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:15.178 17:35:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:15.178 17:35:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:15.178 17:35:46 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:15.178 17:35:46 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:15.178 17:35:46 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:15.178 17:35:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:15.178 "name": "raid_bdev1", 00:14:15.178 "uuid": "e98d97d7-023b-45bb-b7c9-366be1ec1b15", 00:14:15.178 "strip_size_kb": 64, 00:14:15.178 "state": "online", 00:14:15.178 "raid_level": "raid5f", 00:14:15.178 "superblock": false, 00:14:15.178 "num_base_bdevs": 3, 00:14:15.178 "num_base_bdevs_discovered": 3, 00:14:15.178 "num_base_bdevs_operational": 3, 00:14:15.178 "base_bdevs_list": [ 00:14:15.178 { 00:14:15.178 "name": "BaseBdev1", 00:14:15.178 "uuid": "615a36a5-adf7-5e94-a99d-cbedd231e1b3", 00:14:15.178 "is_configured": true, 00:14:15.178 "data_offset": 0, 00:14:15.178 "data_size": 65536 00:14:15.178 }, 00:14:15.178 { 00:14:15.178 "name": "BaseBdev2", 00:14:15.178 "uuid": "ca2d0cbd-93a0-55fb-b9ce-0a9fde370184", 00:14:15.178 "is_configured": true, 00:14:15.178 "data_offset": 0, 00:14:15.178 "data_size": 65536 00:14:15.178 }, 00:14:15.178 { 00:14:15.178 "name": "BaseBdev3", 00:14:15.178 "uuid": "30ab3547-d525-5ef3-aee9-915c4dbd5d35", 00:14:15.178 "is_configured": true, 00:14:15.178 "data_offset": 0, 00:14:15.178 "data_size": 65536 00:14:15.178 } 00:14:15.178 ] 00:14:15.178 }' 00:14:15.178 17:35:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:15.178 17:35:46 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:15.439 17:35:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:14:15.439 17:35:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:14:15.439 17:35:46 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:15.439 17:35:46 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:15.699 [2024-11-27 17:35:46.630151] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:15.699 17:35:46 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:15.699 17:35:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=131072 00:14:15.699 17:35:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:14:15.699 17:35:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:15.699 17:35:46 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:15.699 17:35:46 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:15.699 17:35:46 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:15.699 17:35:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@619 -- # data_offset=0 00:14:15.699 17:35:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:14:15.699 17:35:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:14:15.699 17:35:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:14:15.699 17:35:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:14:15.699 17:35:46 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:14:15.699 17:35:46 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:14:15.699 17:35:46 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@10 -- # local bdev_list 00:14:15.699 17:35:46 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:14:15.699 17:35:46 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@11 -- # local nbd_list 00:14:15.699 17:35:46 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@12 -- # local i 00:14:15.699 17:35:46 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:14:15.699 17:35:46 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:14:15.699 17:35:46 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:14:15.699 [2024-11-27 17:35:46.877601] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000026d0 00:14:15.958 /dev/nbd0 00:14:15.958 17:35:46 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:14:15.958 17:35:46 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:14:15.958 17:35:46 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:14:15.958 17:35:46 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@869 -- # local i 00:14:15.958 17:35:46 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:14:15.958 17:35:46 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:14:15.958 17:35:46 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:14:15.958 17:35:46 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@873 -- # break 00:14:15.958 17:35:46 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:14:15.958 17:35:46 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:14:15.958 17:35:46 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:14:15.958 1+0 records in 00:14:15.958 1+0 records out 00:14:15.958 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000571462 s, 7.2 MB/s 00:14:15.958 17:35:46 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:15.958 17:35:46 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@886 -- # size=4096 00:14:15.958 17:35:46 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:15.958 17:35:46 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:14:15.958 17:35:46 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@889 -- # return 0 00:14:15.958 17:35:46 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:14:15.958 17:35:46 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:14:15.958 17:35:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@629 -- # '[' raid5f = raid5f ']' 00:14:15.958 17:35:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@630 -- # write_unit_size=256 00:14:15.958 17:35:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@631 -- # echo 128 00:14:15.958 17:35:46 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=131072 count=512 oflag=direct 00:14:16.220 512+0 records in 00:14:16.220 512+0 records out 00:14:16.220 67108864 bytes (67 MB, 64 MiB) copied, 0.308549 s, 217 MB/s 00:14:16.220 17:35:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:14:16.220 17:35:47 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:14:16.220 17:35:47 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:14:16.220 17:35:47 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@50 -- # local nbd_list 00:14:16.220 17:35:47 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@51 -- # local i 00:14:16.220 17:35:47 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:14:16.220 17:35:47 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:14:16.498 17:35:47 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:14:16.498 [2024-11-27 17:35:47.460246] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:16.498 17:35:47 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:14:16.498 17:35:47 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:14:16.498 17:35:47 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:14:16.498 17:35:47 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:14:16.498 17:35:47 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:14:16.498 17:35:47 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:14:16.498 17:35:47 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:14:16.498 17:35:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:14:16.498 17:35:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:16.498 17:35:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:16.498 [2024-11-27 17:35:47.477534] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:14:16.498 17:35:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:16.498 17:35:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:14:16.498 17:35:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:16.498 17:35:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:16.498 17:35:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:16.498 17:35:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:16.498 17:35:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:14:16.498 17:35:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:16.498 17:35:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:16.498 17:35:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:16.498 17:35:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:16.498 17:35:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:16.498 17:35:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:16.498 17:35:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:16.498 17:35:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:16.498 17:35:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:16.498 17:35:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:16.498 "name": "raid_bdev1", 00:14:16.498 "uuid": "e98d97d7-023b-45bb-b7c9-366be1ec1b15", 00:14:16.498 "strip_size_kb": 64, 00:14:16.498 "state": "online", 00:14:16.498 "raid_level": "raid5f", 00:14:16.498 "superblock": false, 00:14:16.498 "num_base_bdevs": 3, 00:14:16.498 "num_base_bdevs_discovered": 2, 00:14:16.498 "num_base_bdevs_operational": 2, 00:14:16.498 "base_bdevs_list": [ 00:14:16.498 { 00:14:16.498 "name": null, 00:14:16.498 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:16.498 "is_configured": false, 00:14:16.498 "data_offset": 0, 00:14:16.498 "data_size": 65536 00:14:16.498 }, 00:14:16.498 { 00:14:16.498 "name": "BaseBdev2", 00:14:16.498 "uuid": "ca2d0cbd-93a0-55fb-b9ce-0a9fde370184", 00:14:16.498 "is_configured": true, 00:14:16.498 "data_offset": 0, 00:14:16.498 "data_size": 65536 00:14:16.498 }, 00:14:16.498 { 00:14:16.498 "name": "BaseBdev3", 00:14:16.498 "uuid": "30ab3547-d525-5ef3-aee9-915c4dbd5d35", 00:14:16.498 "is_configured": true, 00:14:16.498 "data_offset": 0, 00:14:16.498 "data_size": 65536 00:14:16.498 } 00:14:16.498 ] 00:14:16.498 }' 00:14:16.498 17:35:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:16.498 17:35:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:16.775 17:35:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:14:16.775 17:35:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:16.775 17:35:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:16.775 [2024-11-27 17:35:47.944721] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:14:16.775 [2024-11-27 17:35:47.948487] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000027cd0 00:14:16.775 [2024-11-27 17:35:47.950528] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:14:16.775 17:35:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:16.775 17:35:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@647 -- # sleep 1 00:14:18.159 17:35:48 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:18.160 17:35:48 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:18.160 17:35:48 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:18.160 17:35:48 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:18.160 17:35:48 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:18.160 17:35:48 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:18.160 17:35:48 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:18.160 17:35:48 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:18.160 17:35:48 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:18.160 17:35:48 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:18.160 17:35:49 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:18.160 "name": "raid_bdev1", 00:14:18.160 "uuid": "e98d97d7-023b-45bb-b7c9-366be1ec1b15", 00:14:18.160 "strip_size_kb": 64, 00:14:18.160 "state": "online", 00:14:18.160 "raid_level": "raid5f", 00:14:18.160 "superblock": false, 00:14:18.160 "num_base_bdevs": 3, 00:14:18.160 "num_base_bdevs_discovered": 3, 00:14:18.160 "num_base_bdevs_operational": 3, 00:14:18.160 "process": { 00:14:18.160 "type": "rebuild", 00:14:18.160 "target": "spare", 00:14:18.160 "progress": { 00:14:18.160 "blocks": 20480, 00:14:18.160 "percent": 15 00:14:18.160 } 00:14:18.160 }, 00:14:18.160 "base_bdevs_list": [ 00:14:18.160 { 00:14:18.160 "name": "spare", 00:14:18.160 "uuid": "d1112ecf-6261-5abb-9637-30e40a4d3bf8", 00:14:18.160 "is_configured": true, 00:14:18.160 "data_offset": 0, 00:14:18.160 "data_size": 65536 00:14:18.160 }, 00:14:18.160 { 00:14:18.160 "name": "BaseBdev2", 00:14:18.160 "uuid": "ca2d0cbd-93a0-55fb-b9ce-0a9fde370184", 00:14:18.160 "is_configured": true, 00:14:18.160 "data_offset": 0, 00:14:18.160 "data_size": 65536 00:14:18.160 }, 00:14:18.160 { 00:14:18.160 "name": "BaseBdev3", 00:14:18.160 "uuid": "30ab3547-d525-5ef3-aee9-915c4dbd5d35", 00:14:18.160 "is_configured": true, 00:14:18.160 "data_offset": 0, 00:14:18.160 "data_size": 65536 00:14:18.160 } 00:14:18.160 ] 00:14:18.160 }' 00:14:18.160 17:35:49 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:18.160 17:35:49 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:18.160 17:35:49 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:18.160 17:35:49 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:18.160 17:35:49 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:14:18.160 17:35:49 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:18.160 17:35:49 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:18.160 [2024-11-27 17:35:49.101284] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:14:18.160 [2024-11-27 17:35:49.157250] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:14:18.160 [2024-11-27 17:35:49.157338] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:18.160 [2024-11-27 17:35:49.157354] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:14:18.160 [2024-11-27 17:35:49.157371] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:14:18.160 17:35:49 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:18.160 17:35:49 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:14:18.160 17:35:49 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:18.160 17:35:49 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:18.160 17:35:49 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:18.160 17:35:49 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:18.160 17:35:49 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:14:18.160 17:35:49 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:18.160 17:35:49 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:18.160 17:35:49 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:18.160 17:35:49 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:18.160 17:35:49 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:18.160 17:35:49 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:18.160 17:35:49 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:18.160 17:35:49 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:18.160 17:35:49 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:18.160 17:35:49 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:18.160 "name": "raid_bdev1", 00:14:18.160 "uuid": "e98d97d7-023b-45bb-b7c9-366be1ec1b15", 00:14:18.160 "strip_size_kb": 64, 00:14:18.160 "state": "online", 00:14:18.160 "raid_level": "raid5f", 00:14:18.160 "superblock": false, 00:14:18.160 "num_base_bdevs": 3, 00:14:18.160 "num_base_bdevs_discovered": 2, 00:14:18.160 "num_base_bdevs_operational": 2, 00:14:18.160 "base_bdevs_list": [ 00:14:18.160 { 00:14:18.160 "name": null, 00:14:18.160 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:18.160 "is_configured": false, 00:14:18.160 "data_offset": 0, 00:14:18.160 "data_size": 65536 00:14:18.160 }, 00:14:18.160 { 00:14:18.160 "name": "BaseBdev2", 00:14:18.160 "uuid": "ca2d0cbd-93a0-55fb-b9ce-0a9fde370184", 00:14:18.160 "is_configured": true, 00:14:18.160 "data_offset": 0, 00:14:18.160 "data_size": 65536 00:14:18.160 }, 00:14:18.160 { 00:14:18.160 "name": "BaseBdev3", 00:14:18.160 "uuid": "30ab3547-d525-5ef3-aee9-915c4dbd5d35", 00:14:18.160 "is_configured": true, 00:14:18.160 "data_offset": 0, 00:14:18.160 "data_size": 65536 00:14:18.160 } 00:14:18.160 ] 00:14:18.160 }' 00:14:18.160 17:35:49 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:18.160 17:35:49 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:18.420 17:35:49 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:14:18.420 17:35:49 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:18.420 17:35:49 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:14:18.420 17:35:49 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=none 00:14:18.420 17:35:49 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:18.420 17:35:49 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:18.420 17:35:49 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:18.420 17:35:49 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:18.420 17:35:49 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:18.420 17:35:49 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:18.420 17:35:49 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:18.420 "name": "raid_bdev1", 00:14:18.420 "uuid": "e98d97d7-023b-45bb-b7c9-366be1ec1b15", 00:14:18.420 "strip_size_kb": 64, 00:14:18.420 "state": "online", 00:14:18.420 "raid_level": "raid5f", 00:14:18.420 "superblock": false, 00:14:18.420 "num_base_bdevs": 3, 00:14:18.420 "num_base_bdevs_discovered": 2, 00:14:18.420 "num_base_bdevs_operational": 2, 00:14:18.420 "base_bdevs_list": [ 00:14:18.420 { 00:14:18.420 "name": null, 00:14:18.420 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:18.420 "is_configured": false, 00:14:18.420 "data_offset": 0, 00:14:18.420 "data_size": 65536 00:14:18.420 }, 00:14:18.420 { 00:14:18.420 "name": "BaseBdev2", 00:14:18.420 "uuid": "ca2d0cbd-93a0-55fb-b9ce-0a9fde370184", 00:14:18.420 "is_configured": true, 00:14:18.420 "data_offset": 0, 00:14:18.420 "data_size": 65536 00:14:18.420 }, 00:14:18.420 { 00:14:18.420 "name": "BaseBdev3", 00:14:18.420 "uuid": "30ab3547-d525-5ef3-aee9-915c4dbd5d35", 00:14:18.420 "is_configured": true, 00:14:18.420 "data_offset": 0, 00:14:18.420 "data_size": 65536 00:14:18.420 } 00:14:18.420 ] 00:14:18.420 }' 00:14:18.420 17:35:49 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:18.680 17:35:49 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:14:18.680 17:35:49 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:18.680 17:35:49 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:14:18.680 17:35:49 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:14:18.680 17:35:49 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:18.680 17:35:49 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:18.680 [2024-11-27 17:35:49.709582] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:14:18.680 [2024-11-27 17:35:49.712447] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000027da0 00:14:18.680 [2024-11-27 17:35:49.714481] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:14:18.680 17:35:49 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:18.680 17:35:49 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@663 -- # sleep 1 00:14:19.620 17:35:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:19.620 17:35:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:19.620 17:35:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:19.620 17:35:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:19.620 17:35:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:19.620 17:35:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:19.620 17:35:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:19.621 17:35:50 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:19.621 17:35:50 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:19.621 17:35:50 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:19.621 17:35:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:19.621 "name": "raid_bdev1", 00:14:19.621 "uuid": "e98d97d7-023b-45bb-b7c9-366be1ec1b15", 00:14:19.621 "strip_size_kb": 64, 00:14:19.621 "state": "online", 00:14:19.621 "raid_level": "raid5f", 00:14:19.621 "superblock": false, 00:14:19.621 "num_base_bdevs": 3, 00:14:19.621 "num_base_bdevs_discovered": 3, 00:14:19.621 "num_base_bdevs_operational": 3, 00:14:19.621 "process": { 00:14:19.621 "type": "rebuild", 00:14:19.621 "target": "spare", 00:14:19.621 "progress": { 00:14:19.621 "blocks": 20480, 00:14:19.621 "percent": 15 00:14:19.621 } 00:14:19.621 }, 00:14:19.621 "base_bdevs_list": [ 00:14:19.621 { 00:14:19.621 "name": "spare", 00:14:19.621 "uuid": "d1112ecf-6261-5abb-9637-30e40a4d3bf8", 00:14:19.621 "is_configured": true, 00:14:19.621 "data_offset": 0, 00:14:19.621 "data_size": 65536 00:14:19.621 }, 00:14:19.621 { 00:14:19.621 "name": "BaseBdev2", 00:14:19.621 "uuid": "ca2d0cbd-93a0-55fb-b9ce-0a9fde370184", 00:14:19.621 "is_configured": true, 00:14:19.621 "data_offset": 0, 00:14:19.621 "data_size": 65536 00:14:19.621 }, 00:14:19.621 { 00:14:19.621 "name": "BaseBdev3", 00:14:19.621 "uuid": "30ab3547-d525-5ef3-aee9-915c4dbd5d35", 00:14:19.621 "is_configured": true, 00:14:19.621 "data_offset": 0, 00:14:19.621 "data_size": 65536 00:14:19.621 } 00:14:19.621 ] 00:14:19.621 }' 00:14:19.621 17:35:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:19.881 17:35:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:19.881 17:35:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:19.881 17:35:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:19.881 17:35:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@666 -- # '[' false = true ']' 00:14:19.881 17:35:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=3 00:14:19.881 17:35:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@693 -- # '[' raid5f = raid1 ']' 00:14:19.881 17:35:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@706 -- # local timeout=458 00:14:19.881 17:35:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:19.881 17:35:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:19.881 17:35:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:19.881 17:35:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:19.881 17:35:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:19.881 17:35:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:19.881 17:35:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:19.881 17:35:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:19.881 17:35:50 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:19.881 17:35:50 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:19.881 17:35:50 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:19.881 17:35:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:19.881 "name": "raid_bdev1", 00:14:19.881 "uuid": "e98d97d7-023b-45bb-b7c9-366be1ec1b15", 00:14:19.881 "strip_size_kb": 64, 00:14:19.881 "state": "online", 00:14:19.881 "raid_level": "raid5f", 00:14:19.881 "superblock": false, 00:14:19.881 "num_base_bdevs": 3, 00:14:19.881 "num_base_bdevs_discovered": 3, 00:14:19.881 "num_base_bdevs_operational": 3, 00:14:19.881 "process": { 00:14:19.881 "type": "rebuild", 00:14:19.881 "target": "spare", 00:14:19.881 "progress": { 00:14:19.881 "blocks": 22528, 00:14:19.881 "percent": 17 00:14:19.881 } 00:14:19.881 }, 00:14:19.881 "base_bdevs_list": [ 00:14:19.881 { 00:14:19.881 "name": "spare", 00:14:19.881 "uuid": "d1112ecf-6261-5abb-9637-30e40a4d3bf8", 00:14:19.881 "is_configured": true, 00:14:19.881 "data_offset": 0, 00:14:19.881 "data_size": 65536 00:14:19.881 }, 00:14:19.881 { 00:14:19.881 "name": "BaseBdev2", 00:14:19.881 "uuid": "ca2d0cbd-93a0-55fb-b9ce-0a9fde370184", 00:14:19.881 "is_configured": true, 00:14:19.881 "data_offset": 0, 00:14:19.881 "data_size": 65536 00:14:19.881 }, 00:14:19.881 { 00:14:19.881 "name": "BaseBdev3", 00:14:19.881 "uuid": "30ab3547-d525-5ef3-aee9-915c4dbd5d35", 00:14:19.881 "is_configured": true, 00:14:19.881 "data_offset": 0, 00:14:19.881 "data_size": 65536 00:14:19.881 } 00:14:19.881 ] 00:14:19.881 }' 00:14:19.881 17:35:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:19.881 17:35:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:19.881 17:35:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:19.881 17:35:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:19.881 17:35:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:21.263 17:35:52 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:21.263 17:35:52 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:21.263 17:35:52 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:21.263 17:35:52 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:21.263 17:35:52 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:21.263 17:35:52 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:21.263 17:35:52 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:21.263 17:35:52 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:21.263 17:35:52 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:21.263 17:35:52 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:21.263 17:35:52 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:21.263 17:35:52 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:21.263 "name": "raid_bdev1", 00:14:21.263 "uuid": "e98d97d7-023b-45bb-b7c9-366be1ec1b15", 00:14:21.263 "strip_size_kb": 64, 00:14:21.263 "state": "online", 00:14:21.263 "raid_level": "raid5f", 00:14:21.263 "superblock": false, 00:14:21.263 "num_base_bdevs": 3, 00:14:21.263 "num_base_bdevs_discovered": 3, 00:14:21.263 "num_base_bdevs_operational": 3, 00:14:21.263 "process": { 00:14:21.263 "type": "rebuild", 00:14:21.263 "target": "spare", 00:14:21.263 "progress": { 00:14:21.264 "blocks": 47104, 00:14:21.264 "percent": 35 00:14:21.264 } 00:14:21.264 }, 00:14:21.264 "base_bdevs_list": [ 00:14:21.264 { 00:14:21.264 "name": "spare", 00:14:21.264 "uuid": "d1112ecf-6261-5abb-9637-30e40a4d3bf8", 00:14:21.264 "is_configured": true, 00:14:21.264 "data_offset": 0, 00:14:21.264 "data_size": 65536 00:14:21.264 }, 00:14:21.264 { 00:14:21.264 "name": "BaseBdev2", 00:14:21.264 "uuid": "ca2d0cbd-93a0-55fb-b9ce-0a9fde370184", 00:14:21.264 "is_configured": true, 00:14:21.264 "data_offset": 0, 00:14:21.264 "data_size": 65536 00:14:21.264 }, 00:14:21.264 { 00:14:21.264 "name": "BaseBdev3", 00:14:21.264 "uuid": "30ab3547-d525-5ef3-aee9-915c4dbd5d35", 00:14:21.264 "is_configured": true, 00:14:21.264 "data_offset": 0, 00:14:21.264 "data_size": 65536 00:14:21.264 } 00:14:21.264 ] 00:14:21.264 }' 00:14:21.264 17:35:52 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:21.264 17:35:52 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:21.264 17:35:52 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:21.264 17:35:52 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:21.264 17:35:52 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:22.205 17:35:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:22.205 17:35:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:22.205 17:35:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:22.205 17:35:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:22.205 17:35:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:22.205 17:35:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:22.205 17:35:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:22.205 17:35:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:22.205 17:35:53 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:22.205 17:35:53 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:22.205 17:35:53 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:22.205 17:35:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:22.205 "name": "raid_bdev1", 00:14:22.205 "uuid": "e98d97d7-023b-45bb-b7c9-366be1ec1b15", 00:14:22.205 "strip_size_kb": 64, 00:14:22.205 "state": "online", 00:14:22.205 "raid_level": "raid5f", 00:14:22.205 "superblock": false, 00:14:22.205 "num_base_bdevs": 3, 00:14:22.205 "num_base_bdevs_discovered": 3, 00:14:22.205 "num_base_bdevs_operational": 3, 00:14:22.205 "process": { 00:14:22.205 "type": "rebuild", 00:14:22.205 "target": "spare", 00:14:22.205 "progress": { 00:14:22.205 "blocks": 69632, 00:14:22.205 "percent": 53 00:14:22.205 } 00:14:22.205 }, 00:14:22.205 "base_bdevs_list": [ 00:14:22.205 { 00:14:22.205 "name": "spare", 00:14:22.205 "uuid": "d1112ecf-6261-5abb-9637-30e40a4d3bf8", 00:14:22.205 "is_configured": true, 00:14:22.205 "data_offset": 0, 00:14:22.205 "data_size": 65536 00:14:22.205 }, 00:14:22.205 { 00:14:22.205 "name": "BaseBdev2", 00:14:22.205 "uuid": "ca2d0cbd-93a0-55fb-b9ce-0a9fde370184", 00:14:22.205 "is_configured": true, 00:14:22.205 "data_offset": 0, 00:14:22.205 "data_size": 65536 00:14:22.205 }, 00:14:22.205 { 00:14:22.205 "name": "BaseBdev3", 00:14:22.205 "uuid": "30ab3547-d525-5ef3-aee9-915c4dbd5d35", 00:14:22.205 "is_configured": true, 00:14:22.205 "data_offset": 0, 00:14:22.205 "data_size": 65536 00:14:22.205 } 00:14:22.205 ] 00:14:22.205 }' 00:14:22.205 17:35:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:22.205 17:35:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:22.205 17:35:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:22.205 17:35:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:22.205 17:35:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:23.144 17:35:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:23.144 17:35:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:23.144 17:35:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:23.144 17:35:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:23.144 17:35:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:23.144 17:35:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:23.144 17:35:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:23.144 17:35:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:23.144 17:35:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:23.144 17:35:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:23.404 17:35:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:23.404 17:35:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:23.404 "name": "raid_bdev1", 00:14:23.404 "uuid": "e98d97d7-023b-45bb-b7c9-366be1ec1b15", 00:14:23.404 "strip_size_kb": 64, 00:14:23.404 "state": "online", 00:14:23.404 "raid_level": "raid5f", 00:14:23.404 "superblock": false, 00:14:23.404 "num_base_bdevs": 3, 00:14:23.404 "num_base_bdevs_discovered": 3, 00:14:23.404 "num_base_bdevs_operational": 3, 00:14:23.404 "process": { 00:14:23.404 "type": "rebuild", 00:14:23.404 "target": "spare", 00:14:23.404 "progress": { 00:14:23.404 "blocks": 92160, 00:14:23.404 "percent": 70 00:14:23.404 } 00:14:23.404 }, 00:14:23.404 "base_bdevs_list": [ 00:14:23.404 { 00:14:23.404 "name": "spare", 00:14:23.404 "uuid": "d1112ecf-6261-5abb-9637-30e40a4d3bf8", 00:14:23.404 "is_configured": true, 00:14:23.404 "data_offset": 0, 00:14:23.404 "data_size": 65536 00:14:23.404 }, 00:14:23.404 { 00:14:23.404 "name": "BaseBdev2", 00:14:23.404 "uuid": "ca2d0cbd-93a0-55fb-b9ce-0a9fde370184", 00:14:23.404 "is_configured": true, 00:14:23.404 "data_offset": 0, 00:14:23.404 "data_size": 65536 00:14:23.404 }, 00:14:23.404 { 00:14:23.404 "name": "BaseBdev3", 00:14:23.404 "uuid": "30ab3547-d525-5ef3-aee9-915c4dbd5d35", 00:14:23.404 "is_configured": true, 00:14:23.404 "data_offset": 0, 00:14:23.404 "data_size": 65536 00:14:23.404 } 00:14:23.404 ] 00:14:23.404 }' 00:14:23.404 17:35:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:23.404 17:35:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:23.404 17:35:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:23.404 17:35:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:23.404 17:35:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:24.343 17:35:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:24.343 17:35:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:24.343 17:35:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:24.343 17:35:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:24.343 17:35:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:24.344 17:35:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:24.344 17:35:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:24.344 17:35:55 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:24.344 17:35:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:24.344 17:35:55 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:24.344 17:35:55 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:24.344 17:35:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:24.344 "name": "raid_bdev1", 00:14:24.344 "uuid": "e98d97d7-023b-45bb-b7c9-366be1ec1b15", 00:14:24.344 "strip_size_kb": 64, 00:14:24.344 "state": "online", 00:14:24.344 "raid_level": "raid5f", 00:14:24.344 "superblock": false, 00:14:24.344 "num_base_bdevs": 3, 00:14:24.344 "num_base_bdevs_discovered": 3, 00:14:24.344 "num_base_bdevs_operational": 3, 00:14:24.344 "process": { 00:14:24.344 "type": "rebuild", 00:14:24.344 "target": "spare", 00:14:24.344 "progress": { 00:14:24.344 "blocks": 116736, 00:14:24.344 "percent": 89 00:14:24.344 } 00:14:24.344 }, 00:14:24.344 "base_bdevs_list": [ 00:14:24.344 { 00:14:24.344 "name": "spare", 00:14:24.344 "uuid": "d1112ecf-6261-5abb-9637-30e40a4d3bf8", 00:14:24.344 "is_configured": true, 00:14:24.344 "data_offset": 0, 00:14:24.344 "data_size": 65536 00:14:24.344 }, 00:14:24.344 { 00:14:24.344 "name": "BaseBdev2", 00:14:24.344 "uuid": "ca2d0cbd-93a0-55fb-b9ce-0a9fde370184", 00:14:24.344 "is_configured": true, 00:14:24.344 "data_offset": 0, 00:14:24.344 "data_size": 65536 00:14:24.344 }, 00:14:24.344 { 00:14:24.344 "name": "BaseBdev3", 00:14:24.344 "uuid": "30ab3547-d525-5ef3-aee9-915c4dbd5d35", 00:14:24.344 "is_configured": true, 00:14:24.344 "data_offset": 0, 00:14:24.344 "data_size": 65536 00:14:24.344 } 00:14:24.344 ] 00:14:24.344 }' 00:14:24.604 17:35:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:24.604 17:35:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:24.604 17:35:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:24.604 17:35:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:24.604 17:35:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:25.173 [2024-11-27 17:35:56.146827] bdev_raid.c:2896:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:14:25.173 [2024-11-27 17:35:56.146942] bdev_raid.c:2558:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:14:25.173 [2024-11-27 17:35:56.147032] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:25.743 17:35:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:25.743 17:35:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:25.743 17:35:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:25.743 17:35:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:25.743 17:35:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:25.743 17:35:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:25.743 17:35:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:25.743 17:35:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:25.743 17:35:56 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:25.743 17:35:56 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:25.743 17:35:56 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:25.743 17:35:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:25.743 "name": "raid_bdev1", 00:14:25.743 "uuid": "e98d97d7-023b-45bb-b7c9-366be1ec1b15", 00:14:25.743 "strip_size_kb": 64, 00:14:25.743 "state": "online", 00:14:25.743 "raid_level": "raid5f", 00:14:25.743 "superblock": false, 00:14:25.743 "num_base_bdevs": 3, 00:14:25.743 "num_base_bdevs_discovered": 3, 00:14:25.743 "num_base_bdevs_operational": 3, 00:14:25.743 "base_bdevs_list": [ 00:14:25.743 { 00:14:25.743 "name": "spare", 00:14:25.743 "uuid": "d1112ecf-6261-5abb-9637-30e40a4d3bf8", 00:14:25.743 "is_configured": true, 00:14:25.743 "data_offset": 0, 00:14:25.743 "data_size": 65536 00:14:25.743 }, 00:14:25.743 { 00:14:25.743 "name": "BaseBdev2", 00:14:25.743 "uuid": "ca2d0cbd-93a0-55fb-b9ce-0a9fde370184", 00:14:25.743 "is_configured": true, 00:14:25.743 "data_offset": 0, 00:14:25.743 "data_size": 65536 00:14:25.743 }, 00:14:25.743 { 00:14:25.743 "name": "BaseBdev3", 00:14:25.743 "uuid": "30ab3547-d525-5ef3-aee9-915c4dbd5d35", 00:14:25.743 "is_configured": true, 00:14:25.743 "data_offset": 0, 00:14:25.743 "data_size": 65536 00:14:25.743 } 00:14:25.743 ] 00:14:25.743 }' 00:14:25.743 17:35:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:25.743 17:35:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:14:25.743 17:35:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:25.743 17:35:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:14:25.743 17:35:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@709 -- # break 00:14:25.743 17:35:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:14:25.743 17:35:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:25.743 17:35:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:14:25.743 17:35:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=none 00:14:25.743 17:35:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:25.743 17:35:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:25.743 17:35:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:25.743 17:35:56 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:25.743 17:35:56 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:25.743 17:35:56 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:25.743 17:35:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:25.743 "name": "raid_bdev1", 00:14:25.743 "uuid": "e98d97d7-023b-45bb-b7c9-366be1ec1b15", 00:14:25.743 "strip_size_kb": 64, 00:14:25.743 "state": "online", 00:14:25.743 "raid_level": "raid5f", 00:14:25.743 "superblock": false, 00:14:25.743 "num_base_bdevs": 3, 00:14:25.743 "num_base_bdevs_discovered": 3, 00:14:25.743 "num_base_bdevs_operational": 3, 00:14:25.743 "base_bdevs_list": [ 00:14:25.743 { 00:14:25.743 "name": "spare", 00:14:25.743 "uuid": "d1112ecf-6261-5abb-9637-30e40a4d3bf8", 00:14:25.743 "is_configured": true, 00:14:25.743 "data_offset": 0, 00:14:25.743 "data_size": 65536 00:14:25.743 }, 00:14:25.743 { 00:14:25.743 "name": "BaseBdev2", 00:14:25.743 "uuid": "ca2d0cbd-93a0-55fb-b9ce-0a9fde370184", 00:14:25.743 "is_configured": true, 00:14:25.743 "data_offset": 0, 00:14:25.743 "data_size": 65536 00:14:25.743 }, 00:14:25.743 { 00:14:25.743 "name": "BaseBdev3", 00:14:25.743 "uuid": "30ab3547-d525-5ef3-aee9-915c4dbd5d35", 00:14:25.743 "is_configured": true, 00:14:25.743 "data_offset": 0, 00:14:25.743 "data_size": 65536 00:14:25.743 } 00:14:25.743 ] 00:14:25.743 }' 00:14:25.743 17:35:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:25.743 17:35:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:14:25.743 17:35:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:25.743 17:35:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:14:25.743 17:35:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:14:25.743 17:35:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:25.743 17:35:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:25.743 17:35:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:25.743 17:35:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:25.743 17:35:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:25.743 17:35:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:25.743 17:35:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:25.743 17:35:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:25.743 17:35:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:25.743 17:35:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:25.743 17:35:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:25.743 17:35:56 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:25.743 17:35:56 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:26.003 17:35:56 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:26.003 17:35:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:26.003 "name": "raid_bdev1", 00:14:26.003 "uuid": "e98d97d7-023b-45bb-b7c9-366be1ec1b15", 00:14:26.003 "strip_size_kb": 64, 00:14:26.003 "state": "online", 00:14:26.003 "raid_level": "raid5f", 00:14:26.003 "superblock": false, 00:14:26.003 "num_base_bdevs": 3, 00:14:26.003 "num_base_bdevs_discovered": 3, 00:14:26.003 "num_base_bdevs_operational": 3, 00:14:26.003 "base_bdevs_list": [ 00:14:26.003 { 00:14:26.003 "name": "spare", 00:14:26.003 "uuid": "d1112ecf-6261-5abb-9637-30e40a4d3bf8", 00:14:26.003 "is_configured": true, 00:14:26.003 "data_offset": 0, 00:14:26.003 "data_size": 65536 00:14:26.003 }, 00:14:26.003 { 00:14:26.003 "name": "BaseBdev2", 00:14:26.003 "uuid": "ca2d0cbd-93a0-55fb-b9ce-0a9fde370184", 00:14:26.003 "is_configured": true, 00:14:26.003 "data_offset": 0, 00:14:26.003 "data_size": 65536 00:14:26.003 }, 00:14:26.003 { 00:14:26.003 "name": "BaseBdev3", 00:14:26.003 "uuid": "30ab3547-d525-5ef3-aee9-915c4dbd5d35", 00:14:26.003 "is_configured": true, 00:14:26.003 "data_offset": 0, 00:14:26.003 "data_size": 65536 00:14:26.003 } 00:14:26.003 ] 00:14:26.003 }' 00:14:26.003 17:35:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:26.003 17:35:56 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:26.262 17:35:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:14:26.262 17:35:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:26.262 17:35:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:26.262 [2024-11-27 17:35:57.361829] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:14:26.262 [2024-11-27 17:35:57.361914] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:14:26.262 [2024-11-27 17:35:57.362008] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:26.262 [2024-11-27 17:35:57.362106] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:14:26.262 [2024-11-27 17:35:57.362167] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:14:26.262 17:35:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:26.262 17:35:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:26.262 17:35:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@720 -- # jq length 00:14:26.262 17:35:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:26.262 17:35:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:26.262 17:35:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:26.262 17:35:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:14:26.262 17:35:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:14:26.262 17:35:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:14:26.262 17:35:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:14:26.262 17:35:57 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:14:26.262 17:35:57 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:14:26.262 17:35:57 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@10 -- # local bdev_list 00:14:26.262 17:35:57 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:14:26.262 17:35:57 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@11 -- # local nbd_list 00:14:26.262 17:35:57 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@12 -- # local i 00:14:26.262 17:35:57 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:14:26.262 17:35:57 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:14:26.262 17:35:57 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:14:26.522 /dev/nbd0 00:14:26.522 17:35:57 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:14:26.522 17:35:57 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:14:26.522 17:35:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:14:26.522 17:35:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@869 -- # local i 00:14:26.522 17:35:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:14:26.522 17:35:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:14:26.522 17:35:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:14:26.522 17:35:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@873 -- # break 00:14:26.522 17:35:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:14:26.522 17:35:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:14:26.522 17:35:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:14:26.522 1+0 records in 00:14:26.522 1+0 records out 00:14:26.522 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000515229 s, 7.9 MB/s 00:14:26.522 17:35:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:26.522 17:35:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@886 -- # size=4096 00:14:26.522 17:35:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:26.522 17:35:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:14:26.522 17:35:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@889 -- # return 0 00:14:26.522 17:35:57 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:14:26.522 17:35:57 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:14:26.522 17:35:57 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:14:26.782 /dev/nbd1 00:14:26.782 17:35:57 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:14:26.782 17:35:57 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:14:26.782 17:35:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@868 -- # local nbd_name=nbd1 00:14:26.782 17:35:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@869 -- # local i 00:14:26.782 17:35:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:14:26.782 17:35:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:14:26.782 17:35:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@872 -- # grep -q -w nbd1 /proc/partitions 00:14:26.782 17:35:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@873 -- # break 00:14:26.782 17:35:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:14:26.782 17:35:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:14:26.782 17:35:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@885 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:14:26.782 1+0 records in 00:14:26.782 1+0 records out 00:14:26.782 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000547875 s, 7.5 MB/s 00:14:26.782 17:35:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:26.782 17:35:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@886 -- # size=4096 00:14:26.782 17:35:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:26.782 17:35:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:14:26.782 17:35:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@889 -- # return 0 00:14:26.782 17:35:57 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:14:26.782 17:35:57 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:14:26.782 17:35:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@738 -- # cmp -i 0 /dev/nbd0 /dev/nbd1 00:14:27.042 17:35:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:14:27.042 17:35:57 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:14:27.042 17:35:57 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:14:27.042 17:35:57 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@50 -- # local nbd_list 00:14:27.042 17:35:57 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@51 -- # local i 00:14:27.042 17:35:57 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:14:27.042 17:35:57 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:14:27.042 17:35:58 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:14:27.042 17:35:58 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:14:27.042 17:35:58 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:14:27.042 17:35:58 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:14:27.042 17:35:58 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:14:27.042 17:35:58 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:14:27.042 17:35:58 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:14:27.042 17:35:58 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:14:27.042 17:35:58 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:14:27.042 17:35:58 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:14:27.303 17:35:58 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:14:27.303 17:35:58 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:14:27.303 17:35:58 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:14:27.303 17:35:58 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:14:27.303 17:35:58 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:14:27.303 17:35:58 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:14:27.303 17:35:58 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:14:27.303 17:35:58 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:14:27.303 17:35:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@743 -- # '[' false = true ']' 00:14:27.303 17:35:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@784 -- # killprocess 91941 00:14:27.303 17:35:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@950 -- # '[' -z 91941 ']' 00:14:27.303 17:35:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@954 -- # kill -0 91941 00:14:27.303 17:35:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@955 -- # uname 00:14:27.303 17:35:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:14:27.303 17:35:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 91941 00:14:27.303 killing process with pid 91941 00:14:27.303 Received shutdown signal, test time was about 60.000000 seconds 00:14:27.303 00:14:27.303 Latency(us) 00:14:27.303 [2024-11-27T17:35:58.495Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:14:27.303 [2024-11-27T17:35:58.495Z] =================================================================================================================== 00:14:27.303 [2024-11-27T17:35:58.495Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:14:27.303 17:35:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:14:27.303 17:35:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:14:27.303 17:35:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 91941' 00:14:27.303 17:35:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@969 -- # kill 91941 00:14:27.303 [2024-11-27 17:35:58.466815] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:14:27.303 17:35:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@974 -- # wait 91941 00:14:27.562 [2024-11-27 17:35:58.507089] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:14:27.562 17:35:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@786 -- # return 0 00:14:27.562 00:14:27.562 real 0m13.623s 00:14:27.562 user 0m17.019s 00:14:27.562 sys 0m2.022s 00:14:27.562 ************************************ 00:14:27.562 END TEST raid5f_rebuild_test 00:14:27.562 ************************************ 00:14:27.562 17:35:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:14:27.562 17:35:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:14:27.823 17:35:58 bdev_raid -- bdev/bdev_raid.sh@991 -- # run_test raid5f_rebuild_test_sb raid_rebuild_test raid5f 3 true false true 00:14:27.823 17:35:58 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 7 -le 1 ']' 00:14:27.823 17:35:58 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:14:27.823 17:35:58 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:14:27.823 ************************************ 00:14:27.823 START TEST raid5f_rebuild_test_sb 00:14:27.823 ************************************ 00:14:27.823 17:35:58 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@1125 -- # raid_rebuild_test raid5f 3 true false true 00:14:27.823 17:35:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@569 -- # local raid_level=raid5f 00:14:27.823 17:35:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=3 00:14:27.823 17:35:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:14:27.823 17:35:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:14:27.823 17:35:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@573 -- # local verify=true 00:14:27.823 17:35:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:14:27.823 17:35:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:14:27.823 17:35:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:14:27.823 17:35:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:14:27.823 17:35:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:14:27.823 17:35:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:14:27.823 17:35:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:14:27.823 17:35:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:14:27.823 17:35:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev3 00:14:27.823 17:35:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:14:27.823 17:35:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:14:27.823 17:35:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3') 00:14:27.823 17:35:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:14:27.823 17:35:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:14:27.823 17:35:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # local strip_size 00:14:27.823 17:35:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@577 -- # local create_arg 00:14:27.823 17:35:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:14:27.823 17:35:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@579 -- # local data_offset 00:14:27.823 17:35:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@581 -- # '[' raid5f '!=' raid1 ']' 00:14:27.823 17:35:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@582 -- # '[' false = true ']' 00:14:27.823 17:35:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@586 -- # strip_size=64 00:14:27.823 17:35:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@587 -- # create_arg+=' -z 64' 00:14:27.823 17:35:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:14:27.823 17:35:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:14:27.823 17:35:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@597 -- # raid_pid=92370 00:14:27.823 17:35:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:14:27.823 17:35:58 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@598 -- # waitforlisten 92370 00:14:27.823 17:35:58 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@831 -- # '[' -z 92370 ']' 00:14:27.823 17:35:58 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:14:27.823 17:35:58 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@836 -- # local max_retries=100 00:14:27.823 17:35:58 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:14:27.823 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:14:27.823 17:35:58 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@840 -- # xtrace_disable 00:14:27.823 17:35:58 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:27.823 [2024-11-27 17:35:58.920692] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:14:27.823 [2024-11-27 17:35:58.920874] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --matchI/O size of 3145728 is greater than zero copy threshold (65536). 00:14:27.823 Zero copy mechanism will not be used. 00:14:27.823 -allocations --file-prefix=spdk_pid92370 ] 00:14:28.082 [2024-11-27 17:35:59.066393] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:14:28.082 [2024-11-27 17:35:59.111730] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:14:28.082 [2024-11-27 17:35:59.154191] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:14:28.082 [2024-11-27 17:35:59.154222] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:14:28.653 17:35:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:14:28.653 17:35:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@864 -- # return 0 00:14:28.653 17:35:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:14:28.653 17:35:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:14:28.653 17:35:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:28.653 17:35:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:28.653 BaseBdev1_malloc 00:14:28.653 17:35:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:28.653 17:35:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:14:28.653 17:35:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:28.653 17:35:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:28.653 [2024-11-27 17:35:59.748498] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:14:28.653 [2024-11-27 17:35:59.748559] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:28.653 [2024-11-27 17:35:59.748579] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:14:28.653 [2024-11-27 17:35:59.748593] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:28.653 [2024-11-27 17:35:59.750598] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:28.653 [2024-11-27 17:35:59.750633] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:14:28.653 BaseBdev1 00:14:28.653 17:35:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:28.653 17:35:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:14:28.653 17:35:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:14:28.653 17:35:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:28.653 17:35:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:28.653 BaseBdev2_malloc 00:14:28.653 17:35:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:28.653 17:35:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:14:28.653 17:35:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:28.653 17:35:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:28.653 [2024-11-27 17:35:59.787995] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:14:28.653 [2024-11-27 17:35:59.788090] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:28.653 [2024-11-27 17:35:59.788132] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:14:28.653 [2024-11-27 17:35:59.788186] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:28.653 [2024-11-27 17:35:59.792611] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:28.653 [2024-11-27 17:35:59.792682] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:14:28.653 BaseBdev2 00:14:28.653 17:35:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:28.653 17:35:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:14:28.653 17:35:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:14:28.653 17:35:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:28.653 17:35:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:28.653 BaseBdev3_malloc 00:14:28.653 17:35:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:28.653 17:35:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev3_malloc -p BaseBdev3 00:14:28.653 17:35:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:28.653 17:35:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:28.653 [2024-11-27 17:35:59.818723] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev3_malloc 00:14:28.653 [2024-11-27 17:35:59.818783] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:28.653 [2024-11-27 17:35:59.818808] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:14:28.653 [2024-11-27 17:35:59.818817] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:28.653 [2024-11-27 17:35:59.820788] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:28.653 [2024-11-27 17:35:59.820865] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:14:28.653 BaseBdev3 00:14:28.653 17:35:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:28.653 17:35:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:14:28.653 17:35:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:28.653 17:35:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:28.653 spare_malloc 00:14:28.653 17:35:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:28.653 17:35:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:14:28.653 17:35:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:28.653 17:35:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:28.913 spare_delay 00:14:28.913 17:35:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:28.913 17:35:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:14:28.913 17:35:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:28.913 17:35:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:28.913 [2024-11-27 17:35:59.859314] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:14:28.913 [2024-11-27 17:35:59.859418] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:28.913 [2024-11-27 17:35:59.859447] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009080 00:14:28.913 [2024-11-27 17:35:59.859456] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:28.913 [2024-11-27 17:35:59.861430] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:28.913 [2024-11-27 17:35:59.861465] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:14:28.913 spare 00:14:28.913 17:35:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:28.913 17:35:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3'\''' -n raid_bdev1 00:14:28.913 17:35:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:28.913 17:35:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:28.913 [2024-11-27 17:35:59.871366] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:14:28.914 [2024-11-27 17:35:59.873076] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:14:28.914 [2024-11-27 17:35:59.873140] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:14:28.914 [2024-11-27 17:35:59.873294] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:14:28.914 [2024-11-27 17:35:59.873313] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:14:28.914 [2024-11-27 17:35:59.873550] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:14:28.914 [2024-11-27 17:35:59.873954] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:14:28.914 [2024-11-27 17:35:59.873970] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:14:28.914 [2024-11-27 17:35:59.874075] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:28.914 17:35:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:28.914 17:35:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:14:28.914 17:35:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:28.914 17:35:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:28.914 17:35:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:28.914 17:35:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:28.914 17:35:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:28.914 17:35:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:28.914 17:35:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:28.914 17:35:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:28.914 17:35:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:28.914 17:35:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:28.914 17:35:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:28.914 17:35:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:28.914 17:35:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:28.914 17:35:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:28.914 17:35:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:28.914 "name": "raid_bdev1", 00:14:28.914 "uuid": "99f3834e-db84-45f3-9a79-784e865c5710", 00:14:28.914 "strip_size_kb": 64, 00:14:28.914 "state": "online", 00:14:28.914 "raid_level": "raid5f", 00:14:28.914 "superblock": true, 00:14:28.914 "num_base_bdevs": 3, 00:14:28.914 "num_base_bdevs_discovered": 3, 00:14:28.914 "num_base_bdevs_operational": 3, 00:14:28.914 "base_bdevs_list": [ 00:14:28.914 { 00:14:28.914 "name": "BaseBdev1", 00:14:28.914 "uuid": "c9bce139-0413-50f0-a27e-db235611d7fd", 00:14:28.914 "is_configured": true, 00:14:28.914 "data_offset": 2048, 00:14:28.914 "data_size": 63488 00:14:28.914 }, 00:14:28.914 { 00:14:28.914 "name": "BaseBdev2", 00:14:28.914 "uuid": "de33ca6b-8e71-5d0e-874a-8324ac844b6e", 00:14:28.914 "is_configured": true, 00:14:28.914 "data_offset": 2048, 00:14:28.914 "data_size": 63488 00:14:28.914 }, 00:14:28.914 { 00:14:28.914 "name": "BaseBdev3", 00:14:28.914 "uuid": "44977377-4710-517f-8a52-2efd69972bbe", 00:14:28.914 "is_configured": true, 00:14:28.914 "data_offset": 2048, 00:14:28.914 "data_size": 63488 00:14:28.914 } 00:14:28.914 ] 00:14:28.914 }' 00:14:28.914 17:35:59 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:28.914 17:35:59 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:29.173 17:36:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:14:29.173 17:36:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:14:29.173 17:36:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:29.173 17:36:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:29.434 [2024-11-27 17:36:00.367044] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:29.434 17:36:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:29.434 17:36:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=126976 00:14:29.434 17:36:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:29.434 17:36:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:14:29.434 17:36:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:29.434 17:36:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:29.434 17:36:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:29.434 17:36:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # data_offset=2048 00:14:29.434 17:36:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:14:29.434 17:36:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:14:29.434 17:36:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:14:29.434 17:36:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:14:29.434 17:36:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:14:29.434 17:36:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:14:29.434 17:36:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # local bdev_list 00:14:29.434 17:36:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:14:29.434 17:36:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # local nbd_list 00:14:29.434 17:36:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@12 -- # local i 00:14:29.434 17:36:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:14:29.434 17:36:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:14:29.434 17:36:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:14:29.694 [2024-11-27 17:36:00.634398] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000026d0 00:14:29.694 /dev/nbd0 00:14:29.694 17:36:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:14:29.694 17:36:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:14:29.694 17:36:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:14:29.694 17:36:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@869 -- # local i 00:14:29.694 17:36:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:14:29.694 17:36:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:14:29.694 17:36:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:14:29.694 17:36:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@873 -- # break 00:14:29.694 17:36:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:14:29.694 17:36:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:14:29.694 17:36:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:14:29.694 1+0 records in 00:14:29.694 1+0 records out 00:14:29.694 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000338268 s, 12.1 MB/s 00:14:29.694 17:36:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:29.694 17:36:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@886 -- # size=4096 00:14:29.694 17:36:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:29.694 17:36:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:14:29.694 17:36:00 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@889 -- # return 0 00:14:29.694 17:36:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:14:29.694 17:36:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:14:29.694 17:36:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@629 -- # '[' raid5f = raid5f ']' 00:14:29.694 17:36:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@630 -- # write_unit_size=256 00:14:29.694 17:36:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@631 -- # echo 128 00:14:29.694 17:36:00 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=131072 count=496 oflag=direct 00:14:29.954 496+0 records in 00:14:29.954 496+0 records out 00:14:29.954 65011712 bytes (65 MB, 62 MiB) copied, 0.280507 s, 232 MB/s 00:14:29.954 17:36:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:14:29.954 17:36:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:14:29.954 17:36:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:14:29.954 17:36:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # local nbd_list 00:14:29.954 17:36:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@51 -- # local i 00:14:29.954 17:36:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:14:29.954 17:36:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:14:30.214 [2024-11-27 17:36:01.211175] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:30.215 17:36:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:14:30.215 17:36:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:14:30.215 17:36:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:14:30.215 17:36:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:14:30.215 17:36:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:14:30.215 17:36:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:14:30.215 17:36:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:14:30.215 17:36:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:14:30.215 17:36:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:14:30.215 17:36:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:30.215 17:36:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:30.215 [2024-11-27 17:36:01.239185] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:14:30.215 17:36:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:30.215 17:36:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:14:30.215 17:36:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:30.215 17:36:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:30.215 17:36:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:30.215 17:36:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:30.215 17:36:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:14:30.215 17:36:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:30.215 17:36:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:30.215 17:36:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:30.215 17:36:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:30.215 17:36:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:30.215 17:36:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:30.215 17:36:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:30.215 17:36:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:30.215 17:36:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:30.215 17:36:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:30.215 "name": "raid_bdev1", 00:14:30.215 "uuid": "99f3834e-db84-45f3-9a79-784e865c5710", 00:14:30.215 "strip_size_kb": 64, 00:14:30.215 "state": "online", 00:14:30.215 "raid_level": "raid5f", 00:14:30.215 "superblock": true, 00:14:30.215 "num_base_bdevs": 3, 00:14:30.215 "num_base_bdevs_discovered": 2, 00:14:30.215 "num_base_bdevs_operational": 2, 00:14:30.215 "base_bdevs_list": [ 00:14:30.215 { 00:14:30.215 "name": null, 00:14:30.215 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:30.215 "is_configured": false, 00:14:30.215 "data_offset": 0, 00:14:30.215 "data_size": 63488 00:14:30.215 }, 00:14:30.215 { 00:14:30.215 "name": "BaseBdev2", 00:14:30.215 "uuid": "de33ca6b-8e71-5d0e-874a-8324ac844b6e", 00:14:30.215 "is_configured": true, 00:14:30.215 "data_offset": 2048, 00:14:30.215 "data_size": 63488 00:14:30.215 }, 00:14:30.215 { 00:14:30.215 "name": "BaseBdev3", 00:14:30.215 "uuid": "44977377-4710-517f-8a52-2efd69972bbe", 00:14:30.215 "is_configured": true, 00:14:30.215 "data_offset": 2048, 00:14:30.215 "data_size": 63488 00:14:30.215 } 00:14:30.215 ] 00:14:30.215 }' 00:14:30.215 17:36:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:30.215 17:36:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:30.783 17:36:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:14:30.783 17:36:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:30.783 17:36:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:30.783 [2024-11-27 17:36:01.706466] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:14:30.783 [2024-11-27 17:36:01.710308] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000255d0 00:14:30.783 [2024-11-27 17:36:01.712570] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:14:30.783 17:36:01 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:30.783 17:36:01 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@647 -- # sleep 1 00:14:31.728 17:36:02 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:31.728 17:36:02 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:31.728 17:36:02 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:31.728 17:36:02 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:31.728 17:36:02 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:31.728 17:36:02 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:31.728 17:36:02 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:31.729 17:36:02 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:31.729 17:36:02 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:31.729 17:36:02 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:31.729 17:36:02 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:31.729 "name": "raid_bdev1", 00:14:31.729 "uuid": "99f3834e-db84-45f3-9a79-784e865c5710", 00:14:31.729 "strip_size_kb": 64, 00:14:31.729 "state": "online", 00:14:31.729 "raid_level": "raid5f", 00:14:31.729 "superblock": true, 00:14:31.729 "num_base_bdevs": 3, 00:14:31.729 "num_base_bdevs_discovered": 3, 00:14:31.729 "num_base_bdevs_operational": 3, 00:14:31.729 "process": { 00:14:31.729 "type": "rebuild", 00:14:31.729 "target": "spare", 00:14:31.729 "progress": { 00:14:31.729 "blocks": 20480, 00:14:31.729 "percent": 16 00:14:31.729 } 00:14:31.729 }, 00:14:31.729 "base_bdevs_list": [ 00:14:31.729 { 00:14:31.729 "name": "spare", 00:14:31.729 "uuid": "a8e64158-5cf7-5b2b-8209-ce201a4b16d0", 00:14:31.729 "is_configured": true, 00:14:31.729 "data_offset": 2048, 00:14:31.729 "data_size": 63488 00:14:31.729 }, 00:14:31.729 { 00:14:31.729 "name": "BaseBdev2", 00:14:31.729 "uuid": "de33ca6b-8e71-5d0e-874a-8324ac844b6e", 00:14:31.729 "is_configured": true, 00:14:31.729 "data_offset": 2048, 00:14:31.729 "data_size": 63488 00:14:31.729 }, 00:14:31.729 { 00:14:31.729 "name": "BaseBdev3", 00:14:31.729 "uuid": "44977377-4710-517f-8a52-2efd69972bbe", 00:14:31.729 "is_configured": true, 00:14:31.729 "data_offset": 2048, 00:14:31.729 "data_size": 63488 00:14:31.729 } 00:14:31.729 ] 00:14:31.729 }' 00:14:31.729 17:36:02 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:31.729 17:36:02 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:31.729 17:36:02 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:31.729 17:36:02 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:31.729 17:36:02 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:14:31.729 17:36:02 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:31.729 17:36:02 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:31.729 [2024-11-27 17:36:02.871470] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:14:31.988 [2024-11-27 17:36:02.919593] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:14:31.988 [2024-11-27 17:36:02.919650] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:31.988 [2024-11-27 17:36:02.919681] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:14:31.988 [2024-11-27 17:36:02.919695] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:14:31.988 17:36:02 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:31.988 17:36:02 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:14:31.988 17:36:02 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:31.988 17:36:02 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:31.988 17:36:02 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:31.988 17:36:02 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:31.988 17:36:02 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:14:31.988 17:36:02 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:31.988 17:36:02 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:31.988 17:36:02 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:31.988 17:36:02 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:31.988 17:36:02 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:31.988 17:36:02 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:31.988 17:36:02 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:31.988 17:36:02 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:31.988 17:36:02 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:31.988 17:36:02 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:31.988 "name": "raid_bdev1", 00:14:31.988 "uuid": "99f3834e-db84-45f3-9a79-784e865c5710", 00:14:31.988 "strip_size_kb": 64, 00:14:31.988 "state": "online", 00:14:31.988 "raid_level": "raid5f", 00:14:31.988 "superblock": true, 00:14:31.988 "num_base_bdevs": 3, 00:14:31.988 "num_base_bdevs_discovered": 2, 00:14:31.988 "num_base_bdevs_operational": 2, 00:14:31.988 "base_bdevs_list": [ 00:14:31.988 { 00:14:31.988 "name": null, 00:14:31.988 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:31.988 "is_configured": false, 00:14:31.988 "data_offset": 0, 00:14:31.988 "data_size": 63488 00:14:31.988 }, 00:14:31.988 { 00:14:31.988 "name": "BaseBdev2", 00:14:31.988 "uuid": "de33ca6b-8e71-5d0e-874a-8324ac844b6e", 00:14:31.988 "is_configured": true, 00:14:31.988 "data_offset": 2048, 00:14:31.988 "data_size": 63488 00:14:31.988 }, 00:14:31.988 { 00:14:31.988 "name": "BaseBdev3", 00:14:31.988 "uuid": "44977377-4710-517f-8a52-2efd69972bbe", 00:14:31.988 "is_configured": true, 00:14:31.988 "data_offset": 2048, 00:14:31.988 "data_size": 63488 00:14:31.988 } 00:14:31.988 ] 00:14:31.988 }' 00:14:31.988 17:36:02 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:31.988 17:36:02 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:32.246 17:36:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:14:32.246 17:36:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:32.246 17:36:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:14:32.246 17:36:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:14:32.246 17:36:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:32.246 17:36:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:32.246 17:36:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:32.246 17:36:03 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:32.246 17:36:03 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:32.246 17:36:03 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:32.246 17:36:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:32.247 "name": "raid_bdev1", 00:14:32.247 "uuid": "99f3834e-db84-45f3-9a79-784e865c5710", 00:14:32.247 "strip_size_kb": 64, 00:14:32.247 "state": "online", 00:14:32.247 "raid_level": "raid5f", 00:14:32.247 "superblock": true, 00:14:32.247 "num_base_bdevs": 3, 00:14:32.247 "num_base_bdevs_discovered": 2, 00:14:32.247 "num_base_bdevs_operational": 2, 00:14:32.247 "base_bdevs_list": [ 00:14:32.247 { 00:14:32.247 "name": null, 00:14:32.247 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:32.247 "is_configured": false, 00:14:32.247 "data_offset": 0, 00:14:32.247 "data_size": 63488 00:14:32.247 }, 00:14:32.247 { 00:14:32.247 "name": "BaseBdev2", 00:14:32.247 "uuid": "de33ca6b-8e71-5d0e-874a-8324ac844b6e", 00:14:32.247 "is_configured": true, 00:14:32.247 "data_offset": 2048, 00:14:32.247 "data_size": 63488 00:14:32.247 }, 00:14:32.247 { 00:14:32.247 "name": "BaseBdev3", 00:14:32.247 "uuid": "44977377-4710-517f-8a52-2efd69972bbe", 00:14:32.247 "is_configured": true, 00:14:32.247 "data_offset": 2048, 00:14:32.247 "data_size": 63488 00:14:32.247 } 00:14:32.247 ] 00:14:32.247 }' 00:14:32.506 17:36:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:32.506 17:36:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:14:32.506 17:36:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:32.506 17:36:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:14:32.506 17:36:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:14:32.506 17:36:03 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:32.507 17:36:03 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:32.507 [2024-11-27 17:36:03.539952] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:14:32.507 [2024-11-27 17:36:03.542971] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000256a0 00:14:32.507 [2024-11-27 17:36:03.545026] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:14:32.507 17:36:03 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:32.507 17:36:03 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@663 -- # sleep 1 00:14:33.446 17:36:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:33.446 17:36:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:33.446 17:36:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:33.446 17:36:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:33.446 17:36:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:33.446 17:36:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:33.446 17:36:04 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:33.446 17:36:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:33.446 17:36:04 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:33.446 17:36:04 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:33.446 17:36:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:33.446 "name": "raid_bdev1", 00:14:33.446 "uuid": "99f3834e-db84-45f3-9a79-784e865c5710", 00:14:33.446 "strip_size_kb": 64, 00:14:33.446 "state": "online", 00:14:33.446 "raid_level": "raid5f", 00:14:33.446 "superblock": true, 00:14:33.446 "num_base_bdevs": 3, 00:14:33.446 "num_base_bdevs_discovered": 3, 00:14:33.446 "num_base_bdevs_operational": 3, 00:14:33.446 "process": { 00:14:33.446 "type": "rebuild", 00:14:33.446 "target": "spare", 00:14:33.446 "progress": { 00:14:33.446 "blocks": 20480, 00:14:33.446 "percent": 16 00:14:33.446 } 00:14:33.446 }, 00:14:33.446 "base_bdevs_list": [ 00:14:33.446 { 00:14:33.446 "name": "spare", 00:14:33.446 "uuid": "a8e64158-5cf7-5b2b-8209-ce201a4b16d0", 00:14:33.446 "is_configured": true, 00:14:33.446 "data_offset": 2048, 00:14:33.446 "data_size": 63488 00:14:33.446 }, 00:14:33.446 { 00:14:33.446 "name": "BaseBdev2", 00:14:33.446 "uuid": "de33ca6b-8e71-5d0e-874a-8324ac844b6e", 00:14:33.446 "is_configured": true, 00:14:33.446 "data_offset": 2048, 00:14:33.446 "data_size": 63488 00:14:33.446 }, 00:14:33.446 { 00:14:33.446 "name": "BaseBdev3", 00:14:33.446 "uuid": "44977377-4710-517f-8a52-2efd69972bbe", 00:14:33.446 "is_configured": true, 00:14:33.446 "data_offset": 2048, 00:14:33.446 "data_size": 63488 00:14:33.446 } 00:14:33.446 ] 00:14:33.446 }' 00:14:33.446 17:36:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:33.707 17:36:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:33.707 17:36:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:33.707 17:36:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:33.707 17:36:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:14:33.707 17:36:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:14:33.707 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:14:33.707 17:36:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=3 00:14:33.707 17:36:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@693 -- # '[' raid5f = raid1 ']' 00:14:33.707 17:36:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@706 -- # local timeout=472 00:14:33.707 17:36:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:33.707 17:36:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:33.707 17:36:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:33.707 17:36:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:33.707 17:36:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:33.707 17:36:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:33.707 17:36:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:33.707 17:36:04 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:33.707 17:36:04 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:33.707 17:36:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:33.707 17:36:04 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:33.707 17:36:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:33.707 "name": "raid_bdev1", 00:14:33.707 "uuid": "99f3834e-db84-45f3-9a79-784e865c5710", 00:14:33.707 "strip_size_kb": 64, 00:14:33.707 "state": "online", 00:14:33.707 "raid_level": "raid5f", 00:14:33.707 "superblock": true, 00:14:33.707 "num_base_bdevs": 3, 00:14:33.707 "num_base_bdevs_discovered": 3, 00:14:33.707 "num_base_bdevs_operational": 3, 00:14:33.707 "process": { 00:14:33.707 "type": "rebuild", 00:14:33.707 "target": "spare", 00:14:33.707 "progress": { 00:14:33.707 "blocks": 22528, 00:14:33.707 "percent": 17 00:14:33.707 } 00:14:33.707 }, 00:14:33.707 "base_bdevs_list": [ 00:14:33.707 { 00:14:33.707 "name": "spare", 00:14:33.708 "uuid": "a8e64158-5cf7-5b2b-8209-ce201a4b16d0", 00:14:33.708 "is_configured": true, 00:14:33.708 "data_offset": 2048, 00:14:33.708 "data_size": 63488 00:14:33.708 }, 00:14:33.708 { 00:14:33.708 "name": "BaseBdev2", 00:14:33.708 "uuid": "de33ca6b-8e71-5d0e-874a-8324ac844b6e", 00:14:33.708 "is_configured": true, 00:14:33.708 "data_offset": 2048, 00:14:33.708 "data_size": 63488 00:14:33.708 }, 00:14:33.708 { 00:14:33.708 "name": "BaseBdev3", 00:14:33.708 "uuid": "44977377-4710-517f-8a52-2efd69972bbe", 00:14:33.708 "is_configured": true, 00:14:33.708 "data_offset": 2048, 00:14:33.708 "data_size": 63488 00:14:33.708 } 00:14:33.708 ] 00:14:33.708 }' 00:14:33.708 17:36:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:33.708 17:36:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:33.708 17:36:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:33.708 17:36:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:33.708 17:36:04 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:35.090 17:36:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:35.090 17:36:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:35.090 17:36:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:35.090 17:36:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:35.090 17:36:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:35.090 17:36:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:35.090 17:36:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:35.090 17:36:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:35.090 17:36:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:35.090 17:36:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:35.090 17:36:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:35.090 17:36:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:35.090 "name": "raid_bdev1", 00:14:35.090 "uuid": "99f3834e-db84-45f3-9a79-784e865c5710", 00:14:35.090 "strip_size_kb": 64, 00:14:35.090 "state": "online", 00:14:35.090 "raid_level": "raid5f", 00:14:35.090 "superblock": true, 00:14:35.090 "num_base_bdevs": 3, 00:14:35.090 "num_base_bdevs_discovered": 3, 00:14:35.090 "num_base_bdevs_operational": 3, 00:14:35.090 "process": { 00:14:35.090 "type": "rebuild", 00:14:35.090 "target": "spare", 00:14:35.090 "progress": { 00:14:35.090 "blocks": 47104, 00:14:35.090 "percent": 37 00:14:35.090 } 00:14:35.090 }, 00:14:35.090 "base_bdevs_list": [ 00:14:35.090 { 00:14:35.090 "name": "spare", 00:14:35.090 "uuid": "a8e64158-5cf7-5b2b-8209-ce201a4b16d0", 00:14:35.090 "is_configured": true, 00:14:35.090 "data_offset": 2048, 00:14:35.090 "data_size": 63488 00:14:35.090 }, 00:14:35.090 { 00:14:35.091 "name": "BaseBdev2", 00:14:35.091 "uuid": "de33ca6b-8e71-5d0e-874a-8324ac844b6e", 00:14:35.091 "is_configured": true, 00:14:35.091 "data_offset": 2048, 00:14:35.091 "data_size": 63488 00:14:35.091 }, 00:14:35.091 { 00:14:35.091 "name": "BaseBdev3", 00:14:35.091 "uuid": "44977377-4710-517f-8a52-2efd69972bbe", 00:14:35.091 "is_configured": true, 00:14:35.091 "data_offset": 2048, 00:14:35.091 "data_size": 63488 00:14:35.091 } 00:14:35.091 ] 00:14:35.091 }' 00:14:35.091 17:36:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:35.091 17:36:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:35.091 17:36:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:35.091 17:36:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:35.091 17:36:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:36.030 17:36:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:36.031 17:36:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:36.031 17:36:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:36.031 17:36:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:36.031 17:36:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:36.031 17:36:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:36.031 17:36:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:36.031 17:36:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:36.031 17:36:07 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:36.031 17:36:07 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:36.031 17:36:07 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:36.031 17:36:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:36.031 "name": "raid_bdev1", 00:14:36.031 "uuid": "99f3834e-db84-45f3-9a79-784e865c5710", 00:14:36.031 "strip_size_kb": 64, 00:14:36.031 "state": "online", 00:14:36.031 "raid_level": "raid5f", 00:14:36.031 "superblock": true, 00:14:36.031 "num_base_bdevs": 3, 00:14:36.031 "num_base_bdevs_discovered": 3, 00:14:36.031 "num_base_bdevs_operational": 3, 00:14:36.031 "process": { 00:14:36.031 "type": "rebuild", 00:14:36.031 "target": "spare", 00:14:36.031 "progress": { 00:14:36.031 "blocks": 69632, 00:14:36.031 "percent": 54 00:14:36.031 } 00:14:36.031 }, 00:14:36.031 "base_bdevs_list": [ 00:14:36.031 { 00:14:36.031 "name": "spare", 00:14:36.031 "uuid": "a8e64158-5cf7-5b2b-8209-ce201a4b16d0", 00:14:36.031 "is_configured": true, 00:14:36.031 "data_offset": 2048, 00:14:36.031 "data_size": 63488 00:14:36.031 }, 00:14:36.031 { 00:14:36.031 "name": "BaseBdev2", 00:14:36.031 "uuid": "de33ca6b-8e71-5d0e-874a-8324ac844b6e", 00:14:36.031 "is_configured": true, 00:14:36.031 "data_offset": 2048, 00:14:36.031 "data_size": 63488 00:14:36.031 }, 00:14:36.031 { 00:14:36.031 "name": "BaseBdev3", 00:14:36.031 "uuid": "44977377-4710-517f-8a52-2efd69972bbe", 00:14:36.031 "is_configured": true, 00:14:36.031 "data_offset": 2048, 00:14:36.031 "data_size": 63488 00:14:36.031 } 00:14:36.031 ] 00:14:36.031 }' 00:14:36.031 17:36:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:36.031 17:36:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:36.031 17:36:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:36.031 17:36:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:36.031 17:36:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:37.414 17:36:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:37.414 17:36:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:37.414 17:36:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:37.414 17:36:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:37.414 17:36:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:37.415 17:36:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:37.415 17:36:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:37.415 17:36:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:37.415 17:36:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:37.415 17:36:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:37.415 17:36:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:37.415 17:36:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:37.415 "name": "raid_bdev1", 00:14:37.415 "uuid": "99f3834e-db84-45f3-9a79-784e865c5710", 00:14:37.415 "strip_size_kb": 64, 00:14:37.415 "state": "online", 00:14:37.415 "raid_level": "raid5f", 00:14:37.415 "superblock": true, 00:14:37.415 "num_base_bdevs": 3, 00:14:37.415 "num_base_bdevs_discovered": 3, 00:14:37.415 "num_base_bdevs_operational": 3, 00:14:37.415 "process": { 00:14:37.415 "type": "rebuild", 00:14:37.415 "target": "spare", 00:14:37.415 "progress": { 00:14:37.415 "blocks": 94208, 00:14:37.415 "percent": 74 00:14:37.415 } 00:14:37.415 }, 00:14:37.415 "base_bdevs_list": [ 00:14:37.415 { 00:14:37.415 "name": "spare", 00:14:37.415 "uuid": "a8e64158-5cf7-5b2b-8209-ce201a4b16d0", 00:14:37.415 "is_configured": true, 00:14:37.415 "data_offset": 2048, 00:14:37.415 "data_size": 63488 00:14:37.415 }, 00:14:37.415 { 00:14:37.415 "name": "BaseBdev2", 00:14:37.415 "uuid": "de33ca6b-8e71-5d0e-874a-8324ac844b6e", 00:14:37.415 "is_configured": true, 00:14:37.415 "data_offset": 2048, 00:14:37.415 "data_size": 63488 00:14:37.415 }, 00:14:37.415 { 00:14:37.415 "name": "BaseBdev3", 00:14:37.415 "uuid": "44977377-4710-517f-8a52-2efd69972bbe", 00:14:37.415 "is_configured": true, 00:14:37.415 "data_offset": 2048, 00:14:37.415 "data_size": 63488 00:14:37.415 } 00:14:37.415 ] 00:14:37.415 }' 00:14:37.415 17:36:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:37.415 17:36:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:37.415 17:36:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:37.415 17:36:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:37.415 17:36:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:38.354 17:36:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:38.354 17:36:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:38.354 17:36:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:38.354 17:36:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:38.354 17:36:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:38.354 17:36:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:38.354 17:36:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:38.354 17:36:09 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:38.354 17:36:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:38.354 17:36:09 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:38.354 17:36:09 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:38.354 17:36:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:38.354 "name": "raid_bdev1", 00:14:38.354 "uuid": "99f3834e-db84-45f3-9a79-784e865c5710", 00:14:38.354 "strip_size_kb": 64, 00:14:38.354 "state": "online", 00:14:38.354 "raid_level": "raid5f", 00:14:38.354 "superblock": true, 00:14:38.354 "num_base_bdevs": 3, 00:14:38.354 "num_base_bdevs_discovered": 3, 00:14:38.354 "num_base_bdevs_operational": 3, 00:14:38.354 "process": { 00:14:38.354 "type": "rebuild", 00:14:38.354 "target": "spare", 00:14:38.354 "progress": { 00:14:38.354 "blocks": 116736, 00:14:38.354 "percent": 91 00:14:38.354 } 00:14:38.354 }, 00:14:38.354 "base_bdevs_list": [ 00:14:38.354 { 00:14:38.354 "name": "spare", 00:14:38.354 "uuid": "a8e64158-5cf7-5b2b-8209-ce201a4b16d0", 00:14:38.354 "is_configured": true, 00:14:38.354 "data_offset": 2048, 00:14:38.354 "data_size": 63488 00:14:38.354 }, 00:14:38.354 { 00:14:38.354 "name": "BaseBdev2", 00:14:38.354 "uuid": "de33ca6b-8e71-5d0e-874a-8324ac844b6e", 00:14:38.354 "is_configured": true, 00:14:38.354 "data_offset": 2048, 00:14:38.354 "data_size": 63488 00:14:38.354 }, 00:14:38.354 { 00:14:38.354 "name": "BaseBdev3", 00:14:38.354 "uuid": "44977377-4710-517f-8a52-2efd69972bbe", 00:14:38.354 "is_configured": true, 00:14:38.354 "data_offset": 2048, 00:14:38.354 "data_size": 63488 00:14:38.354 } 00:14:38.354 ] 00:14:38.354 }' 00:14:38.354 17:36:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:38.354 17:36:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:38.354 17:36:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:38.354 17:36:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:38.354 17:36:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:14:38.613 [2024-11-27 17:36:09.776422] bdev_raid.c:2896:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:14:38.613 [2024-11-27 17:36:09.776481] bdev_raid.c:2558:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:14:38.613 [2024-11-27 17:36:09.776595] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:39.553 17:36:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:14:39.553 17:36:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:39.553 17:36:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:39.553 17:36:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:39.553 17:36:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:39.553 17:36:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:39.553 17:36:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:39.553 17:36:10 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:39.553 17:36:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:39.553 17:36:10 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:39.553 17:36:10 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:39.553 17:36:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:39.553 "name": "raid_bdev1", 00:14:39.553 "uuid": "99f3834e-db84-45f3-9a79-784e865c5710", 00:14:39.553 "strip_size_kb": 64, 00:14:39.553 "state": "online", 00:14:39.553 "raid_level": "raid5f", 00:14:39.553 "superblock": true, 00:14:39.553 "num_base_bdevs": 3, 00:14:39.553 "num_base_bdevs_discovered": 3, 00:14:39.553 "num_base_bdevs_operational": 3, 00:14:39.553 "base_bdevs_list": [ 00:14:39.553 { 00:14:39.553 "name": "spare", 00:14:39.553 "uuid": "a8e64158-5cf7-5b2b-8209-ce201a4b16d0", 00:14:39.553 "is_configured": true, 00:14:39.553 "data_offset": 2048, 00:14:39.553 "data_size": 63488 00:14:39.553 }, 00:14:39.553 { 00:14:39.553 "name": "BaseBdev2", 00:14:39.553 "uuid": "de33ca6b-8e71-5d0e-874a-8324ac844b6e", 00:14:39.553 "is_configured": true, 00:14:39.553 "data_offset": 2048, 00:14:39.553 "data_size": 63488 00:14:39.553 }, 00:14:39.553 { 00:14:39.553 "name": "BaseBdev3", 00:14:39.553 "uuid": "44977377-4710-517f-8a52-2efd69972bbe", 00:14:39.553 "is_configured": true, 00:14:39.553 "data_offset": 2048, 00:14:39.553 "data_size": 63488 00:14:39.553 } 00:14:39.553 ] 00:14:39.553 }' 00:14:39.553 17:36:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:39.553 17:36:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:14:39.553 17:36:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:39.553 17:36:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:14:39.553 17:36:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@709 -- # break 00:14:39.553 17:36:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:14:39.553 17:36:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:39.553 17:36:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:14:39.553 17:36:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:14:39.553 17:36:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:39.553 17:36:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:39.553 17:36:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:39.553 17:36:10 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:39.553 17:36:10 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:39.553 17:36:10 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:39.553 17:36:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:39.553 "name": "raid_bdev1", 00:14:39.553 "uuid": "99f3834e-db84-45f3-9a79-784e865c5710", 00:14:39.553 "strip_size_kb": 64, 00:14:39.553 "state": "online", 00:14:39.553 "raid_level": "raid5f", 00:14:39.553 "superblock": true, 00:14:39.553 "num_base_bdevs": 3, 00:14:39.553 "num_base_bdevs_discovered": 3, 00:14:39.553 "num_base_bdevs_operational": 3, 00:14:39.553 "base_bdevs_list": [ 00:14:39.553 { 00:14:39.553 "name": "spare", 00:14:39.553 "uuid": "a8e64158-5cf7-5b2b-8209-ce201a4b16d0", 00:14:39.553 "is_configured": true, 00:14:39.553 "data_offset": 2048, 00:14:39.553 "data_size": 63488 00:14:39.553 }, 00:14:39.553 { 00:14:39.553 "name": "BaseBdev2", 00:14:39.553 "uuid": "de33ca6b-8e71-5d0e-874a-8324ac844b6e", 00:14:39.553 "is_configured": true, 00:14:39.553 "data_offset": 2048, 00:14:39.553 "data_size": 63488 00:14:39.553 }, 00:14:39.553 { 00:14:39.553 "name": "BaseBdev3", 00:14:39.553 "uuid": "44977377-4710-517f-8a52-2efd69972bbe", 00:14:39.553 "is_configured": true, 00:14:39.553 "data_offset": 2048, 00:14:39.553 "data_size": 63488 00:14:39.553 } 00:14:39.553 ] 00:14:39.553 }' 00:14:39.553 17:36:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:39.553 17:36:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:14:39.553 17:36:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:39.553 17:36:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:14:39.553 17:36:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:14:39.553 17:36:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:39.553 17:36:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:39.553 17:36:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:39.553 17:36:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:39.553 17:36:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:39.553 17:36:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:39.553 17:36:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:39.553 17:36:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:39.553 17:36:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:39.553 17:36:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:39.553 17:36:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:39.553 17:36:10 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:39.554 17:36:10 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:39.554 17:36:10 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:39.554 17:36:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:39.554 "name": "raid_bdev1", 00:14:39.554 "uuid": "99f3834e-db84-45f3-9a79-784e865c5710", 00:14:39.554 "strip_size_kb": 64, 00:14:39.554 "state": "online", 00:14:39.554 "raid_level": "raid5f", 00:14:39.554 "superblock": true, 00:14:39.554 "num_base_bdevs": 3, 00:14:39.554 "num_base_bdevs_discovered": 3, 00:14:39.554 "num_base_bdevs_operational": 3, 00:14:39.554 "base_bdevs_list": [ 00:14:39.554 { 00:14:39.554 "name": "spare", 00:14:39.554 "uuid": "a8e64158-5cf7-5b2b-8209-ce201a4b16d0", 00:14:39.554 "is_configured": true, 00:14:39.554 "data_offset": 2048, 00:14:39.554 "data_size": 63488 00:14:39.554 }, 00:14:39.554 { 00:14:39.554 "name": "BaseBdev2", 00:14:39.554 "uuid": "de33ca6b-8e71-5d0e-874a-8324ac844b6e", 00:14:39.554 "is_configured": true, 00:14:39.554 "data_offset": 2048, 00:14:39.554 "data_size": 63488 00:14:39.554 }, 00:14:39.554 { 00:14:39.554 "name": "BaseBdev3", 00:14:39.554 "uuid": "44977377-4710-517f-8a52-2efd69972bbe", 00:14:39.554 "is_configured": true, 00:14:39.554 "data_offset": 2048, 00:14:39.554 "data_size": 63488 00:14:39.554 } 00:14:39.554 ] 00:14:39.554 }' 00:14:39.554 17:36:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:39.554 17:36:10 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:40.132 17:36:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:14:40.132 17:36:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:40.132 17:36:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:40.132 [2024-11-27 17:36:11.151236] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:14:40.132 [2024-11-27 17:36:11.151317] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:14:40.132 [2024-11-27 17:36:11.151408] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:40.132 [2024-11-27 17:36:11.151497] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:14:40.132 [2024-11-27 17:36:11.151543] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:14:40.132 17:36:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:40.132 17:36:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:40.132 17:36:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # jq length 00:14:40.132 17:36:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:40.132 17:36:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:40.132 17:36:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:40.132 17:36:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:14:40.132 17:36:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:14:40.132 17:36:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:14:40.132 17:36:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:14:40.132 17:36:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:14:40.132 17:36:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:14:40.132 17:36:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # local bdev_list 00:14:40.132 17:36:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:14:40.132 17:36:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # local nbd_list 00:14:40.133 17:36:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@12 -- # local i 00:14:40.133 17:36:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:14:40.133 17:36:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:14:40.133 17:36:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:14:40.414 /dev/nbd0 00:14:40.414 17:36:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:14:40.414 17:36:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:14:40.414 17:36:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:14:40.414 17:36:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@869 -- # local i 00:14:40.414 17:36:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:14:40.414 17:36:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:14:40.414 17:36:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:14:40.414 17:36:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@873 -- # break 00:14:40.414 17:36:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:14:40.414 17:36:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:14:40.414 17:36:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:14:40.414 1+0 records in 00:14:40.414 1+0 records out 00:14:40.414 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000263538 s, 15.5 MB/s 00:14:40.414 17:36:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:40.414 17:36:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@886 -- # size=4096 00:14:40.414 17:36:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:40.414 17:36:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:14:40.414 17:36:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@889 -- # return 0 00:14:40.414 17:36:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:14:40.414 17:36:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:14:40.414 17:36:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:14:40.699 /dev/nbd1 00:14:40.699 17:36:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:14:40.699 17:36:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:14:40.699 17:36:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@868 -- # local nbd_name=nbd1 00:14:40.699 17:36:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@869 -- # local i 00:14:40.699 17:36:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:14:40.699 17:36:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:14:40.699 17:36:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@872 -- # grep -q -w nbd1 /proc/partitions 00:14:40.699 17:36:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@873 -- # break 00:14:40.699 17:36:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:14:40.699 17:36:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:14:40.699 17:36:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@885 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:14:40.699 1+0 records in 00:14:40.699 1+0 records out 00:14:40.699 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000458722 s, 8.9 MB/s 00:14:40.699 17:36:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:40.699 17:36:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@886 -- # size=4096 00:14:40.699 17:36:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:14:40.699 17:36:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:14:40.699 17:36:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@889 -- # return 0 00:14:40.699 17:36:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:14:40.699 17:36:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:14:40.699 17:36:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@738 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:14:40.699 17:36:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:14:40.699 17:36:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:14:40.699 17:36:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:14:40.699 17:36:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # local nbd_list 00:14:40.699 17:36:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@51 -- # local i 00:14:40.699 17:36:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:14:40.699 17:36:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:14:40.974 17:36:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:14:40.974 17:36:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:14:40.974 17:36:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:14:40.974 17:36:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:14:40.974 17:36:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:14:40.974 17:36:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:14:40.974 17:36:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:14:40.974 17:36:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:14:40.974 17:36:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:14:40.974 17:36:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:14:41.234 17:36:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:14:41.234 17:36:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:14:41.234 17:36:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:14:41.234 17:36:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:14:41.234 17:36:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:14:41.234 17:36:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:14:41.234 17:36:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:14:41.234 17:36:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:14:41.234 17:36:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:14:41.235 17:36:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:14:41.235 17:36:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:41.235 17:36:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:41.235 17:36:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:41.235 17:36:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:14:41.235 17:36:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:41.235 17:36:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:41.235 [2024-11-27 17:36:12.211927] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:14:41.235 [2024-11-27 17:36:12.212034] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:41.235 [2024-11-27 17:36:12.212062] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:14:41.235 [2024-11-27 17:36:12.212072] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:41.235 [2024-11-27 17:36:12.214095] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:41.235 [2024-11-27 17:36:12.214133] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:14:41.235 [2024-11-27 17:36:12.214215] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:14:41.235 [2024-11-27 17:36:12.214258] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:14:41.235 [2024-11-27 17:36:12.214370] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:14:41.235 [2024-11-27 17:36:12.214466] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:14:41.235 spare 00:14:41.235 17:36:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:41.235 17:36:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:14:41.235 17:36:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:41.235 17:36:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:41.235 [2024-11-27 17:36:12.314358] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001580 00:14:41.235 [2024-11-27 17:36:12.314427] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 126976, blocklen 512 00:14:41.235 [2024-11-27 17:36:12.314671] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000043d50 00:14:41.235 [2024-11-27 17:36:12.315087] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001580 00:14:41.235 [2024-11-27 17:36:12.315107] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001580 00:14:41.235 [2024-11-27 17:36:12.315243] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:41.235 17:36:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:41.235 17:36:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:14:41.235 17:36:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:41.235 17:36:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:41.235 17:36:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:41.235 17:36:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:41.235 17:36:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:41.235 17:36:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:41.235 17:36:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:41.235 17:36:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:41.235 17:36:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:41.235 17:36:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:41.235 17:36:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:41.235 17:36:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:41.235 17:36:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:41.235 17:36:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:41.235 17:36:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:41.235 "name": "raid_bdev1", 00:14:41.235 "uuid": "99f3834e-db84-45f3-9a79-784e865c5710", 00:14:41.235 "strip_size_kb": 64, 00:14:41.235 "state": "online", 00:14:41.235 "raid_level": "raid5f", 00:14:41.235 "superblock": true, 00:14:41.235 "num_base_bdevs": 3, 00:14:41.235 "num_base_bdevs_discovered": 3, 00:14:41.235 "num_base_bdevs_operational": 3, 00:14:41.235 "base_bdevs_list": [ 00:14:41.235 { 00:14:41.235 "name": "spare", 00:14:41.235 "uuid": "a8e64158-5cf7-5b2b-8209-ce201a4b16d0", 00:14:41.235 "is_configured": true, 00:14:41.235 "data_offset": 2048, 00:14:41.235 "data_size": 63488 00:14:41.235 }, 00:14:41.235 { 00:14:41.235 "name": "BaseBdev2", 00:14:41.235 "uuid": "de33ca6b-8e71-5d0e-874a-8324ac844b6e", 00:14:41.235 "is_configured": true, 00:14:41.235 "data_offset": 2048, 00:14:41.235 "data_size": 63488 00:14:41.235 }, 00:14:41.235 { 00:14:41.235 "name": "BaseBdev3", 00:14:41.235 "uuid": "44977377-4710-517f-8a52-2efd69972bbe", 00:14:41.235 "is_configured": true, 00:14:41.235 "data_offset": 2048, 00:14:41.235 "data_size": 63488 00:14:41.235 } 00:14:41.235 ] 00:14:41.235 }' 00:14:41.235 17:36:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:41.235 17:36:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:41.806 17:36:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:14:41.806 17:36:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:41.806 17:36:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:14:41.806 17:36:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:14:41.806 17:36:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:41.806 17:36:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:41.806 17:36:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:41.806 17:36:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:41.806 17:36:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:41.806 17:36:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:41.806 17:36:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:41.806 "name": "raid_bdev1", 00:14:41.806 "uuid": "99f3834e-db84-45f3-9a79-784e865c5710", 00:14:41.806 "strip_size_kb": 64, 00:14:41.806 "state": "online", 00:14:41.806 "raid_level": "raid5f", 00:14:41.806 "superblock": true, 00:14:41.806 "num_base_bdevs": 3, 00:14:41.806 "num_base_bdevs_discovered": 3, 00:14:41.806 "num_base_bdevs_operational": 3, 00:14:41.806 "base_bdevs_list": [ 00:14:41.806 { 00:14:41.806 "name": "spare", 00:14:41.806 "uuid": "a8e64158-5cf7-5b2b-8209-ce201a4b16d0", 00:14:41.806 "is_configured": true, 00:14:41.806 "data_offset": 2048, 00:14:41.806 "data_size": 63488 00:14:41.806 }, 00:14:41.806 { 00:14:41.806 "name": "BaseBdev2", 00:14:41.806 "uuid": "de33ca6b-8e71-5d0e-874a-8324ac844b6e", 00:14:41.807 "is_configured": true, 00:14:41.807 "data_offset": 2048, 00:14:41.807 "data_size": 63488 00:14:41.807 }, 00:14:41.807 { 00:14:41.807 "name": "BaseBdev3", 00:14:41.807 "uuid": "44977377-4710-517f-8a52-2efd69972bbe", 00:14:41.807 "is_configured": true, 00:14:41.807 "data_offset": 2048, 00:14:41.807 "data_size": 63488 00:14:41.807 } 00:14:41.807 ] 00:14:41.807 }' 00:14:41.807 17:36:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:41.807 17:36:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:14:41.807 17:36:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:41.807 17:36:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:14:41.807 17:36:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:41.807 17:36:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:14:41.807 17:36:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:41.807 17:36:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:41.807 17:36:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:41.807 17:36:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:14:41.807 17:36:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:14:41.807 17:36:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:41.807 17:36:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:41.807 [2024-11-27 17:36:12.979281] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:14:41.807 17:36:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:41.807 17:36:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:14:41.807 17:36:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:41.807 17:36:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:41.807 17:36:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:41.807 17:36:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:41.807 17:36:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:14:41.807 17:36:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:41.807 17:36:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:41.807 17:36:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:41.807 17:36:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:41.807 17:36:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:41.807 17:36:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:41.807 17:36:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:41.807 17:36:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:42.067 17:36:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:42.067 17:36:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:42.067 "name": "raid_bdev1", 00:14:42.067 "uuid": "99f3834e-db84-45f3-9a79-784e865c5710", 00:14:42.067 "strip_size_kb": 64, 00:14:42.067 "state": "online", 00:14:42.067 "raid_level": "raid5f", 00:14:42.067 "superblock": true, 00:14:42.067 "num_base_bdevs": 3, 00:14:42.067 "num_base_bdevs_discovered": 2, 00:14:42.067 "num_base_bdevs_operational": 2, 00:14:42.067 "base_bdevs_list": [ 00:14:42.067 { 00:14:42.067 "name": null, 00:14:42.067 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:42.067 "is_configured": false, 00:14:42.067 "data_offset": 0, 00:14:42.067 "data_size": 63488 00:14:42.067 }, 00:14:42.067 { 00:14:42.067 "name": "BaseBdev2", 00:14:42.067 "uuid": "de33ca6b-8e71-5d0e-874a-8324ac844b6e", 00:14:42.067 "is_configured": true, 00:14:42.067 "data_offset": 2048, 00:14:42.067 "data_size": 63488 00:14:42.067 }, 00:14:42.067 { 00:14:42.067 "name": "BaseBdev3", 00:14:42.067 "uuid": "44977377-4710-517f-8a52-2efd69972bbe", 00:14:42.067 "is_configured": true, 00:14:42.067 "data_offset": 2048, 00:14:42.067 "data_size": 63488 00:14:42.067 } 00:14:42.067 ] 00:14:42.067 }' 00:14:42.067 17:36:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:42.067 17:36:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:42.327 17:36:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:14:42.327 17:36:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:42.327 17:36:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:42.327 [2024-11-27 17:36:13.486755] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:14:42.327 [2024-11-27 17:36:13.486989] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:14:42.327 [2024-11-27 17:36:13.487008] bdev_raid.c:3748:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:14:42.327 [2024-11-27 17:36:13.487047] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:14:42.327 [2024-11-27 17:36:13.490592] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000043e20 00:14:42.327 [2024-11-27 17:36:13.492621] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:14:42.327 17:36:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:42.327 17:36:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@757 -- # sleep 1 00:14:43.708 17:36:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:43.708 17:36:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:43.708 17:36:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:43.708 17:36:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:43.708 17:36:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:43.708 17:36:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:43.708 17:36:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:43.708 17:36:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:43.708 17:36:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:43.708 17:36:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:43.708 17:36:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:43.708 "name": "raid_bdev1", 00:14:43.708 "uuid": "99f3834e-db84-45f3-9a79-784e865c5710", 00:14:43.708 "strip_size_kb": 64, 00:14:43.708 "state": "online", 00:14:43.708 "raid_level": "raid5f", 00:14:43.708 "superblock": true, 00:14:43.708 "num_base_bdevs": 3, 00:14:43.708 "num_base_bdevs_discovered": 3, 00:14:43.708 "num_base_bdevs_operational": 3, 00:14:43.708 "process": { 00:14:43.708 "type": "rebuild", 00:14:43.708 "target": "spare", 00:14:43.708 "progress": { 00:14:43.708 "blocks": 20480, 00:14:43.708 "percent": 16 00:14:43.708 } 00:14:43.708 }, 00:14:43.708 "base_bdevs_list": [ 00:14:43.708 { 00:14:43.708 "name": "spare", 00:14:43.708 "uuid": "a8e64158-5cf7-5b2b-8209-ce201a4b16d0", 00:14:43.708 "is_configured": true, 00:14:43.708 "data_offset": 2048, 00:14:43.708 "data_size": 63488 00:14:43.708 }, 00:14:43.708 { 00:14:43.708 "name": "BaseBdev2", 00:14:43.708 "uuid": "de33ca6b-8e71-5d0e-874a-8324ac844b6e", 00:14:43.708 "is_configured": true, 00:14:43.708 "data_offset": 2048, 00:14:43.708 "data_size": 63488 00:14:43.708 }, 00:14:43.708 { 00:14:43.708 "name": "BaseBdev3", 00:14:43.708 "uuid": "44977377-4710-517f-8a52-2efd69972bbe", 00:14:43.708 "is_configured": true, 00:14:43.708 "data_offset": 2048, 00:14:43.708 "data_size": 63488 00:14:43.708 } 00:14:43.708 ] 00:14:43.708 }' 00:14:43.708 17:36:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:43.708 17:36:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:43.708 17:36:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:43.708 17:36:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:43.708 17:36:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:14:43.708 17:36:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:43.708 17:36:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:43.708 [2024-11-27 17:36:14.655294] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:14:43.709 [2024-11-27 17:36:14.699026] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:14:43.709 [2024-11-27 17:36:14.699127] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:43.709 [2024-11-27 17:36:14.699209] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:14:43.709 [2024-11-27 17:36:14.699231] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:14:43.709 17:36:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:43.709 17:36:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:14:43.709 17:36:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:43.709 17:36:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:43.709 17:36:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:43.709 17:36:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:43.709 17:36:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:14:43.709 17:36:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:43.709 17:36:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:43.709 17:36:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:43.709 17:36:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:43.709 17:36:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:43.709 17:36:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:43.709 17:36:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:43.709 17:36:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:43.709 17:36:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:43.709 17:36:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:43.709 "name": "raid_bdev1", 00:14:43.709 "uuid": "99f3834e-db84-45f3-9a79-784e865c5710", 00:14:43.709 "strip_size_kb": 64, 00:14:43.709 "state": "online", 00:14:43.709 "raid_level": "raid5f", 00:14:43.709 "superblock": true, 00:14:43.709 "num_base_bdevs": 3, 00:14:43.709 "num_base_bdevs_discovered": 2, 00:14:43.709 "num_base_bdevs_operational": 2, 00:14:43.709 "base_bdevs_list": [ 00:14:43.709 { 00:14:43.709 "name": null, 00:14:43.709 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:43.709 "is_configured": false, 00:14:43.709 "data_offset": 0, 00:14:43.709 "data_size": 63488 00:14:43.709 }, 00:14:43.709 { 00:14:43.709 "name": "BaseBdev2", 00:14:43.709 "uuid": "de33ca6b-8e71-5d0e-874a-8324ac844b6e", 00:14:43.709 "is_configured": true, 00:14:43.709 "data_offset": 2048, 00:14:43.709 "data_size": 63488 00:14:43.709 }, 00:14:43.709 { 00:14:43.709 "name": "BaseBdev3", 00:14:43.709 "uuid": "44977377-4710-517f-8a52-2efd69972bbe", 00:14:43.709 "is_configured": true, 00:14:43.709 "data_offset": 2048, 00:14:43.709 "data_size": 63488 00:14:43.709 } 00:14:43.709 ] 00:14:43.709 }' 00:14:43.709 17:36:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:43.709 17:36:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:43.968 17:36:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:14:43.968 17:36:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:43.969 17:36:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:43.969 [2024-11-27 17:36:15.107421] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:14:43.969 [2024-11-27 17:36:15.107474] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:43.969 [2024-11-27 17:36:15.107495] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ab80 00:14:43.969 [2024-11-27 17:36:15.107504] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:43.969 [2024-11-27 17:36:15.107932] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:43.969 [2024-11-27 17:36:15.107949] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:14:43.969 [2024-11-27 17:36:15.108025] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:14:43.969 [2024-11-27 17:36:15.108036] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:14:43.969 [2024-11-27 17:36:15.108046] bdev_raid.c:3748:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:14:43.969 [2024-11-27 17:36:15.108071] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:14:43.969 spare 00:14:43.969 [2024-11-27 17:36:15.110688] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000043ef0 00:14:43.969 [2024-11-27 17:36:15.112694] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:14:43.969 17:36:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:43.969 17:36:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@764 -- # sleep 1 00:14:45.350 17:36:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:14:45.350 17:36:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:45.350 17:36:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:14:45.350 17:36:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:14:45.350 17:36:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:45.350 17:36:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:45.350 17:36:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:45.350 17:36:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:45.350 17:36:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:45.350 17:36:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:45.350 17:36:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:45.350 "name": "raid_bdev1", 00:14:45.350 "uuid": "99f3834e-db84-45f3-9a79-784e865c5710", 00:14:45.350 "strip_size_kb": 64, 00:14:45.350 "state": "online", 00:14:45.350 "raid_level": "raid5f", 00:14:45.350 "superblock": true, 00:14:45.350 "num_base_bdevs": 3, 00:14:45.350 "num_base_bdevs_discovered": 3, 00:14:45.350 "num_base_bdevs_operational": 3, 00:14:45.350 "process": { 00:14:45.350 "type": "rebuild", 00:14:45.350 "target": "spare", 00:14:45.350 "progress": { 00:14:45.350 "blocks": 20480, 00:14:45.350 "percent": 16 00:14:45.350 } 00:14:45.350 }, 00:14:45.350 "base_bdevs_list": [ 00:14:45.350 { 00:14:45.350 "name": "spare", 00:14:45.350 "uuid": "a8e64158-5cf7-5b2b-8209-ce201a4b16d0", 00:14:45.350 "is_configured": true, 00:14:45.350 "data_offset": 2048, 00:14:45.350 "data_size": 63488 00:14:45.350 }, 00:14:45.350 { 00:14:45.350 "name": "BaseBdev2", 00:14:45.350 "uuid": "de33ca6b-8e71-5d0e-874a-8324ac844b6e", 00:14:45.350 "is_configured": true, 00:14:45.350 "data_offset": 2048, 00:14:45.350 "data_size": 63488 00:14:45.350 }, 00:14:45.350 { 00:14:45.350 "name": "BaseBdev3", 00:14:45.350 "uuid": "44977377-4710-517f-8a52-2efd69972bbe", 00:14:45.350 "is_configured": true, 00:14:45.350 "data_offset": 2048, 00:14:45.350 "data_size": 63488 00:14:45.350 } 00:14:45.350 ] 00:14:45.350 }' 00:14:45.350 17:36:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:45.350 17:36:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:14:45.350 17:36:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:45.350 17:36:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:14:45.350 17:36:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:14:45.350 17:36:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:45.350 17:36:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:45.350 [2024-11-27 17:36:16.275341] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:14:45.350 [2024-11-27 17:36:16.319233] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:14:45.350 [2024-11-27 17:36:16.319292] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:45.350 [2024-11-27 17:36:16.319307] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:14:45.350 [2024-11-27 17:36:16.319318] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:14:45.350 17:36:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:45.350 17:36:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:14:45.350 17:36:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:45.350 17:36:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:45.350 17:36:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:45.350 17:36:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:45.350 17:36:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:14:45.350 17:36:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:45.350 17:36:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:45.350 17:36:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:45.350 17:36:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:45.350 17:36:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:45.350 17:36:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:45.350 17:36:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:45.350 17:36:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:45.350 17:36:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:45.350 17:36:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:45.350 "name": "raid_bdev1", 00:14:45.350 "uuid": "99f3834e-db84-45f3-9a79-784e865c5710", 00:14:45.350 "strip_size_kb": 64, 00:14:45.350 "state": "online", 00:14:45.350 "raid_level": "raid5f", 00:14:45.350 "superblock": true, 00:14:45.350 "num_base_bdevs": 3, 00:14:45.350 "num_base_bdevs_discovered": 2, 00:14:45.350 "num_base_bdevs_operational": 2, 00:14:45.350 "base_bdevs_list": [ 00:14:45.350 { 00:14:45.350 "name": null, 00:14:45.350 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:45.350 "is_configured": false, 00:14:45.350 "data_offset": 0, 00:14:45.350 "data_size": 63488 00:14:45.350 }, 00:14:45.350 { 00:14:45.350 "name": "BaseBdev2", 00:14:45.350 "uuid": "de33ca6b-8e71-5d0e-874a-8324ac844b6e", 00:14:45.350 "is_configured": true, 00:14:45.350 "data_offset": 2048, 00:14:45.350 "data_size": 63488 00:14:45.350 }, 00:14:45.350 { 00:14:45.350 "name": "BaseBdev3", 00:14:45.350 "uuid": "44977377-4710-517f-8a52-2efd69972bbe", 00:14:45.350 "is_configured": true, 00:14:45.350 "data_offset": 2048, 00:14:45.350 "data_size": 63488 00:14:45.350 } 00:14:45.350 ] 00:14:45.350 }' 00:14:45.350 17:36:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:45.350 17:36:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:45.609 17:36:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:14:45.609 17:36:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:45.609 17:36:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:14:45.609 17:36:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:14:45.609 17:36:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:45.609 17:36:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:45.609 17:36:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:45.609 17:36:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:45.609 17:36:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:45.869 17:36:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:45.869 17:36:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:45.869 "name": "raid_bdev1", 00:14:45.869 "uuid": "99f3834e-db84-45f3-9a79-784e865c5710", 00:14:45.869 "strip_size_kb": 64, 00:14:45.869 "state": "online", 00:14:45.869 "raid_level": "raid5f", 00:14:45.869 "superblock": true, 00:14:45.869 "num_base_bdevs": 3, 00:14:45.869 "num_base_bdevs_discovered": 2, 00:14:45.869 "num_base_bdevs_operational": 2, 00:14:45.869 "base_bdevs_list": [ 00:14:45.869 { 00:14:45.869 "name": null, 00:14:45.869 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:45.869 "is_configured": false, 00:14:45.869 "data_offset": 0, 00:14:45.869 "data_size": 63488 00:14:45.869 }, 00:14:45.869 { 00:14:45.869 "name": "BaseBdev2", 00:14:45.869 "uuid": "de33ca6b-8e71-5d0e-874a-8324ac844b6e", 00:14:45.869 "is_configured": true, 00:14:45.869 "data_offset": 2048, 00:14:45.869 "data_size": 63488 00:14:45.869 }, 00:14:45.869 { 00:14:45.869 "name": "BaseBdev3", 00:14:45.869 "uuid": "44977377-4710-517f-8a52-2efd69972bbe", 00:14:45.869 "is_configured": true, 00:14:45.869 "data_offset": 2048, 00:14:45.869 "data_size": 63488 00:14:45.869 } 00:14:45.869 ] 00:14:45.869 }' 00:14:45.869 17:36:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:45.869 17:36:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:14:45.869 17:36:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:45.869 17:36:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:14:45.869 17:36:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:14:45.869 17:36:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:45.869 17:36:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:45.869 17:36:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:45.869 17:36:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:14:45.869 17:36:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:45.869 17:36:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:45.869 [2024-11-27 17:36:16.919228] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:14:45.869 [2024-11-27 17:36:16.919283] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:14:45.869 [2024-11-27 17:36:16.919307] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b180 00:14:45.869 [2024-11-27 17:36:16.919320] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:14:45.869 [2024-11-27 17:36:16.919708] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:14:45.869 [2024-11-27 17:36:16.919726] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:14:45.869 [2024-11-27 17:36:16.919788] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:14:45.869 [2024-11-27 17:36:16.919811] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:14:45.869 [2024-11-27 17:36:16.919819] bdev_raid.c:3709:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:14:45.869 [2024-11-27 17:36:16.919829] bdev_raid.c:3884:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:14:45.869 BaseBdev1 00:14:45.869 17:36:16 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:45.869 17:36:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@775 -- # sleep 1 00:14:46.808 17:36:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:14:46.808 17:36:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:46.808 17:36:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:46.808 17:36:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:46.808 17:36:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:46.808 17:36:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:14:46.808 17:36:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:46.808 17:36:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:46.808 17:36:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:46.808 17:36:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:46.808 17:36:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:46.808 17:36:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:46.808 17:36:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:46.808 17:36:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:46.808 17:36:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:46.808 17:36:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:46.808 "name": "raid_bdev1", 00:14:46.808 "uuid": "99f3834e-db84-45f3-9a79-784e865c5710", 00:14:46.808 "strip_size_kb": 64, 00:14:46.808 "state": "online", 00:14:46.808 "raid_level": "raid5f", 00:14:46.808 "superblock": true, 00:14:46.808 "num_base_bdevs": 3, 00:14:46.808 "num_base_bdevs_discovered": 2, 00:14:46.808 "num_base_bdevs_operational": 2, 00:14:46.808 "base_bdevs_list": [ 00:14:46.808 { 00:14:46.808 "name": null, 00:14:46.808 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:46.808 "is_configured": false, 00:14:46.808 "data_offset": 0, 00:14:46.808 "data_size": 63488 00:14:46.808 }, 00:14:46.808 { 00:14:46.808 "name": "BaseBdev2", 00:14:46.808 "uuid": "de33ca6b-8e71-5d0e-874a-8324ac844b6e", 00:14:46.808 "is_configured": true, 00:14:46.808 "data_offset": 2048, 00:14:46.808 "data_size": 63488 00:14:46.808 }, 00:14:46.808 { 00:14:46.808 "name": "BaseBdev3", 00:14:46.808 "uuid": "44977377-4710-517f-8a52-2efd69972bbe", 00:14:46.808 "is_configured": true, 00:14:46.808 "data_offset": 2048, 00:14:46.808 "data_size": 63488 00:14:46.808 } 00:14:46.808 ] 00:14:46.808 }' 00:14:46.808 17:36:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:46.808 17:36:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:47.376 17:36:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:14:47.376 17:36:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:47.376 17:36:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:14:47.376 17:36:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:14:47.376 17:36:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:47.376 17:36:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:47.376 17:36:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:47.376 17:36:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:47.376 17:36:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:47.376 17:36:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:47.376 17:36:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:47.376 "name": "raid_bdev1", 00:14:47.376 "uuid": "99f3834e-db84-45f3-9a79-784e865c5710", 00:14:47.376 "strip_size_kb": 64, 00:14:47.376 "state": "online", 00:14:47.376 "raid_level": "raid5f", 00:14:47.376 "superblock": true, 00:14:47.376 "num_base_bdevs": 3, 00:14:47.376 "num_base_bdevs_discovered": 2, 00:14:47.376 "num_base_bdevs_operational": 2, 00:14:47.376 "base_bdevs_list": [ 00:14:47.376 { 00:14:47.376 "name": null, 00:14:47.376 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:47.376 "is_configured": false, 00:14:47.376 "data_offset": 0, 00:14:47.376 "data_size": 63488 00:14:47.376 }, 00:14:47.376 { 00:14:47.376 "name": "BaseBdev2", 00:14:47.376 "uuid": "de33ca6b-8e71-5d0e-874a-8324ac844b6e", 00:14:47.376 "is_configured": true, 00:14:47.376 "data_offset": 2048, 00:14:47.376 "data_size": 63488 00:14:47.376 }, 00:14:47.376 { 00:14:47.376 "name": "BaseBdev3", 00:14:47.376 "uuid": "44977377-4710-517f-8a52-2efd69972bbe", 00:14:47.376 "is_configured": true, 00:14:47.376 "data_offset": 2048, 00:14:47.376 "data_size": 63488 00:14:47.376 } 00:14:47.376 ] 00:14:47.376 }' 00:14:47.376 17:36:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:47.376 17:36:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:14:47.376 17:36:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:47.376 17:36:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:14:47.376 17:36:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:14:47.376 17:36:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@650 -- # local es=0 00:14:47.376 17:36:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:14:47.376 17:36:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:14:47.376 17:36:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:14:47.376 17:36:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:14:47.376 17:36:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:14:47.376 17:36:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:14:47.376 17:36:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:47.376 17:36:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:47.376 [2024-11-27 17:36:18.552426] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:14:47.376 [2024-11-27 17:36:18.552614] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:14:47.376 [2024-11-27 17:36:18.552632] bdev_raid.c:3709:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:14:47.376 request: 00:14:47.376 { 00:14:47.376 "base_bdev": "BaseBdev1", 00:14:47.376 "raid_bdev": "raid_bdev1", 00:14:47.376 "method": "bdev_raid_add_base_bdev", 00:14:47.376 "req_id": 1 00:14:47.376 } 00:14:47.376 Got JSON-RPC error response 00:14:47.376 response: 00:14:47.376 { 00:14:47.376 "code": -22, 00:14:47.376 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:14:47.376 } 00:14:47.376 17:36:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:14:47.376 17:36:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@653 -- # es=1 00:14:47.376 17:36:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:14:47.376 17:36:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:14:47.376 17:36:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:14:47.376 17:36:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@779 -- # sleep 1 00:14:48.754 17:36:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 2 00:14:48.754 17:36:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:14:48.754 17:36:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:48.754 17:36:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:48.754 17:36:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:48.754 17:36:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:14:48.754 17:36:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:48.754 17:36:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:48.754 17:36:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:48.754 17:36:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:48.754 17:36:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:48.754 17:36:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:48.754 17:36:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:48.754 17:36:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:48.754 17:36:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:48.754 17:36:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:48.754 "name": "raid_bdev1", 00:14:48.754 "uuid": "99f3834e-db84-45f3-9a79-784e865c5710", 00:14:48.754 "strip_size_kb": 64, 00:14:48.754 "state": "online", 00:14:48.754 "raid_level": "raid5f", 00:14:48.754 "superblock": true, 00:14:48.754 "num_base_bdevs": 3, 00:14:48.754 "num_base_bdevs_discovered": 2, 00:14:48.754 "num_base_bdevs_operational": 2, 00:14:48.754 "base_bdevs_list": [ 00:14:48.754 { 00:14:48.754 "name": null, 00:14:48.754 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:48.754 "is_configured": false, 00:14:48.754 "data_offset": 0, 00:14:48.754 "data_size": 63488 00:14:48.754 }, 00:14:48.754 { 00:14:48.754 "name": "BaseBdev2", 00:14:48.754 "uuid": "de33ca6b-8e71-5d0e-874a-8324ac844b6e", 00:14:48.754 "is_configured": true, 00:14:48.754 "data_offset": 2048, 00:14:48.754 "data_size": 63488 00:14:48.754 }, 00:14:48.754 { 00:14:48.754 "name": "BaseBdev3", 00:14:48.754 "uuid": "44977377-4710-517f-8a52-2efd69972bbe", 00:14:48.754 "is_configured": true, 00:14:48.754 "data_offset": 2048, 00:14:48.754 "data_size": 63488 00:14:48.754 } 00:14:48.754 ] 00:14:48.754 }' 00:14:48.754 17:36:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:48.754 17:36:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:49.014 17:36:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:14:49.014 17:36:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:14:49.014 17:36:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:14:49.014 17:36:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:14:49.014 17:36:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:14:49.014 17:36:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:14:49.014 17:36:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:49.014 17:36:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:49.014 17:36:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:49.014 17:36:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:49.014 17:36:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:14:49.014 "name": "raid_bdev1", 00:14:49.014 "uuid": "99f3834e-db84-45f3-9a79-784e865c5710", 00:14:49.014 "strip_size_kb": 64, 00:14:49.014 "state": "online", 00:14:49.014 "raid_level": "raid5f", 00:14:49.014 "superblock": true, 00:14:49.014 "num_base_bdevs": 3, 00:14:49.014 "num_base_bdevs_discovered": 2, 00:14:49.014 "num_base_bdevs_operational": 2, 00:14:49.014 "base_bdevs_list": [ 00:14:49.014 { 00:14:49.014 "name": null, 00:14:49.014 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:49.014 "is_configured": false, 00:14:49.014 "data_offset": 0, 00:14:49.014 "data_size": 63488 00:14:49.014 }, 00:14:49.014 { 00:14:49.014 "name": "BaseBdev2", 00:14:49.014 "uuid": "de33ca6b-8e71-5d0e-874a-8324ac844b6e", 00:14:49.014 "is_configured": true, 00:14:49.014 "data_offset": 2048, 00:14:49.014 "data_size": 63488 00:14:49.014 }, 00:14:49.014 { 00:14:49.014 "name": "BaseBdev3", 00:14:49.014 "uuid": "44977377-4710-517f-8a52-2efd69972bbe", 00:14:49.014 "is_configured": true, 00:14:49.014 "data_offset": 2048, 00:14:49.014 "data_size": 63488 00:14:49.014 } 00:14:49.014 ] 00:14:49.014 }' 00:14:49.014 17:36:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:14:49.014 17:36:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:14:49.014 17:36:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:14:49.014 17:36:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:14:49.014 17:36:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@784 -- # killprocess 92370 00:14:49.014 17:36:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@950 -- # '[' -z 92370 ']' 00:14:49.014 17:36:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@954 -- # kill -0 92370 00:14:49.014 17:36:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@955 -- # uname 00:14:49.014 17:36:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:14:49.014 17:36:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 92370 00:14:49.014 17:36:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:14:49.014 17:36:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:14:49.014 killing process with pid 92370 00:14:49.014 Received shutdown signal, test time was about 60.000000 seconds 00:14:49.014 00:14:49.014 Latency(us) 00:14:49.014 [2024-11-27T17:36:20.206Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:14:49.014 [2024-11-27T17:36:20.206Z] =================================================================================================================== 00:14:49.014 [2024-11-27T17:36:20.206Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:14:49.014 17:36:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@968 -- # echo 'killing process with pid 92370' 00:14:49.014 17:36:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@969 -- # kill 92370 00:14:49.014 [2024-11-27 17:36:20.145236] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:14:49.014 [2024-11-27 17:36:20.145350] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:49.014 [2024-11-27 17:36:20.145410] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:14:49.014 [2024-11-27 17:36:20.145420] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state offline 00:14:49.014 17:36:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@974 -- # wait 92370 00:14:49.014 [2024-11-27 17:36:20.186123] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:14:49.274 17:36:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@786 -- # return 0 00:14:49.274 00:14:49.274 real 0m21.601s 00:14:49.274 user 0m28.166s 00:14:49.274 sys 0m2.806s 00:14:49.274 17:36:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@1126 -- # xtrace_disable 00:14:49.274 ************************************ 00:14:49.274 END TEST raid5f_rebuild_test_sb 00:14:49.274 ************************************ 00:14:49.274 17:36:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:49.534 17:36:20 bdev_raid -- bdev/bdev_raid.sh@985 -- # for n in {3..4} 00:14:49.534 17:36:20 bdev_raid -- bdev/bdev_raid.sh@986 -- # run_test raid5f_state_function_test raid_state_function_test raid5f 4 false 00:14:49.534 17:36:20 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:14:49.534 17:36:20 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:14:49.534 17:36:20 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:14:49.534 ************************************ 00:14:49.534 START TEST raid5f_state_function_test 00:14:49.534 ************************************ 00:14:49.534 17:36:20 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@1125 -- # raid_state_function_test raid5f 4 false 00:14:49.534 17:36:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@205 -- # local raid_level=raid5f 00:14:49.534 17:36:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=4 00:14:49.534 17:36:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@207 -- # local superblock=false 00:14:49.534 17:36:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:14:49.534 17:36:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:14:49.534 17:36:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:14:49.534 17:36:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:14:49.534 17:36:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:14:49.534 17:36:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:14:49.534 17:36:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:14:49.534 17:36:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:14:49.534 17:36:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:14:49.534 17:36:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:14:49.534 17:36:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:14:49.534 17:36:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:14:49.534 17:36:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # echo BaseBdev4 00:14:49.534 17:36:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:14:49.534 17:36:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:14:49.534 17:36:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:14:49.534 17:36:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:14:49.534 17:36:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:14:49.534 17:36:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@211 -- # local strip_size 00:14:49.534 17:36:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:14:49.534 17:36:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:14:49.534 17:36:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@215 -- # '[' raid5f '!=' raid1 ']' 00:14:49.534 17:36:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:14:49.534 17:36:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:14:49.534 17:36:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@222 -- # '[' false = true ']' 00:14:49.534 17:36:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@225 -- # superblock_create_arg= 00:14:49.534 17:36:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@229 -- # raid_pid=93100 00:14:49.534 17:36:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:14:49.534 17:36:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 93100' 00:14:49.534 Process raid pid: 93100 00:14:49.534 17:36:20 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@231 -- # waitforlisten 93100 00:14:49.534 17:36:20 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@831 -- # '[' -z 93100 ']' 00:14:49.534 17:36:20 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:14:49.534 17:36:20 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:14:49.534 17:36:20 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:14:49.534 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:14:49.534 17:36:20 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:14:49.534 17:36:20 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:49.534 [2024-11-27 17:36:20.602523] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:14:49.534 [2024-11-27 17:36:20.602769] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:14:49.794 [2024-11-27 17:36:20.749961] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:14:49.794 [2024-11-27 17:36:20.797669] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:14:49.794 [2024-11-27 17:36:20.840486] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:14:49.794 [2024-11-27 17:36:20.840591] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:14:50.365 17:36:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:14:50.365 17:36:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@864 -- # return 0 00:14:50.365 17:36:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:14:50.365 17:36:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:50.365 17:36:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:50.365 [2024-11-27 17:36:21.422320] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:14:50.365 [2024-11-27 17:36:21.422429] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:14:50.365 [2024-11-27 17:36:21.422462] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:14:50.365 [2024-11-27 17:36:21.422483] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:14:50.365 [2024-11-27 17:36:21.422500] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:14:50.365 [2024-11-27 17:36:21.422522] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:14:50.365 [2024-11-27 17:36:21.422538] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:14:50.365 [2024-11-27 17:36:21.422556] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:14:50.365 17:36:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:50.365 17:36:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:50.365 17:36:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:50.365 17:36:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:50.365 17:36:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:50.365 17:36:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:50.365 17:36:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:50.365 17:36:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:50.365 17:36:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:50.365 17:36:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:50.365 17:36:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:50.365 17:36:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:50.365 17:36:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:50.365 17:36:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:50.365 17:36:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:50.365 17:36:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:50.365 17:36:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:50.365 "name": "Existed_Raid", 00:14:50.365 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:50.365 "strip_size_kb": 64, 00:14:50.365 "state": "configuring", 00:14:50.365 "raid_level": "raid5f", 00:14:50.365 "superblock": false, 00:14:50.365 "num_base_bdevs": 4, 00:14:50.365 "num_base_bdevs_discovered": 0, 00:14:50.365 "num_base_bdevs_operational": 4, 00:14:50.365 "base_bdevs_list": [ 00:14:50.365 { 00:14:50.365 "name": "BaseBdev1", 00:14:50.365 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:50.365 "is_configured": false, 00:14:50.365 "data_offset": 0, 00:14:50.365 "data_size": 0 00:14:50.365 }, 00:14:50.365 { 00:14:50.365 "name": "BaseBdev2", 00:14:50.365 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:50.365 "is_configured": false, 00:14:50.365 "data_offset": 0, 00:14:50.365 "data_size": 0 00:14:50.365 }, 00:14:50.365 { 00:14:50.365 "name": "BaseBdev3", 00:14:50.365 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:50.365 "is_configured": false, 00:14:50.365 "data_offset": 0, 00:14:50.365 "data_size": 0 00:14:50.365 }, 00:14:50.365 { 00:14:50.365 "name": "BaseBdev4", 00:14:50.365 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:50.365 "is_configured": false, 00:14:50.365 "data_offset": 0, 00:14:50.365 "data_size": 0 00:14:50.365 } 00:14:50.365 ] 00:14:50.365 }' 00:14:50.365 17:36:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:50.365 17:36:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:50.935 17:36:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:14:50.935 17:36:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:50.935 17:36:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:50.935 [2024-11-27 17:36:21.913342] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:14:50.935 [2024-11-27 17:36:21.913382] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:14:50.935 17:36:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:50.935 17:36:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:14:50.935 17:36:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:50.935 17:36:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:50.935 [2024-11-27 17:36:21.925349] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:14:50.935 [2024-11-27 17:36:21.925391] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:14:50.935 [2024-11-27 17:36:21.925399] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:14:50.935 [2024-11-27 17:36:21.925408] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:14:50.935 [2024-11-27 17:36:21.925414] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:14:50.935 [2024-11-27 17:36:21.925423] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:14:50.935 [2024-11-27 17:36:21.925429] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:14:50.935 [2024-11-27 17:36:21.925437] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:14:50.935 17:36:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:50.935 17:36:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:14:50.935 17:36:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:50.935 17:36:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:50.935 [2024-11-27 17:36:21.946264] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:14:50.935 BaseBdev1 00:14:50.935 17:36:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:50.935 17:36:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:14:50.935 17:36:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:14:50.935 17:36:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:14:50.935 17:36:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@901 -- # local i 00:14:50.935 17:36:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:14:50.935 17:36:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:14:50.935 17:36:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:14:50.935 17:36:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:50.935 17:36:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:50.935 17:36:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:50.935 17:36:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:14:50.935 17:36:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:50.935 17:36:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:50.935 [ 00:14:50.935 { 00:14:50.935 "name": "BaseBdev1", 00:14:50.935 "aliases": [ 00:14:50.935 "6102f23c-e1d4-4c2f-a62b-a4794a57cbf9" 00:14:50.935 ], 00:14:50.935 "product_name": "Malloc disk", 00:14:50.935 "block_size": 512, 00:14:50.935 "num_blocks": 65536, 00:14:50.935 "uuid": "6102f23c-e1d4-4c2f-a62b-a4794a57cbf9", 00:14:50.935 "assigned_rate_limits": { 00:14:50.935 "rw_ios_per_sec": 0, 00:14:50.935 "rw_mbytes_per_sec": 0, 00:14:50.935 "r_mbytes_per_sec": 0, 00:14:50.935 "w_mbytes_per_sec": 0 00:14:50.935 }, 00:14:50.935 "claimed": true, 00:14:50.935 "claim_type": "exclusive_write", 00:14:50.935 "zoned": false, 00:14:50.935 "supported_io_types": { 00:14:50.935 "read": true, 00:14:50.935 "write": true, 00:14:50.935 "unmap": true, 00:14:50.935 "flush": true, 00:14:50.935 "reset": true, 00:14:50.935 "nvme_admin": false, 00:14:50.935 "nvme_io": false, 00:14:50.935 "nvme_io_md": false, 00:14:50.935 "write_zeroes": true, 00:14:50.935 "zcopy": true, 00:14:50.935 "get_zone_info": false, 00:14:50.935 "zone_management": false, 00:14:50.935 "zone_append": false, 00:14:50.935 "compare": false, 00:14:50.935 "compare_and_write": false, 00:14:50.935 "abort": true, 00:14:50.935 "seek_hole": false, 00:14:50.935 "seek_data": false, 00:14:50.935 "copy": true, 00:14:50.935 "nvme_iov_md": false 00:14:50.935 }, 00:14:50.935 "memory_domains": [ 00:14:50.935 { 00:14:50.935 "dma_device_id": "system", 00:14:50.935 "dma_device_type": 1 00:14:50.935 }, 00:14:50.935 { 00:14:50.935 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:50.935 "dma_device_type": 2 00:14:50.935 } 00:14:50.935 ], 00:14:50.935 "driver_specific": {} 00:14:50.935 } 00:14:50.935 ] 00:14:50.935 17:36:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:50.935 17:36:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:14:50.935 17:36:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:50.935 17:36:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:50.935 17:36:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:50.935 17:36:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:50.935 17:36:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:50.935 17:36:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:50.935 17:36:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:50.935 17:36:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:50.935 17:36:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:50.935 17:36:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:50.936 17:36:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:50.936 17:36:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:50.936 17:36:21 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:50.936 17:36:21 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:50.936 17:36:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:50.936 17:36:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:50.936 "name": "Existed_Raid", 00:14:50.936 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:50.936 "strip_size_kb": 64, 00:14:50.936 "state": "configuring", 00:14:50.936 "raid_level": "raid5f", 00:14:50.936 "superblock": false, 00:14:50.936 "num_base_bdevs": 4, 00:14:50.936 "num_base_bdevs_discovered": 1, 00:14:50.936 "num_base_bdevs_operational": 4, 00:14:50.936 "base_bdevs_list": [ 00:14:50.936 { 00:14:50.936 "name": "BaseBdev1", 00:14:50.936 "uuid": "6102f23c-e1d4-4c2f-a62b-a4794a57cbf9", 00:14:50.936 "is_configured": true, 00:14:50.936 "data_offset": 0, 00:14:50.936 "data_size": 65536 00:14:50.936 }, 00:14:50.936 { 00:14:50.936 "name": "BaseBdev2", 00:14:50.936 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:50.936 "is_configured": false, 00:14:50.936 "data_offset": 0, 00:14:50.936 "data_size": 0 00:14:50.936 }, 00:14:50.936 { 00:14:50.936 "name": "BaseBdev3", 00:14:50.936 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:50.936 "is_configured": false, 00:14:50.936 "data_offset": 0, 00:14:50.936 "data_size": 0 00:14:50.936 }, 00:14:50.936 { 00:14:50.936 "name": "BaseBdev4", 00:14:50.936 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:50.936 "is_configured": false, 00:14:50.936 "data_offset": 0, 00:14:50.936 "data_size": 0 00:14:50.936 } 00:14:50.936 ] 00:14:50.936 }' 00:14:50.936 17:36:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:50.936 17:36:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:51.506 17:36:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:14:51.506 17:36:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:51.506 17:36:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:51.506 [2024-11-27 17:36:22.449514] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:14:51.506 [2024-11-27 17:36:22.449554] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:14:51.506 17:36:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:51.506 17:36:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:14:51.506 17:36:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:51.506 17:36:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:51.506 [2024-11-27 17:36:22.461555] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:14:51.506 [2024-11-27 17:36:22.463357] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:14:51.506 [2024-11-27 17:36:22.463395] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:14:51.506 [2024-11-27 17:36:22.463404] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:14:51.506 [2024-11-27 17:36:22.463412] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:14:51.506 [2024-11-27 17:36:22.463418] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:14:51.506 [2024-11-27 17:36:22.463425] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:14:51.506 17:36:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:51.506 17:36:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:14:51.506 17:36:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:14:51.506 17:36:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:51.506 17:36:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:51.506 17:36:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:51.506 17:36:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:51.506 17:36:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:51.506 17:36:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:51.506 17:36:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:51.506 17:36:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:51.506 17:36:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:51.507 17:36:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:51.507 17:36:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:51.507 17:36:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:51.507 17:36:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:51.507 17:36:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:51.507 17:36:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:51.507 17:36:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:51.507 "name": "Existed_Raid", 00:14:51.507 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:51.507 "strip_size_kb": 64, 00:14:51.507 "state": "configuring", 00:14:51.507 "raid_level": "raid5f", 00:14:51.507 "superblock": false, 00:14:51.507 "num_base_bdevs": 4, 00:14:51.507 "num_base_bdevs_discovered": 1, 00:14:51.507 "num_base_bdevs_operational": 4, 00:14:51.507 "base_bdevs_list": [ 00:14:51.507 { 00:14:51.507 "name": "BaseBdev1", 00:14:51.507 "uuid": "6102f23c-e1d4-4c2f-a62b-a4794a57cbf9", 00:14:51.507 "is_configured": true, 00:14:51.507 "data_offset": 0, 00:14:51.507 "data_size": 65536 00:14:51.507 }, 00:14:51.507 { 00:14:51.507 "name": "BaseBdev2", 00:14:51.507 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:51.507 "is_configured": false, 00:14:51.507 "data_offset": 0, 00:14:51.507 "data_size": 0 00:14:51.507 }, 00:14:51.507 { 00:14:51.507 "name": "BaseBdev3", 00:14:51.507 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:51.507 "is_configured": false, 00:14:51.507 "data_offset": 0, 00:14:51.507 "data_size": 0 00:14:51.507 }, 00:14:51.507 { 00:14:51.507 "name": "BaseBdev4", 00:14:51.507 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:51.507 "is_configured": false, 00:14:51.507 "data_offset": 0, 00:14:51.507 "data_size": 0 00:14:51.507 } 00:14:51.507 ] 00:14:51.507 }' 00:14:51.507 17:36:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:51.507 17:36:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:51.767 17:36:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:14:51.767 17:36:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:51.767 17:36:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:52.028 [2024-11-27 17:36:22.956420] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:14:52.028 BaseBdev2 00:14:52.028 17:36:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:52.028 17:36:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:14:52.028 17:36:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:14:52.028 17:36:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:14:52.028 17:36:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@901 -- # local i 00:14:52.028 17:36:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:14:52.028 17:36:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:14:52.028 17:36:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:14:52.028 17:36:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:52.028 17:36:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:52.028 17:36:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:52.028 17:36:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:14:52.028 17:36:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:52.028 17:36:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:52.028 [ 00:14:52.028 { 00:14:52.028 "name": "BaseBdev2", 00:14:52.028 "aliases": [ 00:14:52.028 "09ffbc2b-5286-4a53-8194-07af72f9e5d6" 00:14:52.028 ], 00:14:52.028 "product_name": "Malloc disk", 00:14:52.028 "block_size": 512, 00:14:52.028 "num_blocks": 65536, 00:14:52.028 "uuid": "09ffbc2b-5286-4a53-8194-07af72f9e5d6", 00:14:52.028 "assigned_rate_limits": { 00:14:52.028 "rw_ios_per_sec": 0, 00:14:52.028 "rw_mbytes_per_sec": 0, 00:14:52.028 "r_mbytes_per_sec": 0, 00:14:52.028 "w_mbytes_per_sec": 0 00:14:52.028 }, 00:14:52.028 "claimed": true, 00:14:52.028 "claim_type": "exclusive_write", 00:14:52.028 "zoned": false, 00:14:52.028 "supported_io_types": { 00:14:52.028 "read": true, 00:14:52.028 "write": true, 00:14:52.028 "unmap": true, 00:14:52.028 "flush": true, 00:14:52.028 "reset": true, 00:14:52.028 "nvme_admin": false, 00:14:52.028 "nvme_io": false, 00:14:52.028 "nvme_io_md": false, 00:14:52.028 "write_zeroes": true, 00:14:52.028 "zcopy": true, 00:14:52.028 "get_zone_info": false, 00:14:52.028 "zone_management": false, 00:14:52.028 "zone_append": false, 00:14:52.028 "compare": false, 00:14:52.028 "compare_and_write": false, 00:14:52.028 "abort": true, 00:14:52.028 "seek_hole": false, 00:14:52.028 "seek_data": false, 00:14:52.028 "copy": true, 00:14:52.028 "nvme_iov_md": false 00:14:52.028 }, 00:14:52.028 "memory_domains": [ 00:14:52.028 { 00:14:52.028 "dma_device_id": "system", 00:14:52.028 "dma_device_type": 1 00:14:52.028 }, 00:14:52.028 { 00:14:52.028 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:52.028 "dma_device_type": 2 00:14:52.028 } 00:14:52.028 ], 00:14:52.028 "driver_specific": {} 00:14:52.028 } 00:14:52.028 ] 00:14:52.028 17:36:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:52.028 17:36:22 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:14:52.028 17:36:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:14:52.028 17:36:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:14:52.028 17:36:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:52.028 17:36:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:52.028 17:36:22 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:52.028 17:36:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:52.028 17:36:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:52.028 17:36:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:52.028 17:36:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:52.028 17:36:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:52.028 17:36:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:52.028 17:36:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:52.028 17:36:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:52.028 17:36:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:52.028 17:36:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:52.028 17:36:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:52.028 17:36:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:52.028 17:36:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:52.028 "name": "Existed_Raid", 00:14:52.028 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:52.028 "strip_size_kb": 64, 00:14:52.028 "state": "configuring", 00:14:52.028 "raid_level": "raid5f", 00:14:52.028 "superblock": false, 00:14:52.028 "num_base_bdevs": 4, 00:14:52.028 "num_base_bdevs_discovered": 2, 00:14:52.028 "num_base_bdevs_operational": 4, 00:14:52.028 "base_bdevs_list": [ 00:14:52.028 { 00:14:52.028 "name": "BaseBdev1", 00:14:52.028 "uuid": "6102f23c-e1d4-4c2f-a62b-a4794a57cbf9", 00:14:52.028 "is_configured": true, 00:14:52.028 "data_offset": 0, 00:14:52.028 "data_size": 65536 00:14:52.028 }, 00:14:52.028 { 00:14:52.028 "name": "BaseBdev2", 00:14:52.028 "uuid": "09ffbc2b-5286-4a53-8194-07af72f9e5d6", 00:14:52.028 "is_configured": true, 00:14:52.028 "data_offset": 0, 00:14:52.028 "data_size": 65536 00:14:52.028 }, 00:14:52.028 { 00:14:52.028 "name": "BaseBdev3", 00:14:52.028 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:52.028 "is_configured": false, 00:14:52.028 "data_offset": 0, 00:14:52.028 "data_size": 0 00:14:52.028 }, 00:14:52.028 { 00:14:52.028 "name": "BaseBdev4", 00:14:52.028 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:52.028 "is_configured": false, 00:14:52.028 "data_offset": 0, 00:14:52.028 "data_size": 0 00:14:52.028 } 00:14:52.028 ] 00:14:52.028 }' 00:14:52.028 17:36:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:52.028 17:36:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:52.289 17:36:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:14:52.289 17:36:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:52.289 17:36:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:52.289 [2024-11-27 17:36:23.434603] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:14:52.289 BaseBdev3 00:14:52.289 17:36:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:52.289 17:36:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:14:52.289 17:36:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:14:52.289 17:36:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:14:52.289 17:36:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@901 -- # local i 00:14:52.289 17:36:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:14:52.289 17:36:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:14:52.289 17:36:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:14:52.289 17:36:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:52.289 17:36:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:52.289 17:36:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:52.289 17:36:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:14:52.289 17:36:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:52.289 17:36:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:52.289 [ 00:14:52.289 { 00:14:52.289 "name": "BaseBdev3", 00:14:52.289 "aliases": [ 00:14:52.289 "be3c6fb6-7856-49fb-8f0f-787e291424f4" 00:14:52.289 ], 00:14:52.289 "product_name": "Malloc disk", 00:14:52.289 "block_size": 512, 00:14:52.289 "num_blocks": 65536, 00:14:52.289 "uuid": "be3c6fb6-7856-49fb-8f0f-787e291424f4", 00:14:52.289 "assigned_rate_limits": { 00:14:52.289 "rw_ios_per_sec": 0, 00:14:52.289 "rw_mbytes_per_sec": 0, 00:14:52.289 "r_mbytes_per_sec": 0, 00:14:52.289 "w_mbytes_per_sec": 0 00:14:52.289 }, 00:14:52.289 "claimed": true, 00:14:52.289 "claim_type": "exclusive_write", 00:14:52.289 "zoned": false, 00:14:52.289 "supported_io_types": { 00:14:52.289 "read": true, 00:14:52.289 "write": true, 00:14:52.289 "unmap": true, 00:14:52.289 "flush": true, 00:14:52.289 "reset": true, 00:14:52.289 "nvme_admin": false, 00:14:52.289 "nvme_io": false, 00:14:52.289 "nvme_io_md": false, 00:14:52.289 "write_zeroes": true, 00:14:52.289 "zcopy": true, 00:14:52.289 "get_zone_info": false, 00:14:52.289 "zone_management": false, 00:14:52.289 "zone_append": false, 00:14:52.289 "compare": false, 00:14:52.289 "compare_and_write": false, 00:14:52.289 "abort": true, 00:14:52.289 "seek_hole": false, 00:14:52.289 "seek_data": false, 00:14:52.289 "copy": true, 00:14:52.289 "nvme_iov_md": false 00:14:52.289 }, 00:14:52.289 "memory_domains": [ 00:14:52.289 { 00:14:52.289 "dma_device_id": "system", 00:14:52.289 "dma_device_type": 1 00:14:52.289 }, 00:14:52.289 { 00:14:52.289 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:52.289 "dma_device_type": 2 00:14:52.289 } 00:14:52.289 ], 00:14:52.289 "driver_specific": {} 00:14:52.289 } 00:14:52.289 ] 00:14:52.289 17:36:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:52.289 17:36:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:14:52.289 17:36:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:14:52.289 17:36:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:14:52.289 17:36:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:52.290 17:36:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:52.290 17:36:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:52.290 17:36:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:52.290 17:36:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:52.290 17:36:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:52.290 17:36:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:52.290 17:36:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:52.290 17:36:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:52.290 17:36:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:52.290 17:36:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:52.290 17:36:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:52.290 17:36:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:52.290 17:36:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:52.549 17:36:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:52.549 17:36:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:52.549 "name": "Existed_Raid", 00:14:52.550 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:52.550 "strip_size_kb": 64, 00:14:52.550 "state": "configuring", 00:14:52.550 "raid_level": "raid5f", 00:14:52.550 "superblock": false, 00:14:52.550 "num_base_bdevs": 4, 00:14:52.550 "num_base_bdevs_discovered": 3, 00:14:52.550 "num_base_bdevs_operational": 4, 00:14:52.550 "base_bdevs_list": [ 00:14:52.550 { 00:14:52.550 "name": "BaseBdev1", 00:14:52.550 "uuid": "6102f23c-e1d4-4c2f-a62b-a4794a57cbf9", 00:14:52.550 "is_configured": true, 00:14:52.550 "data_offset": 0, 00:14:52.550 "data_size": 65536 00:14:52.550 }, 00:14:52.550 { 00:14:52.550 "name": "BaseBdev2", 00:14:52.550 "uuid": "09ffbc2b-5286-4a53-8194-07af72f9e5d6", 00:14:52.550 "is_configured": true, 00:14:52.550 "data_offset": 0, 00:14:52.550 "data_size": 65536 00:14:52.550 }, 00:14:52.550 { 00:14:52.550 "name": "BaseBdev3", 00:14:52.550 "uuid": "be3c6fb6-7856-49fb-8f0f-787e291424f4", 00:14:52.550 "is_configured": true, 00:14:52.550 "data_offset": 0, 00:14:52.550 "data_size": 65536 00:14:52.550 }, 00:14:52.550 { 00:14:52.550 "name": "BaseBdev4", 00:14:52.550 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:52.550 "is_configured": false, 00:14:52.550 "data_offset": 0, 00:14:52.550 "data_size": 0 00:14:52.550 } 00:14:52.550 ] 00:14:52.550 }' 00:14:52.550 17:36:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:52.550 17:36:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:52.810 17:36:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:14:52.810 17:36:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:52.810 17:36:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:52.810 [2024-11-27 17:36:23.956622] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:14:52.810 [2024-11-27 17:36:23.956744] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:14:52.810 [2024-11-27 17:36:23.956758] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 196608, blocklen 512 00:14:52.810 [2024-11-27 17:36:23.957017] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:14:52.810 [2024-11-27 17:36:23.957479] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:14:52.810 [2024-11-27 17:36:23.957541] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:14:52.810 [2024-11-27 17:36:23.957735] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:52.810 BaseBdev4 00:14:52.810 17:36:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:52.810 17:36:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev4 00:14:52.810 17:36:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev4 00:14:52.810 17:36:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:14:52.810 17:36:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@901 -- # local i 00:14:52.810 17:36:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:14:52.810 17:36:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:14:52.810 17:36:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:14:52.810 17:36:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:52.810 17:36:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:52.810 17:36:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:52.810 17:36:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:14:52.811 17:36:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:52.811 17:36:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:52.811 [ 00:14:52.811 { 00:14:52.811 "name": "BaseBdev4", 00:14:52.811 "aliases": [ 00:14:52.811 "cbe94f42-d505-4f88-aed3-74f4f93f98fb" 00:14:52.811 ], 00:14:52.811 "product_name": "Malloc disk", 00:14:52.811 "block_size": 512, 00:14:52.811 "num_blocks": 65536, 00:14:52.811 "uuid": "cbe94f42-d505-4f88-aed3-74f4f93f98fb", 00:14:52.811 "assigned_rate_limits": { 00:14:52.811 "rw_ios_per_sec": 0, 00:14:52.811 "rw_mbytes_per_sec": 0, 00:14:52.811 "r_mbytes_per_sec": 0, 00:14:52.811 "w_mbytes_per_sec": 0 00:14:52.811 }, 00:14:52.811 "claimed": true, 00:14:52.811 "claim_type": "exclusive_write", 00:14:52.811 "zoned": false, 00:14:52.811 "supported_io_types": { 00:14:52.811 "read": true, 00:14:52.811 "write": true, 00:14:52.811 "unmap": true, 00:14:52.811 "flush": true, 00:14:52.811 "reset": true, 00:14:52.811 "nvme_admin": false, 00:14:52.811 "nvme_io": false, 00:14:52.811 "nvme_io_md": false, 00:14:52.811 "write_zeroes": true, 00:14:52.811 "zcopy": true, 00:14:52.811 "get_zone_info": false, 00:14:52.811 "zone_management": false, 00:14:52.811 "zone_append": false, 00:14:52.811 "compare": false, 00:14:52.811 "compare_and_write": false, 00:14:52.811 "abort": true, 00:14:52.811 "seek_hole": false, 00:14:52.811 "seek_data": false, 00:14:52.811 "copy": true, 00:14:52.811 "nvme_iov_md": false 00:14:52.811 }, 00:14:52.811 "memory_domains": [ 00:14:52.811 { 00:14:52.811 "dma_device_id": "system", 00:14:52.811 "dma_device_type": 1 00:14:52.811 }, 00:14:52.811 { 00:14:52.811 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:52.811 "dma_device_type": 2 00:14:52.811 } 00:14:52.811 ], 00:14:52.811 "driver_specific": {} 00:14:52.811 } 00:14:52.811 ] 00:14:52.811 17:36:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:52.811 17:36:23 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:14:52.811 17:36:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:14:52.811 17:36:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:14:52.811 17:36:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 4 00:14:52.811 17:36:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:52.811 17:36:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:52.811 17:36:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:52.811 17:36:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:52.811 17:36:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:52.811 17:36:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:52.811 17:36:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:52.811 17:36:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:52.811 17:36:23 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:53.071 17:36:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:53.071 17:36:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:53.071 17:36:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:53.071 17:36:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:53.071 17:36:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:53.071 17:36:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:53.071 "name": "Existed_Raid", 00:14:53.071 "uuid": "7757d97b-0ff3-4207-9489-a56dffc11664", 00:14:53.071 "strip_size_kb": 64, 00:14:53.071 "state": "online", 00:14:53.071 "raid_level": "raid5f", 00:14:53.071 "superblock": false, 00:14:53.071 "num_base_bdevs": 4, 00:14:53.071 "num_base_bdevs_discovered": 4, 00:14:53.071 "num_base_bdevs_operational": 4, 00:14:53.071 "base_bdevs_list": [ 00:14:53.071 { 00:14:53.071 "name": "BaseBdev1", 00:14:53.071 "uuid": "6102f23c-e1d4-4c2f-a62b-a4794a57cbf9", 00:14:53.071 "is_configured": true, 00:14:53.071 "data_offset": 0, 00:14:53.071 "data_size": 65536 00:14:53.071 }, 00:14:53.071 { 00:14:53.071 "name": "BaseBdev2", 00:14:53.071 "uuid": "09ffbc2b-5286-4a53-8194-07af72f9e5d6", 00:14:53.071 "is_configured": true, 00:14:53.071 "data_offset": 0, 00:14:53.071 "data_size": 65536 00:14:53.071 }, 00:14:53.071 { 00:14:53.071 "name": "BaseBdev3", 00:14:53.071 "uuid": "be3c6fb6-7856-49fb-8f0f-787e291424f4", 00:14:53.071 "is_configured": true, 00:14:53.071 "data_offset": 0, 00:14:53.071 "data_size": 65536 00:14:53.071 }, 00:14:53.071 { 00:14:53.071 "name": "BaseBdev4", 00:14:53.071 "uuid": "cbe94f42-d505-4f88-aed3-74f4f93f98fb", 00:14:53.071 "is_configured": true, 00:14:53.071 "data_offset": 0, 00:14:53.071 "data_size": 65536 00:14:53.071 } 00:14:53.071 ] 00:14:53.071 }' 00:14:53.071 17:36:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:53.071 17:36:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:53.332 17:36:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:14:53.332 17:36:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:14:53.332 17:36:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:14:53.332 17:36:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:14:53.332 17:36:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:14:53.332 17:36:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:14:53.332 17:36:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:14:53.332 17:36:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:53.332 17:36:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:53.332 17:36:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:14:53.332 [2024-11-27 17:36:24.467942] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:53.332 17:36:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:53.332 17:36:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:14:53.332 "name": "Existed_Raid", 00:14:53.332 "aliases": [ 00:14:53.332 "7757d97b-0ff3-4207-9489-a56dffc11664" 00:14:53.332 ], 00:14:53.332 "product_name": "Raid Volume", 00:14:53.332 "block_size": 512, 00:14:53.332 "num_blocks": 196608, 00:14:53.332 "uuid": "7757d97b-0ff3-4207-9489-a56dffc11664", 00:14:53.332 "assigned_rate_limits": { 00:14:53.332 "rw_ios_per_sec": 0, 00:14:53.332 "rw_mbytes_per_sec": 0, 00:14:53.332 "r_mbytes_per_sec": 0, 00:14:53.332 "w_mbytes_per_sec": 0 00:14:53.332 }, 00:14:53.332 "claimed": false, 00:14:53.332 "zoned": false, 00:14:53.332 "supported_io_types": { 00:14:53.332 "read": true, 00:14:53.332 "write": true, 00:14:53.332 "unmap": false, 00:14:53.332 "flush": false, 00:14:53.332 "reset": true, 00:14:53.332 "nvme_admin": false, 00:14:53.332 "nvme_io": false, 00:14:53.332 "nvme_io_md": false, 00:14:53.332 "write_zeroes": true, 00:14:53.332 "zcopy": false, 00:14:53.332 "get_zone_info": false, 00:14:53.332 "zone_management": false, 00:14:53.332 "zone_append": false, 00:14:53.332 "compare": false, 00:14:53.332 "compare_and_write": false, 00:14:53.332 "abort": false, 00:14:53.332 "seek_hole": false, 00:14:53.332 "seek_data": false, 00:14:53.332 "copy": false, 00:14:53.332 "nvme_iov_md": false 00:14:53.332 }, 00:14:53.332 "driver_specific": { 00:14:53.332 "raid": { 00:14:53.332 "uuid": "7757d97b-0ff3-4207-9489-a56dffc11664", 00:14:53.332 "strip_size_kb": 64, 00:14:53.332 "state": "online", 00:14:53.332 "raid_level": "raid5f", 00:14:53.332 "superblock": false, 00:14:53.332 "num_base_bdevs": 4, 00:14:53.332 "num_base_bdevs_discovered": 4, 00:14:53.332 "num_base_bdevs_operational": 4, 00:14:53.332 "base_bdevs_list": [ 00:14:53.332 { 00:14:53.332 "name": "BaseBdev1", 00:14:53.332 "uuid": "6102f23c-e1d4-4c2f-a62b-a4794a57cbf9", 00:14:53.332 "is_configured": true, 00:14:53.332 "data_offset": 0, 00:14:53.332 "data_size": 65536 00:14:53.332 }, 00:14:53.332 { 00:14:53.332 "name": "BaseBdev2", 00:14:53.332 "uuid": "09ffbc2b-5286-4a53-8194-07af72f9e5d6", 00:14:53.332 "is_configured": true, 00:14:53.332 "data_offset": 0, 00:14:53.332 "data_size": 65536 00:14:53.332 }, 00:14:53.332 { 00:14:53.332 "name": "BaseBdev3", 00:14:53.332 "uuid": "be3c6fb6-7856-49fb-8f0f-787e291424f4", 00:14:53.332 "is_configured": true, 00:14:53.332 "data_offset": 0, 00:14:53.332 "data_size": 65536 00:14:53.332 }, 00:14:53.332 { 00:14:53.332 "name": "BaseBdev4", 00:14:53.332 "uuid": "cbe94f42-d505-4f88-aed3-74f4f93f98fb", 00:14:53.332 "is_configured": true, 00:14:53.332 "data_offset": 0, 00:14:53.332 "data_size": 65536 00:14:53.332 } 00:14:53.332 ] 00:14:53.332 } 00:14:53.332 } 00:14:53.332 }' 00:14:53.332 17:36:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:14:53.592 17:36:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:14:53.592 BaseBdev2 00:14:53.592 BaseBdev3 00:14:53.592 BaseBdev4' 00:14:53.592 17:36:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:53.592 17:36:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:14:53.592 17:36:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:53.592 17:36:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:53.592 17:36:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:14:53.592 17:36:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:53.592 17:36:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:53.592 17:36:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:53.592 17:36:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:53.592 17:36:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:53.592 17:36:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:53.592 17:36:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:53.592 17:36:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:14:53.592 17:36:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:53.592 17:36:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:53.592 17:36:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:53.592 17:36:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:53.592 17:36:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:53.592 17:36:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:53.592 17:36:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:14:53.592 17:36:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:53.592 17:36:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:53.592 17:36:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:53.592 17:36:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:53.592 17:36:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:53.592 17:36:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:53.592 17:36:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:53.592 17:36:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:14:53.592 17:36:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:53.592 17:36:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:53.592 17:36:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:53.592 17:36:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:53.592 17:36:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:53.592 17:36:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:53.592 17:36:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:14:53.592 17:36:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:53.592 17:36:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:53.852 [2024-11-27 17:36:24.783254] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:14:53.852 17:36:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:53.852 17:36:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@260 -- # local expected_state 00:14:53.852 17:36:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@261 -- # has_redundancy raid5f 00:14:53.852 17:36:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:14:53.852 17:36:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@199 -- # return 0 00:14:53.852 17:36:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:14:53.852 17:36:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 3 00:14:53.852 17:36:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:53.852 17:36:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:53.852 17:36:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:53.852 17:36:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:53.852 17:36:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:14:53.852 17:36:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:53.852 17:36:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:53.852 17:36:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:53.852 17:36:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:53.852 17:36:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:53.852 17:36:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:53.852 17:36:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:53.852 17:36:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:53.852 17:36:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:53.852 17:36:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:53.852 "name": "Existed_Raid", 00:14:53.852 "uuid": "7757d97b-0ff3-4207-9489-a56dffc11664", 00:14:53.852 "strip_size_kb": 64, 00:14:53.852 "state": "online", 00:14:53.852 "raid_level": "raid5f", 00:14:53.852 "superblock": false, 00:14:53.852 "num_base_bdevs": 4, 00:14:53.852 "num_base_bdevs_discovered": 3, 00:14:53.852 "num_base_bdevs_operational": 3, 00:14:53.852 "base_bdevs_list": [ 00:14:53.852 { 00:14:53.852 "name": null, 00:14:53.852 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:53.852 "is_configured": false, 00:14:53.852 "data_offset": 0, 00:14:53.852 "data_size": 65536 00:14:53.852 }, 00:14:53.852 { 00:14:53.852 "name": "BaseBdev2", 00:14:53.852 "uuid": "09ffbc2b-5286-4a53-8194-07af72f9e5d6", 00:14:53.852 "is_configured": true, 00:14:53.852 "data_offset": 0, 00:14:53.852 "data_size": 65536 00:14:53.853 }, 00:14:53.853 { 00:14:53.853 "name": "BaseBdev3", 00:14:53.853 "uuid": "be3c6fb6-7856-49fb-8f0f-787e291424f4", 00:14:53.853 "is_configured": true, 00:14:53.853 "data_offset": 0, 00:14:53.853 "data_size": 65536 00:14:53.853 }, 00:14:53.853 { 00:14:53.853 "name": "BaseBdev4", 00:14:53.853 "uuid": "cbe94f42-d505-4f88-aed3-74f4f93f98fb", 00:14:53.853 "is_configured": true, 00:14:53.853 "data_offset": 0, 00:14:53.853 "data_size": 65536 00:14:53.853 } 00:14:53.853 ] 00:14:53.853 }' 00:14:53.853 17:36:24 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:53.853 17:36:24 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:54.113 17:36:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:14:54.113 17:36:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:14:54.113 17:36:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:14:54.113 17:36:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:54.113 17:36:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:54.113 17:36:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:54.113 17:36:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:54.114 17:36:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:14:54.114 17:36:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:14:54.114 17:36:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:14:54.114 17:36:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:54.114 17:36:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:54.114 [2024-11-27 17:36:25.273529] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:14:54.114 [2024-11-27 17:36:25.273618] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:14:54.114 [2024-11-27 17:36:25.284779] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:54.114 17:36:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:54.114 17:36:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:14:54.114 17:36:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:14:54.114 17:36:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:54.114 17:36:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:54.114 17:36:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:54.114 17:36:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:14:54.375 17:36:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:54.375 17:36:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:14:54.375 17:36:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:14:54.375 17:36:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:14:54.375 17:36:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:54.375 17:36:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:54.375 [2024-11-27 17:36:25.344692] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:14:54.375 17:36:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:54.375 17:36:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:14:54.375 17:36:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:14:54.375 17:36:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:54.375 17:36:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:14:54.375 17:36:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:54.375 17:36:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:54.375 17:36:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:54.375 17:36:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:14:54.375 17:36:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:14:54.375 17:36:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev4 00:14:54.375 17:36:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:54.375 17:36:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:54.375 [2024-11-27 17:36:25.411804] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev4 00:14:54.375 [2024-11-27 17:36:25.411913] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:14:54.375 17:36:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:54.375 17:36:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:14:54.375 17:36:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:14:54.375 17:36:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:54.375 17:36:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:14:54.375 17:36:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:54.375 17:36:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:54.375 17:36:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:54.375 17:36:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:14:54.375 17:36:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:14:54.375 17:36:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@284 -- # '[' 4 -gt 2 ']' 00:14:54.375 17:36:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:14:54.375 17:36:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:14:54.375 17:36:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:14:54.375 17:36:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:54.375 17:36:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:54.375 BaseBdev2 00:14:54.376 17:36:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:54.376 17:36:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:14:54.376 17:36:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:14:54.376 17:36:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:14:54.376 17:36:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@901 -- # local i 00:14:54.376 17:36:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:14:54.376 17:36:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:14:54.376 17:36:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:14:54.376 17:36:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:54.376 17:36:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:54.376 17:36:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:54.376 17:36:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:14:54.376 17:36:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:54.376 17:36:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:54.376 [ 00:14:54.376 { 00:14:54.376 "name": "BaseBdev2", 00:14:54.376 "aliases": [ 00:14:54.376 "73170cc6-caa1-4cf3-a5c7-73ecf0b2f958" 00:14:54.376 ], 00:14:54.376 "product_name": "Malloc disk", 00:14:54.376 "block_size": 512, 00:14:54.376 "num_blocks": 65536, 00:14:54.376 "uuid": "73170cc6-caa1-4cf3-a5c7-73ecf0b2f958", 00:14:54.376 "assigned_rate_limits": { 00:14:54.376 "rw_ios_per_sec": 0, 00:14:54.376 "rw_mbytes_per_sec": 0, 00:14:54.376 "r_mbytes_per_sec": 0, 00:14:54.376 "w_mbytes_per_sec": 0 00:14:54.376 }, 00:14:54.376 "claimed": false, 00:14:54.376 "zoned": false, 00:14:54.376 "supported_io_types": { 00:14:54.376 "read": true, 00:14:54.376 "write": true, 00:14:54.376 "unmap": true, 00:14:54.376 "flush": true, 00:14:54.376 "reset": true, 00:14:54.376 "nvme_admin": false, 00:14:54.376 "nvme_io": false, 00:14:54.376 "nvme_io_md": false, 00:14:54.376 "write_zeroes": true, 00:14:54.376 "zcopy": true, 00:14:54.376 "get_zone_info": false, 00:14:54.376 "zone_management": false, 00:14:54.376 "zone_append": false, 00:14:54.376 "compare": false, 00:14:54.376 "compare_and_write": false, 00:14:54.376 "abort": true, 00:14:54.376 "seek_hole": false, 00:14:54.376 "seek_data": false, 00:14:54.376 "copy": true, 00:14:54.376 "nvme_iov_md": false 00:14:54.376 }, 00:14:54.376 "memory_domains": [ 00:14:54.376 { 00:14:54.376 "dma_device_id": "system", 00:14:54.376 "dma_device_type": 1 00:14:54.376 }, 00:14:54.376 { 00:14:54.376 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:54.376 "dma_device_type": 2 00:14:54.376 } 00:14:54.376 ], 00:14:54.376 "driver_specific": {} 00:14:54.376 } 00:14:54.376 ] 00:14:54.376 17:36:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:54.376 17:36:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:14:54.376 17:36:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:14:54.376 17:36:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:14:54.376 17:36:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:14:54.376 17:36:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:54.376 17:36:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:54.376 BaseBdev3 00:14:54.376 17:36:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:54.376 17:36:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:14:54.376 17:36:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:14:54.376 17:36:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:14:54.376 17:36:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@901 -- # local i 00:14:54.376 17:36:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:14:54.376 17:36:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:14:54.376 17:36:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:14:54.376 17:36:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:54.376 17:36:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:54.376 17:36:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:54.376 17:36:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:14:54.376 17:36:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:54.376 17:36:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:54.637 [ 00:14:54.637 { 00:14:54.637 "name": "BaseBdev3", 00:14:54.637 "aliases": [ 00:14:54.637 "317ec481-ef9e-4fd7-8648-5492f42ebb25" 00:14:54.637 ], 00:14:54.637 "product_name": "Malloc disk", 00:14:54.637 "block_size": 512, 00:14:54.637 "num_blocks": 65536, 00:14:54.637 "uuid": "317ec481-ef9e-4fd7-8648-5492f42ebb25", 00:14:54.637 "assigned_rate_limits": { 00:14:54.637 "rw_ios_per_sec": 0, 00:14:54.637 "rw_mbytes_per_sec": 0, 00:14:54.637 "r_mbytes_per_sec": 0, 00:14:54.637 "w_mbytes_per_sec": 0 00:14:54.637 }, 00:14:54.637 "claimed": false, 00:14:54.637 "zoned": false, 00:14:54.637 "supported_io_types": { 00:14:54.637 "read": true, 00:14:54.637 "write": true, 00:14:54.637 "unmap": true, 00:14:54.637 "flush": true, 00:14:54.637 "reset": true, 00:14:54.637 "nvme_admin": false, 00:14:54.637 "nvme_io": false, 00:14:54.637 "nvme_io_md": false, 00:14:54.637 "write_zeroes": true, 00:14:54.637 "zcopy": true, 00:14:54.637 "get_zone_info": false, 00:14:54.637 "zone_management": false, 00:14:54.637 "zone_append": false, 00:14:54.637 "compare": false, 00:14:54.637 "compare_and_write": false, 00:14:54.637 "abort": true, 00:14:54.637 "seek_hole": false, 00:14:54.637 "seek_data": false, 00:14:54.637 "copy": true, 00:14:54.637 "nvme_iov_md": false 00:14:54.637 }, 00:14:54.637 "memory_domains": [ 00:14:54.637 { 00:14:54.637 "dma_device_id": "system", 00:14:54.637 "dma_device_type": 1 00:14:54.637 }, 00:14:54.637 { 00:14:54.637 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:54.637 "dma_device_type": 2 00:14:54.637 } 00:14:54.637 ], 00:14:54.637 "driver_specific": {} 00:14:54.637 } 00:14:54.637 ] 00:14:54.637 17:36:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:54.637 17:36:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:14:54.637 17:36:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:14:54.637 17:36:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:14:54.637 17:36:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:14:54.637 17:36:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:54.637 17:36:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:54.637 BaseBdev4 00:14:54.637 17:36:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:54.637 17:36:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev4 00:14:54.637 17:36:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev4 00:14:54.637 17:36:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:14:54.637 17:36:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@901 -- # local i 00:14:54.637 17:36:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:14:54.637 17:36:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:14:54.637 17:36:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:14:54.637 17:36:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:54.637 17:36:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:54.637 17:36:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:54.637 17:36:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:14:54.637 17:36:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:54.637 17:36:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:54.637 [ 00:14:54.637 { 00:14:54.637 "name": "BaseBdev4", 00:14:54.637 "aliases": [ 00:14:54.637 "f5214f25-4720-4ddd-8ac1-03d864841930" 00:14:54.637 ], 00:14:54.637 "product_name": "Malloc disk", 00:14:54.637 "block_size": 512, 00:14:54.637 "num_blocks": 65536, 00:14:54.637 "uuid": "f5214f25-4720-4ddd-8ac1-03d864841930", 00:14:54.637 "assigned_rate_limits": { 00:14:54.637 "rw_ios_per_sec": 0, 00:14:54.637 "rw_mbytes_per_sec": 0, 00:14:54.637 "r_mbytes_per_sec": 0, 00:14:54.637 "w_mbytes_per_sec": 0 00:14:54.637 }, 00:14:54.637 "claimed": false, 00:14:54.637 "zoned": false, 00:14:54.637 "supported_io_types": { 00:14:54.637 "read": true, 00:14:54.637 "write": true, 00:14:54.637 "unmap": true, 00:14:54.637 "flush": true, 00:14:54.637 "reset": true, 00:14:54.637 "nvme_admin": false, 00:14:54.637 "nvme_io": false, 00:14:54.637 "nvme_io_md": false, 00:14:54.637 "write_zeroes": true, 00:14:54.637 "zcopy": true, 00:14:54.637 "get_zone_info": false, 00:14:54.637 "zone_management": false, 00:14:54.637 "zone_append": false, 00:14:54.637 "compare": false, 00:14:54.637 "compare_and_write": false, 00:14:54.637 "abort": true, 00:14:54.637 "seek_hole": false, 00:14:54.637 "seek_data": false, 00:14:54.637 "copy": true, 00:14:54.637 "nvme_iov_md": false 00:14:54.637 }, 00:14:54.637 "memory_domains": [ 00:14:54.637 { 00:14:54.637 "dma_device_id": "system", 00:14:54.637 "dma_device_type": 1 00:14:54.637 }, 00:14:54.637 { 00:14:54.637 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:54.637 "dma_device_type": 2 00:14:54.637 } 00:14:54.637 ], 00:14:54.637 "driver_specific": {} 00:14:54.637 } 00:14:54.637 ] 00:14:54.638 17:36:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:54.638 17:36:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:14:54.638 17:36:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:14:54.638 17:36:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:14:54.638 17:36:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:14:54.638 17:36:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:54.638 17:36:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:54.638 [2024-11-27 17:36:25.638432] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:14:54.638 [2024-11-27 17:36:25.638554] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:14:54.638 [2024-11-27 17:36:25.638592] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:14:54.638 [2024-11-27 17:36:25.640366] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:14:54.638 [2024-11-27 17:36:25.640451] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:14:54.638 17:36:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:54.638 17:36:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:54.638 17:36:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:54.638 17:36:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:54.638 17:36:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:54.638 17:36:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:54.638 17:36:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:54.638 17:36:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:54.638 17:36:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:54.638 17:36:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:54.638 17:36:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:54.638 17:36:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:54.638 17:36:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:54.638 17:36:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:54.638 17:36:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:54.638 17:36:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:54.638 17:36:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:54.638 "name": "Existed_Raid", 00:14:54.638 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:54.638 "strip_size_kb": 64, 00:14:54.638 "state": "configuring", 00:14:54.638 "raid_level": "raid5f", 00:14:54.638 "superblock": false, 00:14:54.638 "num_base_bdevs": 4, 00:14:54.638 "num_base_bdevs_discovered": 3, 00:14:54.638 "num_base_bdevs_operational": 4, 00:14:54.638 "base_bdevs_list": [ 00:14:54.638 { 00:14:54.638 "name": "BaseBdev1", 00:14:54.638 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:54.638 "is_configured": false, 00:14:54.638 "data_offset": 0, 00:14:54.638 "data_size": 0 00:14:54.638 }, 00:14:54.638 { 00:14:54.638 "name": "BaseBdev2", 00:14:54.638 "uuid": "73170cc6-caa1-4cf3-a5c7-73ecf0b2f958", 00:14:54.638 "is_configured": true, 00:14:54.638 "data_offset": 0, 00:14:54.638 "data_size": 65536 00:14:54.638 }, 00:14:54.638 { 00:14:54.638 "name": "BaseBdev3", 00:14:54.638 "uuid": "317ec481-ef9e-4fd7-8648-5492f42ebb25", 00:14:54.638 "is_configured": true, 00:14:54.638 "data_offset": 0, 00:14:54.638 "data_size": 65536 00:14:54.638 }, 00:14:54.638 { 00:14:54.638 "name": "BaseBdev4", 00:14:54.638 "uuid": "f5214f25-4720-4ddd-8ac1-03d864841930", 00:14:54.638 "is_configured": true, 00:14:54.638 "data_offset": 0, 00:14:54.638 "data_size": 65536 00:14:54.638 } 00:14:54.638 ] 00:14:54.638 }' 00:14:54.638 17:36:25 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:54.638 17:36:25 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:55.209 17:36:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:14:55.209 17:36:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:55.209 17:36:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:55.209 [2024-11-27 17:36:26.105579] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:14:55.209 17:36:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:55.209 17:36:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:55.209 17:36:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:55.209 17:36:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:55.209 17:36:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:55.209 17:36:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:55.209 17:36:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:55.209 17:36:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:55.209 17:36:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:55.209 17:36:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:55.209 17:36:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:55.209 17:36:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:55.209 17:36:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:55.209 17:36:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:55.209 17:36:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:55.209 17:36:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:55.209 17:36:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:55.209 "name": "Existed_Raid", 00:14:55.209 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:55.209 "strip_size_kb": 64, 00:14:55.209 "state": "configuring", 00:14:55.209 "raid_level": "raid5f", 00:14:55.209 "superblock": false, 00:14:55.209 "num_base_bdevs": 4, 00:14:55.209 "num_base_bdevs_discovered": 2, 00:14:55.209 "num_base_bdevs_operational": 4, 00:14:55.209 "base_bdevs_list": [ 00:14:55.209 { 00:14:55.209 "name": "BaseBdev1", 00:14:55.209 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:55.209 "is_configured": false, 00:14:55.209 "data_offset": 0, 00:14:55.209 "data_size": 0 00:14:55.209 }, 00:14:55.209 { 00:14:55.209 "name": null, 00:14:55.209 "uuid": "73170cc6-caa1-4cf3-a5c7-73ecf0b2f958", 00:14:55.209 "is_configured": false, 00:14:55.209 "data_offset": 0, 00:14:55.209 "data_size": 65536 00:14:55.209 }, 00:14:55.209 { 00:14:55.209 "name": "BaseBdev3", 00:14:55.209 "uuid": "317ec481-ef9e-4fd7-8648-5492f42ebb25", 00:14:55.209 "is_configured": true, 00:14:55.209 "data_offset": 0, 00:14:55.209 "data_size": 65536 00:14:55.209 }, 00:14:55.209 { 00:14:55.209 "name": "BaseBdev4", 00:14:55.209 "uuid": "f5214f25-4720-4ddd-8ac1-03d864841930", 00:14:55.209 "is_configured": true, 00:14:55.209 "data_offset": 0, 00:14:55.209 "data_size": 65536 00:14:55.209 } 00:14:55.209 ] 00:14:55.209 }' 00:14:55.209 17:36:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:55.209 17:36:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:55.470 17:36:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:55.471 17:36:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:14:55.471 17:36:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:55.471 17:36:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:55.471 17:36:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:55.471 17:36:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:14:55.471 17:36:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:14:55.471 17:36:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:55.471 17:36:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:55.471 [2024-11-27 17:36:26.591717] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:14:55.471 BaseBdev1 00:14:55.471 17:36:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:55.471 17:36:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:14:55.471 17:36:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:14:55.471 17:36:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:14:55.471 17:36:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@901 -- # local i 00:14:55.471 17:36:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:14:55.471 17:36:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:14:55.471 17:36:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:14:55.471 17:36:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:55.471 17:36:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:55.471 17:36:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:55.471 17:36:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:14:55.471 17:36:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:55.471 17:36:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:55.471 [ 00:14:55.471 { 00:14:55.471 "name": "BaseBdev1", 00:14:55.471 "aliases": [ 00:14:55.471 "1308dba5-f552-4ad5-9401-249b7671711e" 00:14:55.471 ], 00:14:55.471 "product_name": "Malloc disk", 00:14:55.471 "block_size": 512, 00:14:55.471 "num_blocks": 65536, 00:14:55.471 "uuid": "1308dba5-f552-4ad5-9401-249b7671711e", 00:14:55.471 "assigned_rate_limits": { 00:14:55.471 "rw_ios_per_sec": 0, 00:14:55.471 "rw_mbytes_per_sec": 0, 00:14:55.471 "r_mbytes_per_sec": 0, 00:14:55.471 "w_mbytes_per_sec": 0 00:14:55.471 }, 00:14:55.471 "claimed": true, 00:14:55.471 "claim_type": "exclusive_write", 00:14:55.471 "zoned": false, 00:14:55.471 "supported_io_types": { 00:14:55.471 "read": true, 00:14:55.471 "write": true, 00:14:55.471 "unmap": true, 00:14:55.471 "flush": true, 00:14:55.471 "reset": true, 00:14:55.471 "nvme_admin": false, 00:14:55.471 "nvme_io": false, 00:14:55.471 "nvme_io_md": false, 00:14:55.471 "write_zeroes": true, 00:14:55.471 "zcopy": true, 00:14:55.471 "get_zone_info": false, 00:14:55.471 "zone_management": false, 00:14:55.471 "zone_append": false, 00:14:55.471 "compare": false, 00:14:55.471 "compare_and_write": false, 00:14:55.471 "abort": true, 00:14:55.471 "seek_hole": false, 00:14:55.471 "seek_data": false, 00:14:55.471 "copy": true, 00:14:55.471 "nvme_iov_md": false 00:14:55.471 }, 00:14:55.471 "memory_domains": [ 00:14:55.471 { 00:14:55.471 "dma_device_id": "system", 00:14:55.471 "dma_device_type": 1 00:14:55.471 }, 00:14:55.471 { 00:14:55.471 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:55.471 "dma_device_type": 2 00:14:55.471 } 00:14:55.471 ], 00:14:55.471 "driver_specific": {} 00:14:55.471 } 00:14:55.471 ] 00:14:55.471 17:36:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:55.471 17:36:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:14:55.471 17:36:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:55.471 17:36:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:55.471 17:36:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:55.471 17:36:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:55.471 17:36:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:55.471 17:36:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:55.471 17:36:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:55.471 17:36:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:55.471 17:36:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:55.471 17:36:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:55.471 17:36:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:55.471 17:36:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:55.471 17:36:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:55.471 17:36:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:55.732 17:36:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:55.732 17:36:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:55.732 "name": "Existed_Raid", 00:14:55.732 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:55.732 "strip_size_kb": 64, 00:14:55.732 "state": "configuring", 00:14:55.732 "raid_level": "raid5f", 00:14:55.732 "superblock": false, 00:14:55.732 "num_base_bdevs": 4, 00:14:55.732 "num_base_bdevs_discovered": 3, 00:14:55.732 "num_base_bdevs_operational": 4, 00:14:55.732 "base_bdevs_list": [ 00:14:55.732 { 00:14:55.732 "name": "BaseBdev1", 00:14:55.732 "uuid": "1308dba5-f552-4ad5-9401-249b7671711e", 00:14:55.732 "is_configured": true, 00:14:55.732 "data_offset": 0, 00:14:55.732 "data_size": 65536 00:14:55.732 }, 00:14:55.732 { 00:14:55.732 "name": null, 00:14:55.732 "uuid": "73170cc6-caa1-4cf3-a5c7-73ecf0b2f958", 00:14:55.732 "is_configured": false, 00:14:55.732 "data_offset": 0, 00:14:55.732 "data_size": 65536 00:14:55.732 }, 00:14:55.732 { 00:14:55.732 "name": "BaseBdev3", 00:14:55.732 "uuid": "317ec481-ef9e-4fd7-8648-5492f42ebb25", 00:14:55.732 "is_configured": true, 00:14:55.732 "data_offset": 0, 00:14:55.732 "data_size": 65536 00:14:55.732 }, 00:14:55.732 { 00:14:55.732 "name": "BaseBdev4", 00:14:55.732 "uuid": "f5214f25-4720-4ddd-8ac1-03d864841930", 00:14:55.732 "is_configured": true, 00:14:55.732 "data_offset": 0, 00:14:55.732 "data_size": 65536 00:14:55.732 } 00:14:55.732 ] 00:14:55.732 }' 00:14:55.732 17:36:26 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:55.732 17:36:26 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:55.993 17:36:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:55.993 17:36:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:14:55.993 17:36:27 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:55.993 17:36:27 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:55.993 17:36:27 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:55.993 17:36:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:14:55.993 17:36:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:14:55.993 17:36:27 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:55.993 17:36:27 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:55.993 [2024-11-27 17:36:27.171046] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:14:55.993 17:36:27 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:55.993 17:36:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:55.993 17:36:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:55.993 17:36:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:55.993 17:36:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:55.993 17:36:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:55.993 17:36:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:55.993 17:36:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:55.993 17:36:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:55.993 17:36:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:55.993 17:36:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:56.254 17:36:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:56.254 17:36:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:56.254 17:36:27 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:56.254 17:36:27 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:56.254 17:36:27 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:56.254 17:36:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:56.254 "name": "Existed_Raid", 00:14:56.254 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:56.254 "strip_size_kb": 64, 00:14:56.254 "state": "configuring", 00:14:56.254 "raid_level": "raid5f", 00:14:56.254 "superblock": false, 00:14:56.254 "num_base_bdevs": 4, 00:14:56.254 "num_base_bdevs_discovered": 2, 00:14:56.254 "num_base_bdevs_operational": 4, 00:14:56.254 "base_bdevs_list": [ 00:14:56.254 { 00:14:56.254 "name": "BaseBdev1", 00:14:56.254 "uuid": "1308dba5-f552-4ad5-9401-249b7671711e", 00:14:56.254 "is_configured": true, 00:14:56.254 "data_offset": 0, 00:14:56.254 "data_size": 65536 00:14:56.254 }, 00:14:56.254 { 00:14:56.254 "name": null, 00:14:56.254 "uuid": "73170cc6-caa1-4cf3-a5c7-73ecf0b2f958", 00:14:56.254 "is_configured": false, 00:14:56.254 "data_offset": 0, 00:14:56.254 "data_size": 65536 00:14:56.254 }, 00:14:56.254 { 00:14:56.254 "name": null, 00:14:56.254 "uuid": "317ec481-ef9e-4fd7-8648-5492f42ebb25", 00:14:56.254 "is_configured": false, 00:14:56.254 "data_offset": 0, 00:14:56.254 "data_size": 65536 00:14:56.254 }, 00:14:56.254 { 00:14:56.254 "name": "BaseBdev4", 00:14:56.254 "uuid": "f5214f25-4720-4ddd-8ac1-03d864841930", 00:14:56.254 "is_configured": true, 00:14:56.254 "data_offset": 0, 00:14:56.254 "data_size": 65536 00:14:56.254 } 00:14:56.254 ] 00:14:56.254 }' 00:14:56.254 17:36:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:56.255 17:36:27 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:56.515 17:36:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:14:56.515 17:36:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:56.515 17:36:27 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:56.515 17:36:27 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:56.515 17:36:27 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:56.515 17:36:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:14:56.515 17:36:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:14:56.515 17:36:27 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:56.515 17:36:27 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:56.515 [2024-11-27 17:36:27.666280] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:14:56.515 17:36:27 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:56.515 17:36:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:56.515 17:36:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:56.515 17:36:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:56.515 17:36:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:56.515 17:36:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:56.515 17:36:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:56.515 17:36:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:56.515 17:36:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:56.515 17:36:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:56.515 17:36:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:56.515 17:36:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:56.515 17:36:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:56.515 17:36:27 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:56.515 17:36:27 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:56.515 17:36:27 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:56.775 17:36:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:56.775 "name": "Existed_Raid", 00:14:56.775 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:56.775 "strip_size_kb": 64, 00:14:56.775 "state": "configuring", 00:14:56.775 "raid_level": "raid5f", 00:14:56.775 "superblock": false, 00:14:56.775 "num_base_bdevs": 4, 00:14:56.775 "num_base_bdevs_discovered": 3, 00:14:56.775 "num_base_bdevs_operational": 4, 00:14:56.775 "base_bdevs_list": [ 00:14:56.775 { 00:14:56.775 "name": "BaseBdev1", 00:14:56.775 "uuid": "1308dba5-f552-4ad5-9401-249b7671711e", 00:14:56.775 "is_configured": true, 00:14:56.775 "data_offset": 0, 00:14:56.775 "data_size": 65536 00:14:56.775 }, 00:14:56.775 { 00:14:56.775 "name": null, 00:14:56.775 "uuid": "73170cc6-caa1-4cf3-a5c7-73ecf0b2f958", 00:14:56.775 "is_configured": false, 00:14:56.775 "data_offset": 0, 00:14:56.775 "data_size": 65536 00:14:56.775 }, 00:14:56.775 { 00:14:56.775 "name": "BaseBdev3", 00:14:56.775 "uuid": "317ec481-ef9e-4fd7-8648-5492f42ebb25", 00:14:56.775 "is_configured": true, 00:14:56.775 "data_offset": 0, 00:14:56.775 "data_size": 65536 00:14:56.775 }, 00:14:56.775 { 00:14:56.775 "name": "BaseBdev4", 00:14:56.775 "uuid": "f5214f25-4720-4ddd-8ac1-03d864841930", 00:14:56.775 "is_configured": true, 00:14:56.775 "data_offset": 0, 00:14:56.775 "data_size": 65536 00:14:56.775 } 00:14:56.775 ] 00:14:56.775 }' 00:14:56.775 17:36:27 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:56.775 17:36:27 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:57.035 17:36:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:57.035 17:36:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:14:57.035 17:36:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:57.035 17:36:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:57.035 17:36:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:57.035 17:36:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:14:57.035 17:36:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:14:57.035 17:36:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:57.035 17:36:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:57.035 [2024-11-27 17:36:28.181453] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:14:57.035 17:36:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:57.035 17:36:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:57.035 17:36:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:57.035 17:36:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:57.035 17:36:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:57.035 17:36:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:57.035 17:36:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:57.035 17:36:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:57.035 17:36:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:57.035 17:36:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:57.035 17:36:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:57.035 17:36:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:57.035 17:36:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:57.035 17:36:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:57.035 17:36:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:57.035 17:36:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:57.296 17:36:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:57.296 "name": "Existed_Raid", 00:14:57.296 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:57.296 "strip_size_kb": 64, 00:14:57.296 "state": "configuring", 00:14:57.296 "raid_level": "raid5f", 00:14:57.296 "superblock": false, 00:14:57.296 "num_base_bdevs": 4, 00:14:57.296 "num_base_bdevs_discovered": 2, 00:14:57.296 "num_base_bdevs_operational": 4, 00:14:57.296 "base_bdevs_list": [ 00:14:57.296 { 00:14:57.296 "name": null, 00:14:57.296 "uuid": "1308dba5-f552-4ad5-9401-249b7671711e", 00:14:57.296 "is_configured": false, 00:14:57.296 "data_offset": 0, 00:14:57.296 "data_size": 65536 00:14:57.296 }, 00:14:57.296 { 00:14:57.296 "name": null, 00:14:57.296 "uuid": "73170cc6-caa1-4cf3-a5c7-73ecf0b2f958", 00:14:57.296 "is_configured": false, 00:14:57.296 "data_offset": 0, 00:14:57.296 "data_size": 65536 00:14:57.296 }, 00:14:57.296 { 00:14:57.296 "name": "BaseBdev3", 00:14:57.296 "uuid": "317ec481-ef9e-4fd7-8648-5492f42ebb25", 00:14:57.296 "is_configured": true, 00:14:57.296 "data_offset": 0, 00:14:57.296 "data_size": 65536 00:14:57.296 }, 00:14:57.296 { 00:14:57.296 "name": "BaseBdev4", 00:14:57.296 "uuid": "f5214f25-4720-4ddd-8ac1-03d864841930", 00:14:57.296 "is_configured": true, 00:14:57.296 "data_offset": 0, 00:14:57.296 "data_size": 65536 00:14:57.296 } 00:14:57.296 ] 00:14:57.296 }' 00:14:57.296 17:36:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:57.296 17:36:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:57.556 17:36:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:57.556 17:36:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:14:57.556 17:36:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:57.556 17:36:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:57.556 17:36:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:57.556 17:36:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:14:57.557 17:36:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:14:57.557 17:36:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:57.557 17:36:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:57.557 [2024-11-27 17:36:28.691116] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:14:57.557 17:36:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:57.557 17:36:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:14:57.557 17:36:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:57.557 17:36:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:14:57.557 17:36:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:57.557 17:36:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:57.557 17:36:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:57.557 17:36:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:57.557 17:36:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:57.557 17:36:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:57.557 17:36:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:57.557 17:36:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:57.557 17:36:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:57.557 17:36:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:57.557 17:36:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:57.557 17:36:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:57.557 17:36:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:57.557 "name": "Existed_Raid", 00:14:57.557 "uuid": "00000000-0000-0000-0000-000000000000", 00:14:57.557 "strip_size_kb": 64, 00:14:57.557 "state": "configuring", 00:14:57.557 "raid_level": "raid5f", 00:14:57.557 "superblock": false, 00:14:57.557 "num_base_bdevs": 4, 00:14:57.557 "num_base_bdevs_discovered": 3, 00:14:57.557 "num_base_bdevs_operational": 4, 00:14:57.557 "base_bdevs_list": [ 00:14:57.557 { 00:14:57.557 "name": null, 00:14:57.557 "uuid": "1308dba5-f552-4ad5-9401-249b7671711e", 00:14:57.557 "is_configured": false, 00:14:57.557 "data_offset": 0, 00:14:57.557 "data_size": 65536 00:14:57.557 }, 00:14:57.557 { 00:14:57.557 "name": "BaseBdev2", 00:14:57.557 "uuid": "73170cc6-caa1-4cf3-a5c7-73ecf0b2f958", 00:14:57.557 "is_configured": true, 00:14:57.557 "data_offset": 0, 00:14:57.557 "data_size": 65536 00:14:57.557 }, 00:14:57.557 { 00:14:57.557 "name": "BaseBdev3", 00:14:57.557 "uuid": "317ec481-ef9e-4fd7-8648-5492f42ebb25", 00:14:57.557 "is_configured": true, 00:14:57.557 "data_offset": 0, 00:14:57.557 "data_size": 65536 00:14:57.557 }, 00:14:57.557 { 00:14:57.557 "name": "BaseBdev4", 00:14:57.557 "uuid": "f5214f25-4720-4ddd-8ac1-03d864841930", 00:14:57.557 "is_configured": true, 00:14:57.557 "data_offset": 0, 00:14:57.557 "data_size": 65536 00:14:57.557 } 00:14:57.557 ] 00:14:57.557 }' 00:14:57.557 17:36:28 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:57.557 17:36:28 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:58.130 17:36:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:58.130 17:36:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:58.130 17:36:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:58.130 17:36:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:14:58.130 17:36:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:58.130 17:36:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:14:58.130 17:36:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:58.130 17:36:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:58.130 17:36:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:58.130 17:36:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:14:58.130 17:36:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:58.130 17:36:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 1308dba5-f552-4ad5-9401-249b7671711e 00:14:58.130 17:36:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:58.130 17:36:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:58.130 [2024-11-27 17:36:29.273364] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:14:58.130 [2024-11-27 17:36:29.273468] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:14:58.130 [2024-11-27 17:36:29.273479] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 196608, blocklen 512 00:14:58.130 [2024-11-27 17:36:29.273738] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002a10 00:14:58.130 [2024-11-27 17:36:29.274162] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:14:58.130 [2024-11-27 17:36:29.274177] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001c80 00:14:58.130 [2024-11-27 17:36:29.274354] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:14:58.130 NewBaseBdev 00:14:58.130 17:36:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:58.130 17:36:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:14:58.130 17:36:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@899 -- # local bdev_name=NewBaseBdev 00:14:58.130 17:36:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:14:58.131 17:36:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@901 -- # local i 00:14:58.131 17:36:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:14:58.131 17:36:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:14:58.131 17:36:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:14:58.131 17:36:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:58.131 17:36:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:58.131 17:36:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:58.131 17:36:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:14:58.131 17:36:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:58.131 17:36:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:58.131 [ 00:14:58.131 { 00:14:58.131 "name": "NewBaseBdev", 00:14:58.131 "aliases": [ 00:14:58.131 "1308dba5-f552-4ad5-9401-249b7671711e" 00:14:58.131 ], 00:14:58.131 "product_name": "Malloc disk", 00:14:58.131 "block_size": 512, 00:14:58.131 "num_blocks": 65536, 00:14:58.131 "uuid": "1308dba5-f552-4ad5-9401-249b7671711e", 00:14:58.131 "assigned_rate_limits": { 00:14:58.131 "rw_ios_per_sec": 0, 00:14:58.131 "rw_mbytes_per_sec": 0, 00:14:58.131 "r_mbytes_per_sec": 0, 00:14:58.131 "w_mbytes_per_sec": 0 00:14:58.131 }, 00:14:58.131 "claimed": true, 00:14:58.131 "claim_type": "exclusive_write", 00:14:58.131 "zoned": false, 00:14:58.131 "supported_io_types": { 00:14:58.131 "read": true, 00:14:58.131 "write": true, 00:14:58.131 "unmap": true, 00:14:58.131 "flush": true, 00:14:58.131 "reset": true, 00:14:58.131 "nvme_admin": false, 00:14:58.131 "nvme_io": false, 00:14:58.131 "nvme_io_md": false, 00:14:58.132 "write_zeroes": true, 00:14:58.132 "zcopy": true, 00:14:58.132 "get_zone_info": false, 00:14:58.132 "zone_management": false, 00:14:58.132 "zone_append": false, 00:14:58.132 "compare": false, 00:14:58.132 "compare_and_write": false, 00:14:58.132 "abort": true, 00:14:58.132 "seek_hole": false, 00:14:58.132 "seek_data": false, 00:14:58.132 "copy": true, 00:14:58.132 "nvme_iov_md": false 00:14:58.132 }, 00:14:58.132 "memory_domains": [ 00:14:58.132 { 00:14:58.132 "dma_device_id": "system", 00:14:58.132 "dma_device_type": 1 00:14:58.132 }, 00:14:58.132 { 00:14:58.132 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:58.132 "dma_device_type": 2 00:14:58.132 } 00:14:58.132 ], 00:14:58.132 "driver_specific": {} 00:14:58.132 } 00:14:58.132 ] 00:14:58.132 17:36:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:58.132 17:36:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@907 -- # return 0 00:14:58.132 17:36:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 4 00:14:58.132 17:36:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:14:58.132 17:36:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:14:58.132 17:36:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:14:58.132 17:36:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:14:58.132 17:36:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:14:58.132 17:36:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:14:58.132 17:36:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:14:58.132 17:36:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:14:58.392 17:36:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:14:58.392 17:36:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:14:58.392 17:36:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:14:58.392 17:36:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:58.392 17:36:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:58.392 17:36:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:58.392 17:36:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:14:58.392 "name": "Existed_Raid", 00:14:58.392 "uuid": "733029d4-ca69-4deb-82c0-a746dec90701", 00:14:58.392 "strip_size_kb": 64, 00:14:58.392 "state": "online", 00:14:58.392 "raid_level": "raid5f", 00:14:58.392 "superblock": false, 00:14:58.392 "num_base_bdevs": 4, 00:14:58.392 "num_base_bdevs_discovered": 4, 00:14:58.392 "num_base_bdevs_operational": 4, 00:14:58.392 "base_bdevs_list": [ 00:14:58.392 { 00:14:58.392 "name": "NewBaseBdev", 00:14:58.392 "uuid": "1308dba5-f552-4ad5-9401-249b7671711e", 00:14:58.392 "is_configured": true, 00:14:58.392 "data_offset": 0, 00:14:58.392 "data_size": 65536 00:14:58.392 }, 00:14:58.392 { 00:14:58.392 "name": "BaseBdev2", 00:14:58.392 "uuid": "73170cc6-caa1-4cf3-a5c7-73ecf0b2f958", 00:14:58.392 "is_configured": true, 00:14:58.392 "data_offset": 0, 00:14:58.392 "data_size": 65536 00:14:58.392 }, 00:14:58.392 { 00:14:58.392 "name": "BaseBdev3", 00:14:58.392 "uuid": "317ec481-ef9e-4fd7-8648-5492f42ebb25", 00:14:58.392 "is_configured": true, 00:14:58.392 "data_offset": 0, 00:14:58.392 "data_size": 65536 00:14:58.392 }, 00:14:58.392 { 00:14:58.392 "name": "BaseBdev4", 00:14:58.392 "uuid": "f5214f25-4720-4ddd-8ac1-03d864841930", 00:14:58.392 "is_configured": true, 00:14:58.392 "data_offset": 0, 00:14:58.392 "data_size": 65536 00:14:58.392 } 00:14:58.392 ] 00:14:58.392 }' 00:14:58.392 17:36:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:14:58.392 17:36:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:58.653 17:36:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:14:58.653 17:36:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:14:58.653 17:36:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:14:58.653 17:36:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:14:58.653 17:36:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@184 -- # local name 00:14:58.653 17:36:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:14:58.653 17:36:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:14:58.653 17:36:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:14:58.653 17:36:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:58.653 17:36:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:58.653 [2024-11-27 17:36:29.780699] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:14:58.653 17:36:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:58.653 17:36:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:14:58.653 "name": "Existed_Raid", 00:14:58.653 "aliases": [ 00:14:58.653 "733029d4-ca69-4deb-82c0-a746dec90701" 00:14:58.653 ], 00:14:58.653 "product_name": "Raid Volume", 00:14:58.653 "block_size": 512, 00:14:58.653 "num_blocks": 196608, 00:14:58.653 "uuid": "733029d4-ca69-4deb-82c0-a746dec90701", 00:14:58.653 "assigned_rate_limits": { 00:14:58.653 "rw_ios_per_sec": 0, 00:14:58.653 "rw_mbytes_per_sec": 0, 00:14:58.653 "r_mbytes_per_sec": 0, 00:14:58.653 "w_mbytes_per_sec": 0 00:14:58.653 }, 00:14:58.653 "claimed": false, 00:14:58.653 "zoned": false, 00:14:58.653 "supported_io_types": { 00:14:58.653 "read": true, 00:14:58.653 "write": true, 00:14:58.653 "unmap": false, 00:14:58.653 "flush": false, 00:14:58.653 "reset": true, 00:14:58.653 "nvme_admin": false, 00:14:58.653 "nvme_io": false, 00:14:58.653 "nvme_io_md": false, 00:14:58.653 "write_zeroes": true, 00:14:58.653 "zcopy": false, 00:14:58.653 "get_zone_info": false, 00:14:58.653 "zone_management": false, 00:14:58.653 "zone_append": false, 00:14:58.653 "compare": false, 00:14:58.653 "compare_and_write": false, 00:14:58.653 "abort": false, 00:14:58.653 "seek_hole": false, 00:14:58.653 "seek_data": false, 00:14:58.653 "copy": false, 00:14:58.653 "nvme_iov_md": false 00:14:58.653 }, 00:14:58.653 "driver_specific": { 00:14:58.653 "raid": { 00:14:58.653 "uuid": "733029d4-ca69-4deb-82c0-a746dec90701", 00:14:58.653 "strip_size_kb": 64, 00:14:58.653 "state": "online", 00:14:58.653 "raid_level": "raid5f", 00:14:58.653 "superblock": false, 00:14:58.653 "num_base_bdevs": 4, 00:14:58.653 "num_base_bdevs_discovered": 4, 00:14:58.653 "num_base_bdevs_operational": 4, 00:14:58.653 "base_bdevs_list": [ 00:14:58.653 { 00:14:58.653 "name": "NewBaseBdev", 00:14:58.653 "uuid": "1308dba5-f552-4ad5-9401-249b7671711e", 00:14:58.653 "is_configured": true, 00:14:58.653 "data_offset": 0, 00:14:58.653 "data_size": 65536 00:14:58.653 }, 00:14:58.653 { 00:14:58.653 "name": "BaseBdev2", 00:14:58.653 "uuid": "73170cc6-caa1-4cf3-a5c7-73ecf0b2f958", 00:14:58.653 "is_configured": true, 00:14:58.653 "data_offset": 0, 00:14:58.653 "data_size": 65536 00:14:58.653 }, 00:14:58.653 { 00:14:58.653 "name": "BaseBdev3", 00:14:58.653 "uuid": "317ec481-ef9e-4fd7-8648-5492f42ebb25", 00:14:58.653 "is_configured": true, 00:14:58.653 "data_offset": 0, 00:14:58.653 "data_size": 65536 00:14:58.653 }, 00:14:58.653 { 00:14:58.653 "name": "BaseBdev4", 00:14:58.653 "uuid": "f5214f25-4720-4ddd-8ac1-03d864841930", 00:14:58.653 "is_configured": true, 00:14:58.653 "data_offset": 0, 00:14:58.653 "data_size": 65536 00:14:58.653 } 00:14:58.653 ] 00:14:58.653 } 00:14:58.653 } 00:14:58.653 }' 00:14:58.653 17:36:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:14:58.914 17:36:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:14:58.914 BaseBdev2 00:14:58.914 BaseBdev3 00:14:58.914 BaseBdev4' 00:14:58.914 17:36:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:58.914 17:36:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:14:58.914 17:36:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:58.914 17:36:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:58.914 17:36:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:14:58.914 17:36:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:58.914 17:36:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:58.914 17:36:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:58.914 17:36:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:58.914 17:36:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:58.914 17:36:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:58.914 17:36:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:58.914 17:36:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:14:58.914 17:36:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:58.914 17:36:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:58.914 17:36:29 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:58.914 17:36:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:58.914 17:36:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:58.914 17:36:29 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:58.914 17:36:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:14:58.914 17:36:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:58.914 17:36:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:58.914 17:36:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:58.914 17:36:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:58.914 17:36:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:58.914 17:36:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:58.914 17:36:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:14:58.914 17:36:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:14:58.914 17:36:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:14:58.914 17:36:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:58.914 17:36:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:58.914 17:36:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:58.914 17:36:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:14:58.915 17:36:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:14:58.915 17:36:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:14:58.915 17:36:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:58.915 17:36:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:58.915 [2024-11-27 17:36:30.087984] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:14:58.915 [2024-11-27 17:36:30.088008] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:14:58.915 [2024-11-27 17:36:30.088069] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:14:58.915 [2024-11-27 17:36:30.088313] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:14:58.915 [2024-11-27 17:36:30.088331] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name Existed_Raid, state offline 00:14:58.915 17:36:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:58.915 17:36:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@326 -- # killprocess 93100 00:14:58.915 17:36:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@950 -- # '[' -z 93100 ']' 00:14:58.915 17:36:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@954 -- # kill -0 93100 00:14:58.915 17:36:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@955 -- # uname 00:14:58.915 17:36:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:14:59.175 17:36:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 93100 00:14:59.175 17:36:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:14:59.175 17:36:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:14:59.175 17:36:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 93100' 00:14:59.175 killing process with pid 93100 00:14:59.175 17:36:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@969 -- # kill 93100 00:14:59.175 [2024-11-27 17:36:30.137783] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:14:59.175 17:36:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@974 -- # wait 93100 00:14:59.175 [2024-11-27 17:36:30.177755] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:14:59.436 17:36:30 bdev_raid.raid5f_state_function_test -- bdev/bdev_raid.sh@328 -- # return 0 00:14:59.436 00:14:59.436 real 0m9.928s 00:14:59.436 user 0m16.968s 00:14:59.436 sys 0m2.178s 00:14:59.436 17:36:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:14:59.436 ************************************ 00:14:59.436 END TEST raid5f_state_function_test 00:14:59.436 ************************************ 00:14:59.436 17:36:30 bdev_raid.raid5f_state_function_test -- common/autotest_common.sh@10 -- # set +x 00:14:59.436 17:36:30 bdev_raid -- bdev/bdev_raid.sh@987 -- # run_test raid5f_state_function_test_sb raid_state_function_test raid5f 4 true 00:14:59.436 17:36:30 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:14:59.436 17:36:30 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:14:59.436 17:36:30 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:14:59.436 ************************************ 00:14:59.436 START TEST raid5f_state_function_test_sb 00:14:59.436 ************************************ 00:14:59.436 17:36:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@1125 -- # raid_state_function_test raid5f 4 true 00:14:59.436 17:36:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@205 -- # local raid_level=raid5f 00:14:59.436 17:36:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=4 00:14:59.436 17:36:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:14:59.436 17:36:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:14:59.436 17:36:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:14:59.436 17:36:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:14:59.436 17:36:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:14:59.436 17:36:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:14:59.436 17:36:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:14:59.436 17:36:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:14:59.436 17:36:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:14:59.436 17:36:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:14:59.436 17:36:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev3 00:14:59.436 17:36:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:14:59.436 17:36:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:14:59.436 17:36:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # echo BaseBdev4 00:14:59.436 17:36:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:14:59.436 17:36:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:14:59.436 17:36:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:14:59.436 17:36:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:14:59.436 17:36:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:14:59.436 17:36:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@211 -- # local strip_size 00:14:59.436 17:36:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:14:59.436 17:36:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:14:59.436 17:36:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@215 -- # '[' raid5f '!=' raid1 ']' 00:14:59.436 17:36:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@216 -- # strip_size=64 00:14:59.436 17:36:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@217 -- # strip_size_create_arg='-z 64' 00:14:59.436 17:36:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:14:59.436 17:36:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:14:59.436 17:36:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@229 -- # raid_pid=93750 00:14:59.436 17:36:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:14:59.436 17:36:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 93750' 00:14:59.436 Process raid pid: 93750 00:14:59.436 17:36:30 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@231 -- # waitforlisten 93750 00:14:59.436 17:36:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@831 -- # '[' -z 93750 ']' 00:14:59.436 17:36:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:14:59.436 17:36:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@836 -- # local max_retries=100 00:14:59.436 17:36:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:14:59.436 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:14:59.436 17:36:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@840 -- # xtrace_disable 00:14:59.436 17:36:30 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:14:59.436 [2024-11-27 17:36:30.606236] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:14:59.436 [2024-11-27 17:36:30.606425] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:14:59.697 [2024-11-27 17:36:30.753301] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:14:59.697 [2024-11-27 17:36:30.799332] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:14:59.697 [2024-11-27 17:36:30.842122] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:14:59.697 [2024-11-27 17:36:30.842236] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:00.268 17:36:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:15:00.268 17:36:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@864 -- # return 0 00:15:00.268 17:36:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:15:00.268 17:36:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:00.268 17:36:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:00.268 [2024-11-27 17:36:31.428051] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:15:00.268 [2024-11-27 17:36:31.428192] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:15:00.268 [2024-11-27 17:36:31.428224] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:15:00.268 [2024-11-27 17:36:31.428247] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:15:00.268 [2024-11-27 17:36:31.428264] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:15:00.268 [2024-11-27 17:36:31.428296] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:15:00.268 [2024-11-27 17:36:31.428317] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:15:00.268 [2024-11-27 17:36:31.428336] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:15:00.268 17:36:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:00.268 17:36:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:15:00.268 17:36:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:00.269 17:36:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:00.269 17:36:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:00.269 17:36:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:00.269 17:36:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:00.269 17:36:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:00.269 17:36:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:00.269 17:36:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:00.269 17:36:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:00.269 17:36:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:00.269 17:36:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:00.269 17:36:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:00.269 17:36:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:00.529 17:36:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:00.529 17:36:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:00.529 "name": "Existed_Raid", 00:15:00.529 "uuid": "0503e220-fcc0-431f-9e3b-79ffd8880670", 00:15:00.529 "strip_size_kb": 64, 00:15:00.529 "state": "configuring", 00:15:00.529 "raid_level": "raid5f", 00:15:00.529 "superblock": true, 00:15:00.529 "num_base_bdevs": 4, 00:15:00.529 "num_base_bdevs_discovered": 0, 00:15:00.529 "num_base_bdevs_operational": 4, 00:15:00.529 "base_bdevs_list": [ 00:15:00.529 { 00:15:00.529 "name": "BaseBdev1", 00:15:00.529 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:00.529 "is_configured": false, 00:15:00.529 "data_offset": 0, 00:15:00.529 "data_size": 0 00:15:00.529 }, 00:15:00.529 { 00:15:00.529 "name": "BaseBdev2", 00:15:00.529 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:00.529 "is_configured": false, 00:15:00.529 "data_offset": 0, 00:15:00.529 "data_size": 0 00:15:00.529 }, 00:15:00.529 { 00:15:00.529 "name": "BaseBdev3", 00:15:00.529 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:00.529 "is_configured": false, 00:15:00.529 "data_offset": 0, 00:15:00.529 "data_size": 0 00:15:00.529 }, 00:15:00.529 { 00:15:00.529 "name": "BaseBdev4", 00:15:00.529 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:00.529 "is_configured": false, 00:15:00.529 "data_offset": 0, 00:15:00.529 "data_size": 0 00:15:00.529 } 00:15:00.529 ] 00:15:00.529 }' 00:15:00.529 17:36:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:00.529 17:36:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:00.790 17:36:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:15:00.790 17:36:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:00.790 17:36:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:00.790 [2024-11-27 17:36:31.883136] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:15:00.790 [2024-11-27 17:36:31.883238] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:15:00.790 17:36:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:00.790 17:36:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:15:00.790 17:36:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:00.790 17:36:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:00.790 [2024-11-27 17:36:31.895168] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:15:00.790 [2024-11-27 17:36:31.895251] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:15:00.790 [2024-11-27 17:36:31.895278] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:15:00.790 [2024-11-27 17:36:31.895300] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:15:00.790 [2024-11-27 17:36:31.895317] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:15:00.790 [2024-11-27 17:36:31.895338] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:15:00.790 [2024-11-27 17:36:31.895355] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:15:00.790 [2024-11-27 17:36:31.895374] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:15:00.790 17:36:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:00.790 17:36:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:15:00.790 17:36:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:00.790 17:36:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:00.790 [2024-11-27 17:36:31.916183] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:15:00.790 BaseBdev1 00:15:00.790 17:36:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:00.790 17:36:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:15:00.790 17:36:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:15:00.790 17:36:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:15:00.790 17:36:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:15:00.790 17:36:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:15:00.790 17:36:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:15:00.790 17:36:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:15:00.790 17:36:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:00.790 17:36:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:00.790 17:36:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:00.790 17:36:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:15:00.790 17:36:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:00.790 17:36:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:00.790 [ 00:15:00.790 { 00:15:00.790 "name": "BaseBdev1", 00:15:00.790 "aliases": [ 00:15:00.790 "47126875-78d1-4e7e-9b3e-80cec5bc212d" 00:15:00.790 ], 00:15:00.790 "product_name": "Malloc disk", 00:15:00.790 "block_size": 512, 00:15:00.790 "num_blocks": 65536, 00:15:00.791 "uuid": "47126875-78d1-4e7e-9b3e-80cec5bc212d", 00:15:00.791 "assigned_rate_limits": { 00:15:00.791 "rw_ios_per_sec": 0, 00:15:00.791 "rw_mbytes_per_sec": 0, 00:15:00.791 "r_mbytes_per_sec": 0, 00:15:00.791 "w_mbytes_per_sec": 0 00:15:00.791 }, 00:15:00.791 "claimed": true, 00:15:00.791 "claim_type": "exclusive_write", 00:15:00.791 "zoned": false, 00:15:00.791 "supported_io_types": { 00:15:00.791 "read": true, 00:15:00.791 "write": true, 00:15:00.791 "unmap": true, 00:15:00.791 "flush": true, 00:15:00.791 "reset": true, 00:15:00.791 "nvme_admin": false, 00:15:00.791 "nvme_io": false, 00:15:00.791 "nvme_io_md": false, 00:15:00.791 "write_zeroes": true, 00:15:00.791 "zcopy": true, 00:15:00.791 "get_zone_info": false, 00:15:00.791 "zone_management": false, 00:15:00.791 "zone_append": false, 00:15:00.791 "compare": false, 00:15:00.791 "compare_and_write": false, 00:15:00.791 "abort": true, 00:15:00.791 "seek_hole": false, 00:15:00.791 "seek_data": false, 00:15:00.791 "copy": true, 00:15:00.791 "nvme_iov_md": false 00:15:00.791 }, 00:15:00.791 "memory_domains": [ 00:15:00.791 { 00:15:00.791 "dma_device_id": "system", 00:15:00.791 "dma_device_type": 1 00:15:00.791 }, 00:15:00.791 { 00:15:00.791 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:00.791 "dma_device_type": 2 00:15:00.791 } 00:15:00.791 ], 00:15:00.791 "driver_specific": {} 00:15:00.791 } 00:15:00.791 ] 00:15:00.791 17:36:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:00.791 17:36:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:15:00.791 17:36:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:15:00.791 17:36:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:00.791 17:36:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:00.791 17:36:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:00.791 17:36:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:00.791 17:36:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:00.791 17:36:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:00.791 17:36:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:00.791 17:36:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:00.791 17:36:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:00.791 17:36:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:00.791 17:36:31 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:00.791 17:36:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:00.791 17:36:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:01.051 17:36:31 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:01.051 17:36:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:01.051 "name": "Existed_Raid", 00:15:01.051 "uuid": "2fe24581-9e34-4528-ad61-c7260d30ca40", 00:15:01.051 "strip_size_kb": 64, 00:15:01.051 "state": "configuring", 00:15:01.051 "raid_level": "raid5f", 00:15:01.051 "superblock": true, 00:15:01.051 "num_base_bdevs": 4, 00:15:01.051 "num_base_bdevs_discovered": 1, 00:15:01.051 "num_base_bdevs_operational": 4, 00:15:01.051 "base_bdevs_list": [ 00:15:01.051 { 00:15:01.051 "name": "BaseBdev1", 00:15:01.051 "uuid": "47126875-78d1-4e7e-9b3e-80cec5bc212d", 00:15:01.051 "is_configured": true, 00:15:01.051 "data_offset": 2048, 00:15:01.051 "data_size": 63488 00:15:01.051 }, 00:15:01.051 { 00:15:01.051 "name": "BaseBdev2", 00:15:01.051 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:01.051 "is_configured": false, 00:15:01.051 "data_offset": 0, 00:15:01.051 "data_size": 0 00:15:01.051 }, 00:15:01.051 { 00:15:01.051 "name": "BaseBdev3", 00:15:01.051 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:01.051 "is_configured": false, 00:15:01.051 "data_offset": 0, 00:15:01.051 "data_size": 0 00:15:01.051 }, 00:15:01.051 { 00:15:01.051 "name": "BaseBdev4", 00:15:01.051 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:01.051 "is_configured": false, 00:15:01.051 "data_offset": 0, 00:15:01.051 "data_size": 0 00:15:01.051 } 00:15:01.051 ] 00:15:01.051 }' 00:15:01.051 17:36:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:01.051 17:36:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:01.311 17:36:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:15:01.311 17:36:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:01.311 17:36:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:01.311 [2024-11-27 17:36:32.383369] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:15:01.311 [2024-11-27 17:36:32.383457] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:15:01.311 17:36:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:01.311 17:36:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:15:01.311 17:36:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:01.311 17:36:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:01.311 [2024-11-27 17:36:32.395421] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:15:01.311 [2024-11-27 17:36:32.397157] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:15:01.311 [2024-11-27 17:36:32.397191] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:15:01.311 [2024-11-27 17:36:32.397199] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev3 00:15:01.311 [2024-11-27 17:36:32.397207] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev3 doesn't exist now 00:15:01.311 [2024-11-27 17:36:32.397213] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev4 00:15:01.311 [2024-11-27 17:36:32.397221] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev4 doesn't exist now 00:15:01.311 17:36:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:01.311 17:36:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:15:01.312 17:36:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:15:01.312 17:36:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:15:01.312 17:36:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:01.312 17:36:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:01.312 17:36:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:01.312 17:36:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:01.312 17:36:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:01.312 17:36:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:01.312 17:36:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:01.312 17:36:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:01.312 17:36:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:01.312 17:36:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:01.312 17:36:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:01.312 17:36:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:01.312 17:36:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:01.312 17:36:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:01.312 17:36:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:01.312 "name": "Existed_Raid", 00:15:01.312 "uuid": "c438b21a-ec43-4443-9684-58450606cfc7", 00:15:01.312 "strip_size_kb": 64, 00:15:01.312 "state": "configuring", 00:15:01.312 "raid_level": "raid5f", 00:15:01.312 "superblock": true, 00:15:01.312 "num_base_bdevs": 4, 00:15:01.312 "num_base_bdevs_discovered": 1, 00:15:01.312 "num_base_bdevs_operational": 4, 00:15:01.312 "base_bdevs_list": [ 00:15:01.312 { 00:15:01.312 "name": "BaseBdev1", 00:15:01.312 "uuid": "47126875-78d1-4e7e-9b3e-80cec5bc212d", 00:15:01.312 "is_configured": true, 00:15:01.312 "data_offset": 2048, 00:15:01.312 "data_size": 63488 00:15:01.312 }, 00:15:01.312 { 00:15:01.312 "name": "BaseBdev2", 00:15:01.312 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:01.312 "is_configured": false, 00:15:01.312 "data_offset": 0, 00:15:01.312 "data_size": 0 00:15:01.312 }, 00:15:01.312 { 00:15:01.312 "name": "BaseBdev3", 00:15:01.312 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:01.312 "is_configured": false, 00:15:01.312 "data_offset": 0, 00:15:01.312 "data_size": 0 00:15:01.312 }, 00:15:01.312 { 00:15:01.312 "name": "BaseBdev4", 00:15:01.312 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:01.312 "is_configured": false, 00:15:01.312 "data_offset": 0, 00:15:01.312 "data_size": 0 00:15:01.312 } 00:15:01.312 ] 00:15:01.312 }' 00:15:01.312 17:36:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:01.312 17:36:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:01.883 17:36:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:15:01.883 17:36:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:01.883 17:36:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:01.883 [2024-11-27 17:36:32.871380] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:15:01.883 BaseBdev2 00:15:01.883 17:36:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:01.883 17:36:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:15:01.883 17:36:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:15:01.883 17:36:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:15:01.883 17:36:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:15:01.883 17:36:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:15:01.883 17:36:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:15:01.883 17:36:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:15:01.883 17:36:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:01.883 17:36:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:01.883 17:36:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:01.883 17:36:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:15:01.883 17:36:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:01.883 17:36:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:01.883 [ 00:15:01.883 { 00:15:01.883 "name": "BaseBdev2", 00:15:01.883 "aliases": [ 00:15:01.883 "252dcc2a-faab-4713-9875-c2123fe04122" 00:15:01.883 ], 00:15:01.883 "product_name": "Malloc disk", 00:15:01.883 "block_size": 512, 00:15:01.883 "num_blocks": 65536, 00:15:01.883 "uuid": "252dcc2a-faab-4713-9875-c2123fe04122", 00:15:01.883 "assigned_rate_limits": { 00:15:01.883 "rw_ios_per_sec": 0, 00:15:01.883 "rw_mbytes_per_sec": 0, 00:15:01.883 "r_mbytes_per_sec": 0, 00:15:01.883 "w_mbytes_per_sec": 0 00:15:01.883 }, 00:15:01.883 "claimed": true, 00:15:01.883 "claim_type": "exclusive_write", 00:15:01.883 "zoned": false, 00:15:01.883 "supported_io_types": { 00:15:01.883 "read": true, 00:15:01.883 "write": true, 00:15:01.883 "unmap": true, 00:15:01.883 "flush": true, 00:15:01.883 "reset": true, 00:15:01.883 "nvme_admin": false, 00:15:01.883 "nvme_io": false, 00:15:01.883 "nvme_io_md": false, 00:15:01.883 "write_zeroes": true, 00:15:01.883 "zcopy": true, 00:15:01.883 "get_zone_info": false, 00:15:01.883 "zone_management": false, 00:15:01.883 "zone_append": false, 00:15:01.883 "compare": false, 00:15:01.883 "compare_and_write": false, 00:15:01.883 "abort": true, 00:15:01.883 "seek_hole": false, 00:15:01.883 "seek_data": false, 00:15:01.883 "copy": true, 00:15:01.883 "nvme_iov_md": false 00:15:01.883 }, 00:15:01.883 "memory_domains": [ 00:15:01.883 { 00:15:01.883 "dma_device_id": "system", 00:15:01.883 "dma_device_type": 1 00:15:01.883 }, 00:15:01.883 { 00:15:01.883 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:01.883 "dma_device_type": 2 00:15:01.883 } 00:15:01.883 ], 00:15:01.883 "driver_specific": {} 00:15:01.883 } 00:15:01.883 ] 00:15:01.883 17:36:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:01.883 17:36:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:15:01.883 17:36:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:15:01.883 17:36:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:15:01.883 17:36:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:15:01.883 17:36:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:01.883 17:36:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:01.883 17:36:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:01.883 17:36:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:01.883 17:36:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:01.883 17:36:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:01.883 17:36:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:01.883 17:36:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:01.883 17:36:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:01.883 17:36:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:01.883 17:36:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:01.883 17:36:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:01.883 17:36:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:01.883 17:36:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:01.883 17:36:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:01.883 "name": "Existed_Raid", 00:15:01.883 "uuid": "c438b21a-ec43-4443-9684-58450606cfc7", 00:15:01.883 "strip_size_kb": 64, 00:15:01.883 "state": "configuring", 00:15:01.883 "raid_level": "raid5f", 00:15:01.883 "superblock": true, 00:15:01.883 "num_base_bdevs": 4, 00:15:01.883 "num_base_bdevs_discovered": 2, 00:15:01.883 "num_base_bdevs_operational": 4, 00:15:01.883 "base_bdevs_list": [ 00:15:01.883 { 00:15:01.883 "name": "BaseBdev1", 00:15:01.883 "uuid": "47126875-78d1-4e7e-9b3e-80cec5bc212d", 00:15:01.883 "is_configured": true, 00:15:01.883 "data_offset": 2048, 00:15:01.883 "data_size": 63488 00:15:01.883 }, 00:15:01.883 { 00:15:01.883 "name": "BaseBdev2", 00:15:01.883 "uuid": "252dcc2a-faab-4713-9875-c2123fe04122", 00:15:01.883 "is_configured": true, 00:15:01.883 "data_offset": 2048, 00:15:01.883 "data_size": 63488 00:15:01.883 }, 00:15:01.883 { 00:15:01.883 "name": "BaseBdev3", 00:15:01.883 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:01.883 "is_configured": false, 00:15:01.883 "data_offset": 0, 00:15:01.883 "data_size": 0 00:15:01.883 }, 00:15:01.883 { 00:15:01.883 "name": "BaseBdev4", 00:15:01.883 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:01.883 "is_configured": false, 00:15:01.883 "data_offset": 0, 00:15:01.883 "data_size": 0 00:15:01.883 } 00:15:01.883 ] 00:15:01.883 }' 00:15:01.883 17:36:32 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:01.883 17:36:32 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:02.483 17:36:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:15:02.483 17:36:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:02.483 17:36:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:02.483 BaseBdev3 00:15:02.483 [2024-11-27 17:36:33.381411] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:15:02.483 17:36:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:02.483 17:36:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev3 00:15:02.483 17:36:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:15:02.483 17:36:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:15:02.483 17:36:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:15:02.483 17:36:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:15:02.483 17:36:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:15:02.483 17:36:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:15:02.483 17:36:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:02.483 17:36:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:02.483 17:36:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:02.483 17:36:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:15:02.483 17:36:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:02.483 17:36:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:02.483 [ 00:15:02.483 { 00:15:02.483 "name": "BaseBdev3", 00:15:02.483 "aliases": [ 00:15:02.483 "0a98f647-8ce5-44ab-8930-8208aa224dce" 00:15:02.483 ], 00:15:02.483 "product_name": "Malloc disk", 00:15:02.483 "block_size": 512, 00:15:02.483 "num_blocks": 65536, 00:15:02.483 "uuid": "0a98f647-8ce5-44ab-8930-8208aa224dce", 00:15:02.483 "assigned_rate_limits": { 00:15:02.483 "rw_ios_per_sec": 0, 00:15:02.483 "rw_mbytes_per_sec": 0, 00:15:02.483 "r_mbytes_per_sec": 0, 00:15:02.483 "w_mbytes_per_sec": 0 00:15:02.483 }, 00:15:02.483 "claimed": true, 00:15:02.483 "claim_type": "exclusive_write", 00:15:02.483 "zoned": false, 00:15:02.483 "supported_io_types": { 00:15:02.483 "read": true, 00:15:02.483 "write": true, 00:15:02.483 "unmap": true, 00:15:02.483 "flush": true, 00:15:02.483 "reset": true, 00:15:02.483 "nvme_admin": false, 00:15:02.483 "nvme_io": false, 00:15:02.483 "nvme_io_md": false, 00:15:02.483 "write_zeroes": true, 00:15:02.483 "zcopy": true, 00:15:02.483 "get_zone_info": false, 00:15:02.483 "zone_management": false, 00:15:02.483 "zone_append": false, 00:15:02.483 "compare": false, 00:15:02.483 "compare_and_write": false, 00:15:02.483 "abort": true, 00:15:02.483 "seek_hole": false, 00:15:02.484 "seek_data": false, 00:15:02.484 "copy": true, 00:15:02.484 "nvme_iov_md": false 00:15:02.484 }, 00:15:02.484 "memory_domains": [ 00:15:02.484 { 00:15:02.484 "dma_device_id": "system", 00:15:02.484 "dma_device_type": 1 00:15:02.484 }, 00:15:02.484 { 00:15:02.484 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:02.484 "dma_device_type": 2 00:15:02.484 } 00:15:02.484 ], 00:15:02.484 "driver_specific": {} 00:15:02.484 } 00:15:02.484 ] 00:15:02.484 17:36:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:02.484 17:36:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:15:02.484 17:36:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:15:02.484 17:36:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:15:02.484 17:36:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:15:02.484 17:36:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:02.484 17:36:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:02.484 17:36:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:02.484 17:36:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:02.484 17:36:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:02.484 17:36:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:02.484 17:36:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:02.484 17:36:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:02.484 17:36:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:02.484 17:36:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:02.484 17:36:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:02.484 17:36:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:02.484 17:36:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:02.484 17:36:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:02.484 17:36:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:02.484 "name": "Existed_Raid", 00:15:02.484 "uuid": "c438b21a-ec43-4443-9684-58450606cfc7", 00:15:02.484 "strip_size_kb": 64, 00:15:02.484 "state": "configuring", 00:15:02.484 "raid_level": "raid5f", 00:15:02.484 "superblock": true, 00:15:02.484 "num_base_bdevs": 4, 00:15:02.484 "num_base_bdevs_discovered": 3, 00:15:02.484 "num_base_bdevs_operational": 4, 00:15:02.484 "base_bdevs_list": [ 00:15:02.484 { 00:15:02.484 "name": "BaseBdev1", 00:15:02.484 "uuid": "47126875-78d1-4e7e-9b3e-80cec5bc212d", 00:15:02.484 "is_configured": true, 00:15:02.484 "data_offset": 2048, 00:15:02.484 "data_size": 63488 00:15:02.484 }, 00:15:02.484 { 00:15:02.484 "name": "BaseBdev2", 00:15:02.484 "uuid": "252dcc2a-faab-4713-9875-c2123fe04122", 00:15:02.484 "is_configured": true, 00:15:02.484 "data_offset": 2048, 00:15:02.484 "data_size": 63488 00:15:02.484 }, 00:15:02.484 { 00:15:02.484 "name": "BaseBdev3", 00:15:02.484 "uuid": "0a98f647-8ce5-44ab-8930-8208aa224dce", 00:15:02.484 "is_configured": true, 00:15:02.484 "data_offset": 2048, 00:15:02.484 "data_size": 63488 00:15:02.484 }, 00:15:02.484 { 00:15:02.484 "name": "BaseBdev4", 00:15:02.484 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:02.484 "is_configured": false, 00:15:02.484 "data_offset": 0, 00:15:02.484 "data_size": 0 00:15:02.484 } 00:15:02.484 ] 00:15:02.484 }' 00:15:02.484 17:36:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:02.484 17:36:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:02.744 17:36:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:15:02.744 17:36:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:02.744 17:36:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:02.744 [2024-11-27 17:36:33.887612] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:15:02.744 [2024-11-27 17:36:33.887901] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:15:02.744 [2024-11-27 17:36:33.887955] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:15:02.744 [2024-11-27 17:36:33.888234] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:15:02.744 BaseBdev4 00:15:02.744 [2024-11-27 17:36:33.888692] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:15:02.744 [2024-11-27 17:36:33.888787] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:15:02.744 [2024-11-27 17:36:33.888963] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:02.744 17:36:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:02.744 17:36:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev4 00:15:02.744 17:36:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev4 00:15:02.744 17:36:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:15:02.744 17:36:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:15:02.744 17:36:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:15:02.744 17:36:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:15:02.744 17:36:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:15:02.744 17:36:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:02.744 17:36:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:02.744 17:36:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:02.744 17:36:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:15:02.744 17:36:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:02.744 17:36:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:02.744 [ 00:15:02.744 { 00:15:02.744 "name": "BaseBdev4", 00:15:02.744 "aliases": [ 00:15:02.744 "256e5f62-8598-4d89-8ff4-529c197c5381" 00:15:02.744 ], 00:15:02.744 "product_name": "Malloc disk", 00:15:02.744 "block_size": 512, 00:15:02.744 "num_blocks": 65536, 00:15:02.744 "uuid": "256e5f62-8598-4d89-8ff4-529c197c5381", 00:15:02.744 "assigned_rate_limits": { 00:15:02.744 "rw_ios_per_sec": 0, 00:15:02.744 "rw_mbytes_per_sec": 0, 00:15:02.744 "r_mbytes_per_sec": 0, 00:15:02.744 "w_mbytes_per_sec": 0 00:15:02.744 }, 00:15:02.744 "claimed": true, 00:15:02.744 "claim_type": "exclusive_write", 00:15:02.744 "zoned": false, 00:15:02.744 "supported_io_types": { 00:15:02.744 "read": true, 00:15:02.745 "write": true, 00:15:02.745 "unmap": true, 00:15:02.745 "flush": true, 00:15:02.745 "reset": true, 00:15:02.745 "nvme_admin": false, 00:15:02.745 "nvme_io": false, 00:15:02.745 "nvme_io_md": false, 00:15:02.745 "write_zeroes": true, 00:15:02.745 "zcopy": true, 00:15:02.745 "get_zone_info": false, 00:15:02.745 "zone_management": false, 00:15:02.745 "zone_append": false, 00:15:02.745 "compare": false, 00:15:02.745 "compare_and_write": false, 00:15:02.745 "abort": true, 00:15:02.745 "seek_hole": false, 00:15:02.745 "seek_data": false, 00:15:02.745 "copy": true, 00:15:02.745 "nvme_iov_md": false 00:15:02.745 }, 00:15:02.745 "memory_domains": [ 00:15:02.745 { 00:15:02.745 "dma_device_id": "system", 00:15:02.745 "dma_device_type": 1 00:15:02.745 }, 00:15:02.745 { 00:15:02.745 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:02.745 "dma_device_type": 2 00:15:02.745 } 00:15:02.745 ], 00:15:02.745 "driver_specific": {} 00:15:02.745 } 00:15:02.745 ] 00:15:02.745 17:36:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:02.745 17:36:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:15:02.745 17:36:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:15:02.745 17:36:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:15:02.745 17:36:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 4 00:15:02.745 17:36:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:02.745 17:36:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:02.745 17:36:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:02.745 17:36:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:02.745 17:36:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:02.745 17:36:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:02.745 17:36:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:03.004 17:36:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:03.004 17:36:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:03.004 17:36:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:03.004 17:36:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:03.004 17:36:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:03.004 17:36:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:03.004 17:36:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:03.004 17:36:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:03.004 "name": "Existed_Raid", 00:15:03.004 "uuid": "c438b21a-ec43-4443-9684-58450606cfc7", 00:15:03.004 "strip_size_kb": 64, 00:15:03.004 "state": "online", 00:15:03.004 "raid_level": "raid5f", 00:15:03.004 "superblock": true, 00:15:03.004 "num_base_bdevs": 4, 00:15:03.004 "num_base_bdevs_discovered": 4, 00:15:03.004 "num_base_bdevs_operational": 4, 00:15:03.004 "base_bdevs_list": [ 00:15:03.004 { 00:15:03.004 "name": "BaseBdev1", 00:15:03.004 "uuid": "47126875-78d1-4e7e-9b3e-80cec5bc212d", 00:15:03.004 "is_configured": true, 00:15:03.004 "data_offset": 2048, 00:15:03.004 "data_size": 63488 00:15:03.004 }, 00:15:03.004 { 00:15:03.004 "name": "BaseBdev2", 00:15:03.004 "uuid": "252dcc2a-faab-4713-9875-c2123fe04122", 00:15:03.004 "is_configured": true, 00:15:03.004 "data_offset": 2048, 00:15:03.004 "data_size": 63488 00:15:03.004 }, 00:15:03.004 { 00:15:03.004 "name": "BaseBdev3", 00:15:03.004 "uuid": "0a98f647-8ce5-44ab-8930-8208aa224dce", 00:15:03.004 "is_configured": true, 00:15:03.004 "data_offset": 2048, 00:15:03.004 "data_size": 63488 00:15:03.004 }, 00:15:03.004 { 00:15:03.004 "name": "BaseBdev4", 00:15:03.004 "uuid": "256e5f62-8598-4d89-8ff4-529c197c5381", 00:15:03.004 "is_configured": true, 00:15:03.004 "data_offset": 2048, 00:15:03.004 "data_size": 63488 00:15:03.004 } 00:15:03.004 ] 00:15:03.004 }' 00:15:03.004 17:36:33 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:03.004 17:36:33 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:03.264 17:36:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:15:03.264 17:36:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:15:03.264 17:36:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:15:03.264 17:36:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:15:03.264 17:36:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:15:03.264 17:36:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:15:03.264 17:36:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:15:03.264 17:36:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:15:03.264 17:36:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:03.264 17:36:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:03.264 [2024-11-27 17:36:34.419056] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:03.264 17:36:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:03.525 17:36:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:15:03.525 "name": "Existed_Raid", 00:15:03.525 "aliases": [ 00:15:03.525 "c438b21a-ec43-4443-9684-58450606cfc7" 00:15:03.525 ], 00:15:03.525 "product_name": "Raid Volume", 00:15:03.525 "block_size": 512, 00:15:03.525 "num_blocks": 190464, 00:15:03.525 "uuid": "c438b21a-ec43-4443-9684-58450606cfc7", 00:15:03.525 "assigned_rate_limits": { 00:15:03.525 "rw_ios_per_sec": 0, 00:15:03.525 "rw_mbytes_per_sec": 0, 00:15:03.525 "r_mbytes_per_sec": 0, 00:15:03.525 "w_mbytes_per_sec": 0 00:15:03.525 }, 00:15:03.525 "claimed": false, 00:15:03.525 "zoned": false, 00:15:03.525 "supported_io_types": { 00:15:03.525 "read": true, 00:15:03.525 "write": true, 00:15:03.525 "unmap": false, 00:15:03.525 "flush": false, 00:15:03.525 "reset": true, 00:15:03.525 "nvme_admin": false, 00:15:03.525 "nvme_io": false, 00:15:03.525 "nvme_io_md": false, 00:15:03.525 "write_zeroes": true, 00:15:03.525 "zcopy": false, 00:15:03.525 "get_zone_info": false, 00:15:03.525 "zone_management": false, 00:15:03.525 "zone_append": false, 00:15:03.525 "compare": false, 00:15:03.525 "compare_and_write": false, 00:15:03.525 "abort": false, 00:15:03.525 "seek_hole": false, 00:15:03.525 "seek_data": false, 00:15:03.525 "copy": false, 00:15:03.525 "nvme_iov_md": false 00:15:03.525 }, 00:15:03.525 "driver_specific": { 00:15:03.525 "raid": { 00:15:03.525 "uuid": "c438b21a-ec43-4443-9684-58450606cfc7", 00:15:03.525 "strip_size_kb": 64, 00:15:03.525 "state": "online", 00:15:03.525 "raid_level": "raid5f", 00:15:03.525 "superblock": true, 00:15:03.525 "num_base_bdevs": 4, 00:15:03.525 "num_base_bdevs_discovered": 4, 00:15:03.525 "num_base_bdevs_operational": 4, 00:15:03.525 "base_bdevs_list": [ 00:15:03.525 { 00:15:03.525 "name": "BaseBdev1", 00:15:03.525 "uuid": "47126875-78d1-4e7e-9b3e-80cec5bc212d", 00:15:03.525 "is_configured": true, 00:15:03.525 "data_offset": 2048, 00:15:03.525 "data_size": 63488 00:15:03.525 }, 00:15:03.525 { 00:15:03.525 "name": "BaseBdev2", 00:15:03.525 "uuid": "252dcc2a-faab-4713-9875-c2123fe04122", 00:15:03.525 "is_configured": true, 00:15:03.525 "data_offset": 2048, 00:15:03.525 "data_size": 63488 00:15:03.525 }, 00:15:03.525 { 00:15:03.525 "name": "BaseBdev3", 00:15:03.525 "uuid": "0a98f647-8ce5-44ab-8930-8208aa224dce", 00:15:03.525 "is_configured": true, 00:15:03.525 "data_offset": 2048, 00:15:03.525 "data_size": 63488 00:15:03.525 }, 00:15:03.525 { 00:15:03.525 "name": "BaseBdev4", 00:15:03.525 "uuid": "256e5f62-8598-4d89-8ff4-529c197c5381", 00:15:03.525 "is_configured": true, 00:15:03.525 "data_offset": 2048, 00:15:03.525 "data_size": 63488 00:15:03.525 } 00:15:03.525 ] 00:15:03.525 } 00:15:03.525 } 00:15:03.525 }' 00:15:03.525 17:36:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:15:03.525 17:36:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:15:03.525 BaseBdev2 00:15:03.525 BaseBdev3 00:15:03.525 BaseBdev4' 00:15:03.525 17:36:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:03.525 17:36:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:15:03.525 17:36:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:03.525 17:36:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:15:03.525 17:36:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:03.525 17:36:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:03.525 17:36:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:03.525 17:36:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:03.525 17:36:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:15:03.525 17:36:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:15:03.525 17:36:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:03.525 17:36:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:15:03.525 17:36:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:03.525 17:36:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:03.525 17:36:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:03.525 17:36:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:03.525 17:36:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:15:03.525 17:36:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:15:03.525 17:36:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:03.525 17:36:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:15:03.525 17:36:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:03.525 17:36:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:03.525 17:36:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:03.525 17:36:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:03.525 17:36:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:15:03.525 17:36:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:15:03.525 17:36:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:03.525 17:36:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:15:03.525 17:36:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:03.525 17:36:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:03.525 17:36:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:03.525 17:36:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:03.786 17:36:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:15:03.786 17:36:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:15:03.786 17:36:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:15:03.786 17:36:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:03.786 17:36:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:03.786 [2024-11-27 17:36:34.750307] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:15:03.786 17:36:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:03.786 17:36:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@260 -- # local expected_state 00:15:03.786 17:36:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@261 -- # has_redundancy raid5f 00:15:03.786 17:36:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@198 -- # case $1 in 00:15:03.786 17:36:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@199 -- # return 0 00:15:03.786 17:36:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:15:03.786 17:36:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 3 00:15:03.786 17:36:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:03.786 17:36:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:03.786 17:36:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:03.786 17:36:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:03.786 17:36:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:15:03.786 17:36:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:03.786 17:36:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:03.786 17:36:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:03.786 17:36:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:03.786 17:36:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:03.786 17:36:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:03.786 17:36:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:03.786 17:36:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:03.786 17:36:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:03.786 17:36:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:03.786 "name": "Existed_Raid", 00:15:03.786 "uuid": "c438b21a-ec43-4443-9684-58450606cfc7", 00:15:03.786 "strip_size_kb": 64, 00:15:03.786 "state": "online", 00:15:03.786 "raid_level": "raid5f", 00:15:03.786 "superblock": true, 00:15:03.786 "num_base_bdevs": 4, 00:15:03.786 "num_base_bdevs_discovered": 3, 00:15:03.786 "num_base_bdevs_operational": 3, 00:15:03.786 "base_bdevs_list": [ 00:15:03.786 { 00:15:03.786 "name": null, 00:15:03.786 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:03.786 "is_configured": false, 00:15:03.786 "data_offset": 0, 00:15:03.786 "data_size": 63488 00:15:03.786 }, 00:15:03.786 { 00:15:03.786 "name": "BaseBdev2", 00:15:03.786 "uuid": "252dcc2a-faab-4713-9875-c2123fe04122", 00:15:03.786 "is_configured": true, 00:15:03.786 "data_offset": 2048, 00:15:03.786 "data_size": 63488 00:15:03.786 }, 00:15:03.786 { 00:15:03.786 "name": "BaseBdev3", 00:15:03.786 "uuid": "0a98f647-8ce5-44ab-8930-8208aa224dce", 00:15:03.786 "is_configured": true, 00:15:03.786 "data_offset": 2048, 00:15:03.786 "data_size": 63488 00:15:03.786 }, 00:15:03.786 { 00:15:03.786 "name": "BaseBdev4", 00:15:03.786 "uuid": "256e5f62-8598-4d89-8ff4-529c197c5381", 00:15:03.786 "is_configured": true, 00:15:03.786 "data_offset": 2048, 00:15:03.786 "data_size": 63488 00:15:03.786 } 00:15:03.786 ] 00:15:03.786 }' 00:15:03.786 17:36:34 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:03.786 17:36:34 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:04.046 17:36:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:15:04.046 17:36:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:15:04.046 17:36:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:04.046 17:36:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:04.046 17:36:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:15:04.046 17:36:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:04.046 17:36:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:04.306 17:36:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:15:04.306 17:36:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:15:04.306 17:36:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:15:04.306 17:36:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:04.306 17:36:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:04.306 [2024-11-27 17:36:35.252638] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:15:04.306 [2024-11-27 17:36:35.252785] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:15:04.306 [2024-11-27 17:36:35.263944] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:04.306 17:36:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:04.306 17:36:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:15:04.306 17:36:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:15:04.306 17:36:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:04.306 17:36:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:15:04.306 17:36:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:04.306 17:36:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:04.306 17:36:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:04.306 17:36:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:15:04.306 17:36:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:15:04.306 17:36:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev3 00:15:04.306 17:36:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:04.306 17:36:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:04.306 [2024-11-27 17:36:35.319840] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:15:04.307 17:36:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:04.307 17:36:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:15:04.307 17:36:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:15:04.307 17:36:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:04.307 17:36:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:04.307 17:36:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:15:04.307 17:36:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:04.307 17:36:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:04.307 17:36:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:15:04.307 17:36:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:15:04.307 17:36:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev4 00:15:04.307 17:36:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:04.307 17:36:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:04.307 [2024-11-27 17:36:35.386811] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev4 00:15:04.307 [2024-11-27 17:36:35.386851] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:15:04.307 17:36:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:04.307 17:36:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:15:04.307 17:36:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:15:04.307 17:36:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:15:04.307 17:36:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:04.307 17:36:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:04.307 17:36:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:04.307 17:36:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:04.307 17:36:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:15:04.307 17:36:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:15:04.307 17:36:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@284 -- # '[' 4 -gt 2 ']' 00:15:04.307 17:36:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i = 1 )) 00:15:04.307 17:36:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:15:04.307 17:36:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2 00:15:04.307 17:36:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:04.307 17:36:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:04.307 BaseBdev2 00:15:04.307 17:36:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:04.307 17:36:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev2 00:15:04.307 17:36:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:15:04.307 17:36:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:15:04.307 17:36:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:15:04.307 17:36:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:15:04.307 17:36:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:15:04.307 17:36:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:15:04.307 17:36:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:04.307 17:36:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:04.307 17:36:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:04.307 17:36:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:15:04.307 17:36:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:04.307 17:36:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:04.568 [ 00:15:04.568 { 00:15:04.568 "name": "BaseBdev2", 00:15:04.568 "aliases": [ 00:15:04.568 "a0d39b27-bd48-4dcb-bab5-d2d76df1dce1" 00:15:04.568 ], 00:15:04.568 "product_name": "Malloc disk", 00:15:04.568 "block_size": 512, 00:15:04.568 "num_blocks": 65536, 00:15:04.568 "uuid": "a0d39b27-bd48-4dcb-bab5-d2d76df1dce1", 00:15:04.568 "assigned_rate_limits": { 00:15:04.568 "rw_ios_per_sec": 0, 00:15:04.568 "rw_mbytes_per_sec": 0, 00:15:04.568 "r_mbytes_per_sec": 0, 00:15:04.568 "w_mbytes_per_sec": 0 00:15:04.568 }, 00:15:04.568 "claimed": false, 00:15:04.568 "zoned": false, 00:15:04.568 "supported_io_types": { 00:15:04.568 "read": true, 00:15:04.568 "write": true, 00:15:04.568 "unmap": true, 00:15:04.568 "flush": true, 00:15:04.568 "reset": true, 00:15:04.568 "nvme_admin": false, 00:15:04.568 "nvme_io": false, 00:15:04.568 "nvme_io_md": false, 00:15:04.568 "write_zeroes": true, 00:15:04.568 "zcopy": true, 00:15:04.568 "get_zone_info": false, 00:15:04.568 "zone_management": false, 00:15:04.568 "zone_append": false, 00:15:04.568 "compare": false, 00:15:04.568 "compare_and_write": false, 00:15:04.568 "abort": true, 00:15:04.568 "seek_hole": false, 00:15:04.568 "seek_data": false, 00:15:04.568 "copy": true, 00:15:04.568 "nvme_iov_md": false 00:15:04.568 }, 00:15:04.568 "memory_domains": [ 00:15:04.568 { 00:15:04.568 "dma_device_id": "system", 00:15:04.568 "dma_device_type": 1 00:15:04.568 }, 00:15:04.568 { 00:15:04.568 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:04.568 "dma_device_type": 2 00:15:04.568 } 00:15:04.568 ], 00:15:04.568 "driver_specific": {} 00:15:04.568 } 00:15:04.568 ] 00:15:04.568 17:36:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:04.568 17:36:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:15:04.568 17:36:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:15:04.568 17:36:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:15:04.568 17:36:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3 00:15:04.568 17:36:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:04.568 17:36:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:04.568 BaseBdev3 00:15:04.568 17:36:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:04.568 17:36:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev3 00:15:04.568 17:36:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev3 00:15:04.568 17:36:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:15:04.568 17:36:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:15:04.568 17:36:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:15:04.568 17:36:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:15:04.568 17:36:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:15:04.568 17:36:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:04.568 17:36:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:04.568 17:36:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:04.568 17:36:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 -t 2000 00:15:04.568 17:36:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:04.568 17:36:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:04.568 [ 00:15:04.568 { 00:15:04.568 "name": "BaseBdev3", 00:15:04.568 "aliases": [ 00:15:04.568 "25d165fd-1155-4bcf-8de5-86bbc8ff2b73" 00:15:04.568 ], 00:15:04.568 "product_name": "Malloc disk", 00:15:04.568 "block_size": 512, 00:15:04.568 "num_blocks": 65536, 00:15:04.568 "uuid": "25d165fd-1155-4bcf-8de5-86bbc8ff2b73", 00:15:04.568 "assigned_rate_limits": { 00:15:04.568 "rw_ios_per_sec": 0, 00:15:04.568 "rw_mbytes_per_sec": 0, 00:15:04.568 "r_mbytes_per_sec": 0, 00:15:04.568 "w_mbytes_per_sec": 0 00:15:04.568 }, 00:15:04.568 "claimed": false, 00:15:04.568 "zoned": false, 00:15:04.568 "supported_io_types": { 00:15:04.568 "read": true, 00:15:04.568 "write": true, 00:15:04.568 "unmap": true, 00:15:04.568 "flush": true, 00:15:04.568 "reset": true, 00:15:04.568 "nvme_admin": false, 00:15:04.568 "nvme_io": false, 00:15:04.568 "nvme_io_md": false, 00:15:04.568 "write_zeroes": true, 00:15:04.568 "zcopy": true, 00:15:04.568 "get_zone_info": false, 00:15:04.568 "zone_management": false, 00:15:04.568 "zone_append": false, 00:15:04.568 "compare": false, 00:15:04.568 "compare_and_write": false, 00:15:04.568 "abort": true, 00:15:04.568 "seek_hole": false, 00:15:04.568 "seek_data": false, 00:15:04.568 "copy": true, 00:15:04.568 "nvme_iov_md": false 00:15:04.568 }, 00:15:04.568 "memory_domains": [ 00:15:04.568 { 00:15:04.568 "dma_device_id": "system", 00:15:04.568 "dma_device_type": 1 00:15:04.568 }, 00:15:04.568 { 00:15:04.568 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:04.568 "dma_device_type": 2 00:15:04.568 } 00:15:04.568 ], 00:15:04.568 "driver_specific": {} 00:15:04.568 } 00:15:04.568 ] 00:15:04.568 17:36:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:04.568 17:36:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:15:04.568 17:36:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:15:04.568 17:36:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:15:04.568 17:36:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@287 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4 00:15:04.568 17:36:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:04.568 17:36:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:04.568 BaseBdev4 00:15:04.568 17:36:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:04.568 17:36:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@288 -- # waitforbdev BaseBdev4 00:15:04.568 17:36:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev4 00:15:04.568 17:36:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:15:04.568 17:36:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:15:04.568 17:36:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:15:04.568 17:36:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:15:04.568 17:36:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:15:04.568 17:36:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:04.568 17:36:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:04.569 17:36:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:04.569 17:36:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 -t 2000 00:15:04.569 17:36:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:04.569 17:36:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:04.569 [ 00:15:04.569 { 00:15:04.569 "name": "BaseBdev4", 00:15:04.569 "aliases": [ 00:15:04.569 "72652557-d354-46d7-859f-fe0c012292c0" 00:15:04.569 ], 00:15:04.569 "product_name": "Malloc disk", 00:15:04.569 "block_size": 512, 00:15:04.569 "num_blocks": 65536, 00:15:04.569 "uuid": "72652557-d354-46d7-859f-fe0c012292c0", 00:15:04.569 "assigned_rate_limits": { 00:15:04.569 "rw_ios_per_sec": 0, 00:15:04.569 "rw_mbytes_per_sec": 0, 00:15:04.569 "r_mbytes_per_sec": 0, 00:15:04.569 "w_mbytes_per_sec": 0 00:15:04.569 }, 00:15:04.569 "claimed": false, 00:15:04.569 "zoned": false, 00:15:04.569 "supported_io_types": { 00:15:04.569 "read": true, 00:15:04.569 "write": true, 00:15:04.569 "unmap": true, 00:15:04.569 "flush": true, 00:15:04.569 "reset": true, 00:15:04.569 "nvme_admin": false, 00:15:04.569 "nvme_io": false, 00:15:04.569 "nvme_io_md": false, 00:15:04.569 "write_zeroes": true, 00:15:04.569 "zcopy": true, 00:15:04.569 "get_zone_info": false, 00:15:04.569 "zone_management": false, 00:15:04.569 "zone_append": false, 00:15:04.569 "compare": false, 00:15:04.569 "compare_and_write": false, 00:15:04.569 "abort": true, 00:15:04.569 "seek_hole": false, 00:15:04.569 "seek_data": false, 00:15:04.569 "copy": true, 00:15:04.569 "nvme_iov_md": false 00:15:04.569 }, 00:15:04.569 "memory_domains": [ 00:15:04.569 { 00:15:04.569 "dma_device_id": "system", 00:15:04.569 "dma_device_type": 1 00:15:04.569 }, 00:15:04.569 { 00:15:04.569 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:04.569 "dma_device_type": 2 00:15:04.569 } 00:15:04.569 ], 00:15:04.569 "driver_specific": {} 00:15:04.569 } 00:15:04.569 ] 00:15:04.569 17:36:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:04.569 17:36:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:15:04.569 17:36:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i++ )) 00:15:04.569 17:36:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@286 -- # (( i < num_base_bdevs )) 00:15:04.569 17:36:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@290 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n Existed_Raid 00:15:04.569 17:36:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:04.569 17:36:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:04.569 [2024-11-27 17:36:35.620806] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:15:04.569 [2024-11-27 17:36:35.620929] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:15:04.569 [2024-11-27 17:36:35.620953] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:15:04.569 [2024-11-27 17:36:35.622699] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:15:04.569 [2024-11-27 17:36:35.622743] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:15:04.569 17:36:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:04.569 17:36:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@291 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:15:04.569 17:36:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:04.569 17:36:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:04.569 17:36:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:04.569 17:36:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:04.569 17:36:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:04.569 17:36:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:04.569 17:36:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:04.569 17:36:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:04.569 17:36:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:04.569 17:36:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:04.569 17:36:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:04.569 17:36:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:04.569 17:36:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:04.569 17:36:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:04.569 17:36:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:04.569 "name": "Existed_Raid", 00:15:04.569 "uuid": "2056e9bd-2d3c-4aed-af3a-ea529340a991", 00:15:04.569 "strip_size_kb": 64, 00:15:04.569 "state": "configuring", 00:15:04.569 "raid_level": "raid5f", 00:15:04.569 "superblock": true, 00:15:04.569 "num_base_bdevs": 4, 00:15:04.569 "num_base_bdevs_discovered": 3, 00:15:04.569 "num_base_bdevs_operational": 4, 00:15:04.569 "base_bdevs_list": [ 00:15:04.569 { 00:15:04.569 "name": "BaseBdev1", 00:15:04.569 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:04.569 "is_configured": false, 00:15:04.569 "data_offset": 0, 00:15:04.569 "data_size": 0 00:15:04.569 }, 00:15:04.569 { 00:15:04.569 "name": "BaseBdev2", 00:15:04.569 "uuid": "a0d39b27-bd48-4dcb-bab5-d2d76df1dce1", 00:15:04.569 "is_configured": true, 00:15:04.569 "data_offset": 2048, 00:15:04.569 "data_size": 63488 00:15:04.569 }, 00:15:04.569 { 00:15:04.569 "name": "BaseBdev3", 00:15:04.569 "uuid": "25d165fd-1155-4bcf-8de5-86bbc8ff2b73", 00:15:04.569 "is_configured": true, 00:15:04.569 "data_offset": 2048, 00:15:04.569 "data_size": 63488 00:15:04.569 }, 00:15:04.569 { 00:15:04.569 "name": "BaseBdev4", 00:15:04.569 "uuid": "72652557-d354-46d7-859f-fe0c012292c0", 00:15:04.569 "is_configured": true, 00:15:04.569 "data_offset": 2048, 00:15:04.569 "data_size": 63488 00:15:04.569 } 00:15:04.569 ] 00:15:04.569 }' 00:15:04.569 17:36:35 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:04.569 17:36:35 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:05.140 17:36:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@293 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev2 00:15:05.140 17:36:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:05.140 17:36:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:05.140 [2024-11-27 17:36:36.092022] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:15:05.140 17:36:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:05.140 17:36:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@294 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:15:05.140 17:36:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:05.140 17:36:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:05.140 17:36:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:05.140 17:36:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:05.140 17:36:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:05.140 17:36:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:05.140 17:36:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:05.140 17:36:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:05.140 17:36:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:05.140 17:36:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:05.140 17:36:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:05.140 17:36:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:05.140 17:36:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:05.140 17:36:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:05.140 17:36:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:05.140 "name": "Existed_Raid", 00:15:05.140 "uuid": "2056e9bd-2d3c-4aed-af3a-ea529340a991", 00:15:05.140 "strip_size_kb": 64, 00:15:05.140 "state": "configuring", 00:15:05.140 "raid_level": "raid5f", 00:15:05.140 "superblock": true, 00:15:05.140 "num_base_bdevs": 4, 00:15:05.140 "num_base_bdevs_discovered": 2, 00:15:05.140 "num_base_bdevs_operational": 4, 00:15:05.140 "base_bdevs_list": [ 00:15:05.140 { 00:15:05.140 "name": "BaseBdev1", 00:15:05.140 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:05.140 "is_configured": false, 00:15:05.140 "data_offset": 0, 00:15:05.140 "data_size": 0 00:15:05.140 }, 00:15:05.140 { 00:15:05.140 "name": null, 00:15:05.140 "uuid": "a0d39b27-bd48-4dcb-bab5-d2d76df1dce1", 00:15:05.140 "is_configured": false, 00:15:05.140 "data_offset": 0, 00:15:05.140 "data_size": 63488 00:15:05.140 }, 00:15:05.140 { 00:15:05.140 "name": "BaseBdev3", 00:15:05.140 "uuid": "25d165fd-1155-4bcf-8de5-86bbc8ff2b73", 00:15:05.140 "is_configured": true, 00:15:05.140 "data_offset": 2048, 00:15:05.140 "data_size": 63488 00:15:05.140 }, 00:15:05.140 { 00:15:05.140 "name": "BaseBdev4", 00:15:05.140 "uuid": "72652557-d354-46d7-859f-fe0c012292c0", 00:15:05.140 "is_configured": true, 00:15:05.140 "data_offset": 2048, 00:15:05.140 "data_size": 63488 00:15:05.140 } 00:15:05.140 ] 00:15:05.140 }' 00:15:05.140 17:36:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:05.140 17:36:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:05.400 17:36:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:15:05.400 17:36:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:05.400 17:36:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:05.400 17:36:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:05.400 17:36:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:05.400 17:36:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@295 -- # [[ false == \f\a\l\s\e ]] 00:15:05.400 17:36:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@297 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1 00:15:05.400 17:36:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:05.400 17:36:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:05.660 BaseBdev1 00:15:05.660 [2024-11-27 17:36:36.594228] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:15:05.660 17:36:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:05.660 17:36:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@298 -- # waitforbdev BaseBdev1 00:15:05.660 17:36:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:15:05.660 17:36:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:15:05.660 17:36:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:15:05.660 17:36:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:15:05.660 17:36:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:15:05.660 17:36:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:15:05.660 17:36:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:05.660 17:36:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:05.660 17:36:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:05.660 17:36:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:15:05.660 17:36:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:05.660 17:36:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:05.660 [ 00:15:05.660 { 00:15:05.660 "name": "BaseBdev1", 00:15:05.660 "aliases": [ 00:15:05.660 "5de85ecc-f29d-4986-b7c0-eb7b587f60bb" 00:15:05.660 ], 00:15:05.660 "product_name": "Malloc disk", 00:15:05.660 "block_size": 512, 00:15:05.660 "num_blocks": 65536, 00:15:05.660 "uuid": "5de85ecc-f29d-4986-b7c0-eb7b587f60bb", 00:15:05.660 "assigned_rate_limits": { 00:15:05.660 "rw_ios_per_sec": 0, 00:15:05.660 "rw_mbytes_per_sec": 0, 00:15:05.660 "r_mbytes_per_sec": 0, 00:15:05.660 "w_mbytes_per_sec": 0 00:15:05.660 }, 00:15:05.660 "claimed": true, 00:15:05.660 "claim_type": "exclusive_write", 00:15:05.660 "zoned": false, 00:15:05.660 "supported_io_types": { 00:15:05.660 "read": true, 00:15:05.660 "write": true, 00:15:05.660 "unmap": true, 00:15:05.660 "flush": true, 00:15:05.660 "reset": true, 00:15:05.660 "nvme_admin": false, 00:15:05.660 "nvme_io": false, 00:15:05.660 "nvme_io_md": false, 00:15:05.660 "write_zeroes": true, 00:15:05.660 "zcopy": true, 00:15:05.660 "get_zone_info": false, 00:15:05.660 "zone_management": false, 00:15:05.660 "zone_append": false, 00:15:05.660 "compare": false, 00:15:05.660 "compare_and_write": false, 00:15:05.660 "abort": true, 00:15:05.660 "seek_hole": false, 00:15:05.660 "seek_data": false, 00:15:05.660 "copy": true, 00:15:05.660 "nvme_iov_md": false 00:15:05.660 }, 00:15:05.660 "memory_domains": [ 00:15:05.660 { 00:15:05.660 "dma_device_id": "system", 00:15:05.660 "dma_device_type": 1 00:15:05.660 }, 00:15:05.660 { 00:15:05.660 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:05.660 "dma_device_type": 2 00:15:05.660 } 00:15:05.660 ], 00:15:05.660 "driver_specific": {} 00:15:05.660 } 00:15:05.660 ] 00:15:05.660 17:36:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:05.660 17:36:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:15:05.660 17:36:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@299 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:15:05.660 17:36:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:05.660 17:36:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:05.660 17:36:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:05.660 17:36:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:05.661 17:36:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:05.661 17:36:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:05.661 17:36:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:05.661 17:36:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:05.661 17:36:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:05.661 17:36:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:05.661 17:36:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:05.661 17:36:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:05.661 17:36:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:05.661 17:36:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:05.661 17:36:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:05.661 "name": "Existed_Raid", 00:15:05.661 "uuid": "2056e9bd-2d3c-4aed-af3a-ea529340a991", 00:15:05.661 "strip_size_kb": 64, 00:15:05.661 "state": "configuring", 00:15:05.661 "raid_level": "raid5f", 00:15:05.661 "superblock": true, 00:15:05.661 "num_base_bdevs": 4, 00:15:05.661 "num_base_bdevs_discovered": 3, 00:15:05.661 "num_base_bdevs_operational": 4, 00:15:05.661 "base_bdevs_list": [ 00:15:05.661 { 00:15:05.661 "name": "BaseBdev1", 00:15:05.661 "uuid": "5de85ecc-f29d-4986-b7c0-eb7b587f60bb", 00:15:05.661 "is_configured": true, 00:15:05.661 "data_offset": 2048, 00:15:05.661 "data_size": 63488 00:15:05.661 }, 00:15:05.661 { 00:15:05.661 "name": null, 00:15:05.661 "uuid": "a0d39b27-bd48-4dcb-bab5-d2d76df1dce1", 00:15:05.661 "is_configured": false, 00:15:05.661 "data_offset": 0, 00:15:05.661 "data_size": 63488 00:15:05.661 }, 00:15:05.661 { 00:15:05.661 "name": "BaseBdev3", 00:15:05.661 "uuid": "25d165fd-1155-4bcf-8de5-86bbc8ff2b73", 00:15:05.661 "is_configured": true, 00:15:05.661 "data_offset": 2048, 00:15:05.661 "data_size": 63488 00:15:05.661 }, 00:15:05.661 { 00:15:05.661 "name": "BaseBdev4", 00:15:05.661 "uuid": "72652557-d354-46d7-859f-fe0c012292c0", 00:15:05.661 "is_configured": true, 00:15:05.661 "data_offset": 2048, 00:15:05.661 "data_size": 63488 00:15:05.661 } 00:15:05.661 ] 00:15:05.661 }' 00:15:05.661 17:36:36 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:05.661 17:36:36 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:05.920 17:36:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:05.920 17:36:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:05.920 17:36:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:05.920 17:36:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:15:05.920 17:36:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:05.920 17:36:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@300 -- # [[ true == \t\r\u\e ]] 00:15:05.920 17:36:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@302 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev3 00:15:05.920 17:36:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:05.920 17:36:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:05.920 [2024-11-27 17:36:37.101407] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev3 00:15:05.920 17:36:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:05.920 17:36:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@303 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:15:05.920 17:36:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:05.920 17:36:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:05.920 17:36:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:05.920 17:36:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:05.920 17:36:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:05.920 17:36:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:05.920 17:36:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:05.920 17:36:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:05.920 17:36:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:06.179 17:36:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:06.179 17:36:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:06.179 17:36:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:06.179 17:36:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:06.179 17:36:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:06.179 17:36:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:06.179 "name": "Existed_Raid", 00:15:06.179 "uuid": "2056e9bd-2d3c-4aed-af3a-ea529340a991", 00:15:06.179 "strip_size_kb": 64, 00:15:06.179 "state": "configuring", 00:15:06.179 "raid_level": "raid5f", 00:15:06.179 "superblock": true, 00:15:06.179 "num_base_bdevs": 4, 00:15:06.179 "num_base_bdevs_discovered": 2, 00:15:06.179 "num_base_bdevs_operational": 4, 00:15:06.179 "base_bdevs_list": [ 00:15:06.179 { 00:15:06.179 "name": "BaseBdev1", 00:15:06.179 "uuid": "5de85ecc-f29d-4986-b7c0-eb7b587f60bb", 00:15:06.179 "is_configured": true, 00:15:06.179 "data_offset": 2048, 00:15:06.179 "data_size": 63488 00:15:06.179 }, 00:15:06.179 { 00:15:06.179 "name": null, 00:15:06.179 "uuid": "a0d39b27-bd48-4dcb-bab5-d2d76df1dce1", 00:15:06.179 "is_configured": false, 00:15:06.179 "data_offset": 0, 00:15:06.179 "data_size": 63488 00:15:06.179 }, 00:15:06.179 { 00:15:06.179 "name": null, 00:15:06.179 "uuid": "25d165fd-1155-4bcf-8de5-86bbc8ff2b73", 00:15:06.179 "is_configured": false, 00:15:06.179 "data_offset": 0, 00:15:06.179 "data_size": 63488 00:15:06.179 }, 00:15:06.179 { 00:15:06.179 "name": "BaseBdev4", 00:15:06.179 "uuid": "72652557-d354-46d7-859f-fe0c012292c0", 00:15:06.179 "is_configured": true, 00:15:06.179 "data_offset": 2048, 00:15:06.179 "data_size": 63488 00:15:06.179 } 00:15:06.179 ] 00:15:06.179 }' 00:15:06.179 17:36:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:06.179 17:36:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:06.440 17:36:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:06.440 17:36:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:06.440 17:36:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:06.440 17:36:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:15:06.440 17:36:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:06.440 17:36:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@304 -- # [[ false == \f\a\l\s\e ]] 00:15:06.440 17:36:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@306 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev3 00:15:06.440 17:36:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:06.440 17:36:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:06.440 [2024-11-27 17:36:37.548692] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:15:06.440 17:36:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:06.440 17:36:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@307 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:15:06.440 17:36:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:06.440 17:36:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:06.440 17:36:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:06.440 17:36:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:06.440 17:36:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:06.440 17:36:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:06.440 17:36:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:06.440 17:36:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:06.440 17:36:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:06.440 17:36:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:06.440 17:36:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:06.440 17:36:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:06.440 17:36:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:06.440 17:36:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:06.440 17:36:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:06.440 "name": "Existed_Raid", 00:15:06.440 "uuid": "2056e9bd-2d3c-4aed-af3a-ea529340a991", 00:15:06.440 "strip_size_kb": 64, 00:15:06.440 "state": "configuring", 00:15:06.440 "raid_level": "raid5f", 00:15:06.440 "superblock": true, 00:15:06.440 "num_base_bdevs": 4, 00:15:06.440 "num_base_bdevs_discovered": 3, 00:15:06.440 "num_base_bdevs_operational": 4, 00:15:06.440 "base_bdevs_list": [ 00:15:06.440 { 00:15:06.440 "name": "BaseBdev1", 00:15:06.440 "uuid": "5de85ecc-f29d-4986-b7c0-eb7b587f60bb", 00:15:06.440 "is_configured": true, 00:15:06.440 "data_offset": 2048, 00:15:06.440 "data_size": 63488 00:15:06.440 }, 00:15:06.440 { 00:15:06.440 "name": null, 00:15:06.440 "uuid": "a0d39b27-bd48-4dcb-bab5-d2d76df1dce1", 00:15:06.440 "is_configured": false, 00:15:06.440 "data_offset": 0, 00:15:06.440 "data_size": 63488 00:15:06.440 }, 00:15:06.440 { 00:15:06.440 "name": "BaseBdev3", 00:15:06.440 "uuid": "25d165fd-1155-4bcf-8de5-86bbc8ff2b73", 00:15:06.440 "is_configured": true, 00:15:06.440 "data_offset": 2048, 00:15:06.440 "data_size": 63488 00:15:06.440 }, 00:15:06.440 { 00:15:06.440 "name": "BaseBdev4", 00:15:06.440 "uuid": "72652557-d354-46d7-859f-fe0c012292c0", 00:15:06.440 "is_configured": true, 00:15:06.440 "data_offset": 2048, 00:15:06.440 "data_size": 63488 00:15:06.440 } 00:15:06.440 ] 00:15:06.440 }' 00:15:06.440 17:36:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:06.440 17:36:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:07.011 17:36:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:07.011 17:36:37 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # jq '.[0].base_bdevs_list[2].is_configured' 00:15:07.011 17:36:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:07.011 17:36:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:07.011 17:36:37 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:07.011 17:36:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@308 -- # [[ true == \t\r\u\e ]] 00:15:07.011 17:36:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@310 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:15:07.011 17:36:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:07.011 17:36:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:07.011 [2024-11-27 17:36:38.019858] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:15:07.011 17:36:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:07.011 17:36:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@311 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:15:07.011 17:36:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:07.011 17:36:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:07.011 17:36:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:07.011 17:36:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:07.011 17:36:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:07.011 17:36:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:07.011 17:36:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:07.011 17:36:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:07.011 17:36:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:07.011 17:36:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:07.011 17:36:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:07.011 17:36:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:07.011 17:36:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:07.011 17:36:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:07.011 17:36:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:07.011 "name": "Existed_Raid", 00:15:07.011 "uuid": "2056e9bd-2d3c-4aed-af3a-ea529340a991", 00:15:07.011 "strip_size_kb": 64, 00:15:07.011 "state": "configuring", 00:15:07.011 "raid_level": "raid5f", 00:15:07.011 "superblock": true, 00:15:07.011 "num_base_bdevs": 4, 00:15:07.011 "num_base_bdevs_discovered": 2, 00:15:07.011 "num_base_bdevs_operational": 4, 00:15:07.011 "base_bdevs_list": [ 00:15:07.011 { 00:15:07.011 "name": null, 00:15:07.011 "uuid": "5de85ecc-f29d-4986-b7c0-eb7b587f60bb", 00:15:07.011 "is_configured": false, 00:15:07.011 "data_offset": 0, 00:15:07.011 "data_size": 63488 00:15:07.011 }, 00:15:07.011 { 00:15:07.011 "name": null, 00:15:07.011 "uuid": "a0d39b27-bd48-4dcb-bab5-d2d76df1dce1", 00:15:07.011 "is_configured": false, 00:15:07.011 "data_offset": 0, 00:15:07.011 "data_size": 63488 00:15:07.011 }, 00:15:07.011 { 00:15:07.011 "name": "BaseBdev3", 00:15:07.011 "uuid": "25d165fd-1155-4bcf-8de5-86bbc8ff2b73", 00:15:07.011 "is_configured": true, 00:15:07.011 "data_offset": 2048, 00:15:07.011 "data_size": 63488 00:15:07.011 }, 00:15:07.011 { 00:15:07.011 "name": "BaseBdev4", 00:15:07.011 "uuid": "72652557-d354-46d7-859f-fe0c012292c0", 00:15:07.011 "is_configured": true, 00:15:07.011 "data_offset": 2048, 00:15:07.011 "data_size": 63488 00:15:07.011 } 00:15:07.011 ] 00:15:07.011 }' 00:15:07.011 17:36:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:07.011 17:36:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:07.272 17:36:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:07.272 17:36:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # jq '.[0].base_bdevs_list[0].is_configured' 00:15:07.272 17:36:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:07.272 17:36:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:07.272 17:36:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:07.532 17:36:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@312 -- # [[ false == \f\a\l\s\e ]] 00:15:07.532 17:36:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@314 -- # rpc_cmd bdev_raid_add_base_bdev Existed_Raid BaseBdev2 00:15:07.532 17:36:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:07.532 17:36:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:07.532 [2024-11-27 17:36:38.485373] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:15:07.532 17:36:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:07.532 17:36:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@315 -- # verify_raid_bdev_state Existed_Raid configuring raid5f 64 4 00:15:07.532 17:36:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:07.532 17:36:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:07.532 17:36:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:07.532 17:36:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:07.532 17:36:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:07.532 17:36:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:07.532 17:36:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:07.532 17:36:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:07.532 17:36:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:07.532 17:36:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:07.532 17:36:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:07.532 17:36:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:07.532 17:36:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:07.532 17:36:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:07.532 17:36:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:07.532 "name": "Existed_Raid", 00:15:07.532 "uuid": "2056e9bd-2d3c-4aed-af3a-ea529340a991", 00:15:07.532 "strip_size_kb": 64, 00:15:07.532 "state": "configuring", 00:15:07.532 "raid_level": "raid5f", 00:15:07.532 "superblock": true, 00:15:07.532 "num_base_bdevs": 4, 00:15:07.532 "num_base_bdevs_discovered": 3, 00:15:07.532 "num_base_bdevs_operational": 4, 00:15:07.532 "base_bdevs_list": [ 00:15:07.532 { 00:15:07.532 "name": null, 00:15:07.532 "uuid": "5de85ecc-f29d-4986-b7c0-eb7b587f60bb", 00:15:07.532 "is_configured": false, 00:15:07.532 "data_offset": 0, 00:15:07.532 "data_size": 63488 00:15:07.532 }, 00:15:07.532 { 00:15:07.532 "name": "BaseBdev2", 00:15:07.532 "uuid": "a0d39b27-bd48-4dcb-bab5-d2d76df1dce1", 00:15:07.532 "is_configured": true, 00:15:07.532 "data_offset": 2048, 00:15:07.532 "data_size": 63488 00:15:07.532 }, 00:15:07.532 { 00:15:07.532 "name": "BaseBdev3", 00:15:07.532 "uuid": "25d165fd-1155-4bcf-8de5-86bbc8ff2b73", 00:15:07.532 "is_configured": true, 00:15:07.532 "data_offset": 2048, 00:15:07.532 "data_size": 63488 00:15:07.532 }, 00:15:07.532 { 00:15:07.532 "name": "BaseBdev4", 00:15:07.532 "uuid": "72652557-d354-46d7-859f-fe0c012292c0", 00:15:07.532 "is_configured": true, 00:15:07.532 "data_offset": 2048, 00:15:07.532 "data_size": 63488 00:15:07.532 } 00:15:07.532 ] 00:15:07.532 }' 00:15:07.532 17:36:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:07.532 17:36:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:07.792 17:36:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:07.792 17:36:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:07.792 17:36:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:07.792 17:36:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # jq '.[0].base_bdevs_list[1].is_configured' 00:15:07.792 17:36:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:08.054 17:36:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@316 -- # [[ true == \t\r\u\e ]] 00:15:08.054 17:36:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:08.054 17:36:38 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # jq -r '.[0].base_bdevs_list[0].uuid' 00:15:08.054 17:36:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:08.054 17:36:38 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:08.054 17:36:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:08.054 17:36:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@318 -- # rpc_cmd bdev_malloc_create 32 512 -b NewBaseBdev -u 5de85ecc-f29d-4986-b7c0-eb7b587f60bb 00:15:08.054 17:36:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:08.054 17:36:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:08.054 [2024-11-27 17:36:39.043297] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev NewBaseBdev is claimed 00:15:08.054 [2024-11-27 17:36:39.043555] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:15:08.054 NewBaseBdev 00:15:08.054 [2024-11-27 17:36:39.043601] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:15:08.054 [2024-11-27 17:36:39.043878] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002a10 00:15:08.054 [2024-11-27 17:36:39.044325] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:15:08.054 [2024-11-27 17:36:39.044340] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001c80 00:15:08.054 [2024-11-27 17:36:39.044433] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:08.054 17:36:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:08.054 17:36:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@319 -- # waitforbdev NewBaseBdev 00:15:08.054 17:36:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@899 -- # local bdev_name=NewBaseBdev 00:15:08.054 17:36:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:15:08.054 17:36:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@901 -- # local i 00:15:08.054 17:36:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:15:08.054 17:36:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:15:08.054 17:36:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:15:08.054 17:36:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:08.054 17:36:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:08.054 17:36:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:08.054 17:36:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev -t 2000 00:15:08.054 17:36:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:08.054 17:36:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:08.054 [ 00:15:08.054 { 00:15:08.054 "name": "NewBaseBdev", 00:15:08.054 "aliases": [ 00:15:08.054 "5de85ecc-f29d-4986-b7c0-eb7b587f60bb" 00:15:08.054 ], 00:15:08.054 "product_name": "Malloc disk", 00:15:08.054 "block_size": 512, 00:15:08.054 "num_blocks": 65536, 00:15:08.054 "uuid": "5de85ecc-f29d-4986-b7c0-eb7b587f60bb", 00:15:08.054 "assigned_rate_limits": { 00:15:08.054 "rw_ios_per_sec": 0, 00:15:08.054 "rw_mbytes_per_sec": 0, 00:15:08.054 "r_mbytes_per_sec": 0, 00:15:08.054 "w_mbytes_per_sec": 0 00:15:08.054 }, 00:15:08.054 "claimed": true, 00:15:08.054 "claim_type": "exclusive_write", 00:15:08.054 "zoned": false, 00:15:08.054 "supported_io_types": { 00:15:08.054 "read": true, 00:15:08.054 "write": true, 00:15:08.054 "unmap": true, 00:15:08.054 "flush": true, 00:15:08.054 "reset": true, 00:15:08.054 "nvme_admin": false, 00:15:08.054 "nvme_io": false, 00:15:08.054 "nvme_io_md": false, 00:15:08.054 "write_zeroes": true, 00:15:08.054 "zcopy": true, 00:15:08.054 "get_zone_info": false, 00:15:08.054 "zone_management": false, 00:15:08.054 "zone_append": false, 00:15:08.054 "compare": false, 00:15:08.054 "compare_and_write": false, 00:15:08.054 "abort": true, 00:15:08.054 "seek_hole": false, 00:15:08.054 "seek_data": false, 00:15:08.054 "copy": true, 00:15:08.054 "nvme_iov_md": false 00:15:08.054 }, 00:15:08.054 "memory_domains": [ 00:15:08.054 { 00:15:08.054 "dma_device_id": "system", 00:15:08.054 "dma_device_type": 1 00:15:08.054 }, 00:15:08.054 { 00:15:08.054 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:15:08.054 "dma_device_type": 2 00:15:08.054 } 00:15:08.054 ], 00:15:08.054 "driver_specific": {} 00:15:08.054 } 00:15:08.054 ] 00:15:08.054 17:36:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:08.054 17:36:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@907 -- # return 0 00:15:08.054 17:36:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@320 -- # verify_raid_bdev_state Existed_Raid online raid5f 64 4 00:15:08.054 17:36:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:15:08.054 17:36:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:08.054 17:36:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:08.054 17:36:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:08.054 17:36:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:08.054 17:36:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:08.054 17:36:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:08.054 17:36:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:08.054 17:36:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:08.054 17:36:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:08.054 17:36:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:15:08.055 17:36:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:08.055 17:36:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:08.055 17:36:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:08.055 17:36:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:08.055 "name": "Existed_Raid", 00:15:08.055 "uuid": "2056e9bd-2d3c-4aed-af3a-ea529340a991", 00:15:08.055 "strip_size_kb": 64, 00:15:08.055 "state": "online", 00:15:08.055 "raid_level": "raid5f", 00:15:08.055 "superblock": true, 00:15:08.055 "num_base_bdevs": 4, 00:15:08.055 "num_base_bdevs_discovered": 4, 00:15:08.055 "num_base_bdevs_operational": 4, 00:15:08.055 "base_bdevs_list": [ 00:15:08.055 { 00:15:08.055 "name": "NewBaseBdev", 00:15:08.055 "uuid": "5de85ecc-f29d-4986-b7c0-eb7b587f60bb", 00:15:08.055 "is_configured": true, 00:15:08.055 "data_offset": 2048, 00:15:08.055 "data_size": 63488 00:15:08.055 }, 00:15:08.055 { 00:15:08.055 "name": "BaseBdev2", 00:15:08.055 "uuid": "a0d39b27-bd48-4dcb-bab5-d2d76df1dce1", 00:15:08.055 "is_configured": true, 00:15:08.055 "data_offset": 2048, 00:15:08.055 "data_size": 63488 00:15:08.055 }, 00:15:08.055 { 00:15:08.055 "name": "BaseBdev3", 00:15:08.055 "uuid": "25d165fd-1155-4bcf-8de5-86bbc8ff2b73", 00:15:08.055 "is_configured": true, 00:15:08.055 "data_offset": 2048, 00:15:08.055 "data_size": 63488 00:15:08.055 }, 00:15:08.055 { 00:15:08.055 "name": "BaseBdev4", 00:15:08.055 "uuid": "72652557-d354-46d7-859f-fe0c012292c0", 00:15:08.055 "is_configured": true, 00:15:08.055 "data_offset": 2048, 00:15:08.055 "data_size": 63488 00:15:08.055 } 00:15:08.055 ] 00:15:08.055 }' 00:15:08.055 17:36:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:08.055 17:36:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:08.624 17:36:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@321 -- # verify_raid_bdev_properties Existed_Raid 00:15:08.624 17:36:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:15:08.624 17:36:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:15:08.624 17:36:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:15:08.624 17:36:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@184 -- # local name 00:15:08.624 17:36:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:15:08.624 17:36:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:15:08.624 17:36:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:15:08.624 17:36:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:08.624 17:36:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:08.624 [2024-11-27 17:36:39.522856] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:08.624 17:36:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:08.624 17:36:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:15:08.624 "name": "Existed_Raid", 00:15:08.624 "aliases": [ 00:15:08.624 "2056e9bd-2d3c-4aed-af3a-ea529340a991" 00:15:08.624 ], 00:15:08.624 "product_name": "Raid Volume", 00:15:08.624 "block_size": 512, 00:15:08.624 "num_blocks": 190464, 00:15:08.624 "uuid": "2056e9bd-2d3c-4aed-af3a-ea529340a991", 00:15:08.624 "assigned_rate_limits": { 00:15:08.624 "rw_ios_per_sec": 0, 00:15:08.624 "rw_mbytes_per_sec": 0, 00:15:08.624 "r_mbytes_per_sec": 0, 00:15:08.624 "w_mbytes_per_sec": 0 00:15:08.624 }, 00:15:08.624 "claimed": false, 00:15:08.624 "zoned": false, 00:15:08.624 "supported_io_types": { 00:15:08.624 "read": true, 00:15:08.624 "write": true, 00:15:08.624 "unmap": false, 00:15:08.624 "flush": false, 00:15:08.624 "reset": true, 00:15:08.624 "nvme_admin": false, 00:15:08.624 "nvme_io": false, 00:15:08.624 "nvme_io_md": false, 00:15:08.624 "write_zeroes": true, 00:15:08.624 "zcopy": false, 00:15:08.624 "get_zone_info": false, 00:15:08.624 "zone_management": false, 00:15:08.624 "zone_append": false, 00:15:08.624 "compare": false, 00:15:08.624 "compare_and_write": false, 00:15:08.624 "abort": false, 00:15:08.624 "seek_hole": false, 00:15:08.624 "seek_data": false, 00:15:08.624 "copy": false, 00:15:08.624 "nvme_iov_md": false 00:15:08.625 }, 00:15:08.625 "driver_specific": { 00:15:08.625 "raid": { 00:15:08.625 "uuid": "2056e9bd-2d3c-4aed-af3a-ea529340a991", 00:15:08.625 "strip_size_kb": 64, 00:15:08.625 "state": "online", 00:15:08.625 "raid_level": "raid5f", 00:15:08.625 "superblock": true, 00:15:08.625 "num_base_bdevs": 4, 00:15:08.625 "num_base_bdevs_discovered": 4, 00:15:08.625 "num_base_bdevs_operational": 4, 00:15:08.625 "base_bdevs_list": [ 00:15:08.625 { 00:15:08.625 "name": "NewBaseBdev", 00:15:08.625 "uuid": "5de85ecc-f29d-4986-b7c0-eb7b587f60bb", 00:15:08.625 "is_configured": true, 00:15:08.625 "data_offset": 2048, 00:15:08.625 "data_size": 63488 00:15:08.625 }, 00:15:08.625 { 00:15:08.625 "name": "BaseBdev2", 00:15:08.625 "uuid": "a0d39b27-bd48-4dcb-bab5-d2d76df1dce1", 00:15:08.625 "is_configured": true, 00:15:08.625 "data_offset": 2048, 00:15:08.625 "data_size": 63488 00:15:08.625 }, 00:15:08.625 { 00:15:08.625 "name": "BaseBdev3", 00:15:08.625 "uuid": "25d165fd-1155-4bcf-8de5-86bbc8ff2b73", 00:15:08.625 "is_configured": true, 00:15:08.625 "data_offset": 2048, 00:15:08.625 "data_size": 63488 00:15:08.625 }, 00:15:08.625 { 00:15:08.625 "name": "BaseBdev4", 00:15:08.625 "uuid": "72652557-d354-46d7-859f-fe0c012292c0", 00:15:08.625 "is_configured": true, 00:15:08.625 "data_offset": 2048, 00:15:08.625 "data_size": 63488 00:15:08.625 } 00:15:08.625 ] 00:15:08.625 } 00:15:08.625 } 00:15:08.625 }' 00:15:08.625 17:36:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:15:08.625 17:36:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@188 -- # base_bdev_names='NewBaseBdev 00:15:08.625 BaseBdev2 00:15:08.625 BaseBdev3 00:15:08.625 BaseBdev4' 00:15:08.625 17:36:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:08.625 17:36:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:15:08.625 17:36:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:08.625 17:36:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b NewBaseBdev 00:15:08.625 17:36:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:08.625 17:36:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:08.625 17:36:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:08.625 17:36:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:08.625 17:36:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:15:08.625 17:36:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:15:08.625 17:36:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:08.625 17:36:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:15:08.625 17:36:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:08.625 17:36:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:08.625 17:36:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:08.625 17:36:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:08.625 17:36:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:15:08.625 17:36:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:15:08.625 17:36:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:08.625 17:36:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev3 00:15:08.625 17:36:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:08.625 17:36:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:08.625 17:36:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:08.625 17:36:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:08.889 17:36:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:15:08.889 17:36:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:15:08.889 17:36:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:08.889 17:36:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev4 00:15:08.889 17:36:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:08.889 17:36:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:08.889 17:36:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:08.889 17:36:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:08.889 17:36:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:15:08.889 17:36:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:15:08.889 17:36:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@323 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:15:08.889 17:36:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:08.889 17:36:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:08.889 [2024-11-27 17:36:39.874095] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:15:08.889 [2024-11-27 17:36:39.874123] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:15:08.889 [2024-11-27 17:36:39.874193] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:08.889 [2024-11-27 17:36:39.874433] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:08.889 [2024-11-27 17:36:39.874443] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name Existed_Raid, state offline 00:15:08.897 17:36:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:08.897 17:36:39 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@326 -- # killprocess 93750 00:15:08.897 17:36:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@950 -- # '[' -z 93750 ']' 00:15:08.897 17:36:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@954 -- # kill -0 93750 00:15:08.897 17:36:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@955 -- # uname 00:15:08.897 17:36:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:15:08.897 17:36:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 93750 00:15:08.897 17:36:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:15:08.897 17:36:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:15:08.897 17:36:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@968 -- # echo 'killing process with pid 93750' 00:15:08.897 killing process with pid 93750 00:15:08.897 17:36:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@969 -- # kill 93750 00:15:08.897 [2024-11-27 17:36:39.915368] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:15:08.897 17:36:39 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@974 -- # wait 93750 00:15:08.897 [2024-11-27 17:36:39.955805] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:15:09.158 17:36:40 bdev_raid.raid5f_state_function_test_sb -- bdev/bdev_raid.sh@328 -- # return 0 00:15:09.158 00:15:09.158 real 0m9.695s 00:15:09.158 user 0m16.563s 00:15:09.158 sys 0m2.076s 00:15:09.158 17:36:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@1126 -- # xtrace_disable 00:15:09.158 17:36:40 bdev_raid.raid5f_state_function_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:09.158 ************************************ 00:15:09.158 END TEST raid5f_state_function_test_sb 00:15:09.158 ************************************ 00:15:09.158 17:36:40 bdev_raid -- bdev/bdev_raid.sh@988 -- # run_test raid5f_superblock_test raid_superblock_test raid5f 4 00:15:09.158 17:36:40 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:15:09.158 17:36:40 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:15:09.158 17:36:40 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:15:09.158 ************************************ 00:15:09.158 START TEST raid5f_superblock_test 00:15:09.158 ************************************ 00:15:09.158 17:36:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@1125 -- # raid_superblock_test raid5f 4 00:15:09.158 17:36:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@393 -- # local raid_level=raid5f 00:15:09.158 17:36:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=4 00:15:09.158 17:36:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:15:09.158 17:36:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:15:09.158 17:36:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:15:09.158 17:36:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:15:09.158 17:36:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:15:09.158 17:36:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:15:09.158 17:36:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:15:09.158 17:36:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@399 -- # local strip_size 00:15:09.158 17:36:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:15:09.158 17:36:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:15:09.158 17:36:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:15:09.158 17:36:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@404 -- # '[' raid5f '!=' raid1 ']' 00:15:09.158 17:36:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@405 -- # strip_size=64 00:15:09.158 17:36:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@406 -- # strip_size_create_arg='-z 64' 00:15:09.158 17:36:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@412 -- # raid_pid=94404 00:15:09.158 17:36:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:15:09.158 17:36:40 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@413 -- # waitforlisten 94404 00:15:09.158 17:36:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@831 -- # '[' -z 94404 ']' 00:15:09.158 17:36:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:15:09.158 17:36:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:15:09.158 17:36:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:15:09.158 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:15:09.158 17:36:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:15:09.158 17:36:40 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:09.418 [2024-11-27 17:36:40.368599] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:15:09.418 [2024-11-27 17:36:40.368787] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid94404 ] 00:15:09.418 [2024-11-27 17:36:40.512445] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:15:09.418 [2024-11-27 17:36:40.556094] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:15:09.418 [2024-11-27 17:36:40.598498] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:09.418 [2024-11-27 17:36:40.598536] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:10.360 17:36:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:15:10.360 17:36:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@864 -- # return 0 00:15:10.360 17:36:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:15:10.360 17:36:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:15:10.360 17:36:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:15:10.360 17:36:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:15:10.360 17:36:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:15:10.360 17:36:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:15:10.360 17:36:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:15:10.360 17:36:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:15:10.360 17:36:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc1 00:15:10.360 17:36:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:10.360 17:36:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:10.360 malloc1 00:15:10.360 17:36:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:10.360 17:36:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:15:10.360 17:36:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:10.360 17:36:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:10.360 [2024-11-27 17:36:41.216963] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:15:10.360 [2024-11-27 17:36:41.217100] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:10.360 [2024-11-27 17:36:41.217132] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:15:10.360 [2024-11-27 17:36:41.217182] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:10.360 [2024-11-27 17:36:41.219239] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:10.360 [2024-11-27 17:36:41.219311] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:15:10.360 pt1 00:15:10.360 17:36:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:10.360 17:36:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:15:10.360 17:36:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:15:10.360 17:36:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:15:10.360 17:36:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:15:10.360 17:36:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:15:10.360 17:36:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:15:10.360 17:36:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:15:10.360 17:36:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:15:10.360 17:36:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc2 00:15:10.360 17:36:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:10.360 17:36:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:10.360 malloc2 00:15:10.360 17:36:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:10.360 17:36:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:15:10.360 17:36:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:10.360 17:36:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:10.360 [2024-11-27 17:36:41.260736] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:15:10.360 [2024-11-27 17:36:41.260974] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:10.360 [2024-11-27 17:36:41.261067] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:15:10.360 [2024-11-27 17:36:41.261220] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:10.360 [2024-11-27 17:36:41.265175] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:10.360 [2024-11-27 17:36:41.265295] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:15:10.360 pt2 00:15:10.360 17:36:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:10.360 17:36:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:15:10.360 17:36:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:15:10.360 17:36:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc3 00:15:10.360 17:36:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt3 00:15:10.361 17:36:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000003 00:15:10.361 17:36:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:15:10.361 17:36:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:15:10.361 17:36:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:15:10.361 17:36:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc3 00:15:10.361 17:36:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:10.361 17:36:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:10.361 malloc3 00:15:10.361 17:36:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:10.361 17:36:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:15:10.361 17:36:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:10.361 17:36:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:10.361 [2024-11-27 17:36:41.290978] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:15:10.361 [2024-11-27 17:36:41.291091] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:10.361 [2024-11-27 17:36:41.291123] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:15:10.361 [2024-11-27 17:36:41.291165] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:10.361 [2024-11-27 17:36:41.293090] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:10.361 [2024-11-27 17:36:41.293166] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:15:10.361 pt3 00:15:10.361 17:36:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:10.361 17:36:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:15:10.361 17:36:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:15:10.361 17:36:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc4 00:15:10.361 17:36:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt4 00:15:10.361 17:36:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000004 00:15:10.361 17:36:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:15:10.361 17:36:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:15:10.361 17:36:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:15:10.361 17:36:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 512 -b malloc4 00:15:10.361 17:36:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:10.361 17:36:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:10.361 malloc4 00:15:10.361 17:36:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:10.361 17:36:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:15:10.361 17:36:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:10.361 17:36:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:10.361 [2024-11-27 17:36:41.323391] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:15:10.361 [2024-11-27 17:36:41.323492] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:10.361 [2024-11-27 17:36:41.323524] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:15:10.361 [2024-11-27 17:36:41.323553] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:10.361 [2024-11-27 17:36:41.325492] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:10.361 [2024-11-27 17:36:41.325558] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:15:10.361 pt4 00:15:10.361 17:36:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:10.361 17:36:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:15:10.361 17:36:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:15:10.361 17:36:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''pt1 pt2 pt3 pt4'\''' -n raid_bdev1 -s 00:15:10.361 17:36:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:10.361 17:36:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:10.361 [2024-11-27 17:36:41.335412] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:15:10.361 [2024-11-27 17:36:41.337180] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:15:10.361 [2024-11-27 17:36:41.337275] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:15:10.361 [2024-11-27 17:36:41.337334] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:15:10.361 [2024-11-27 17:36:41.337532] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:15:10.361 [2024-11-27 17:36:41.337577] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:15:10.361 [2024-11-27 17:36:41.337819] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:15:10.361 [2024-11-27 17:36:41.338305] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:15:10.361 [2024-11-27 17:36:41.338350] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:15:10.361 [2024-11-27 17:36:41.338504] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:10.361 17:36:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:10.361 17:36:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 4 00:15:10.361 17:36:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:10.361 17:36:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:10.361 17:36:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:10.361 17:36:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:10.361 17:36:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:10.361 17:36:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:10.361 17:36:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:10.361 17:36:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:10.361 17:36:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:10.361 17:36:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:10.361 17:36:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:10.361 17:36:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:10.361 17:36:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:10.361 17:36:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:10.361 17:36:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:10.361 "name": "raid_bdev1", 00:15:10.361 "uuid": "38dff1dc-17b5-4c86-ad2a-f6e7bbae90f6", 00:15:10.361 "strip_size_kb": 64, 00:15:10.361 "state": "online", 00:15:10.361 "raid_level": "raid5f", 00:15:10.361 "superblock": true, 00:15:10.361 "num_base_bdevs": 4, 00:15:10.361 "num_base_bdevs_discovered": 4, 00:15:10.361 "num_base_bdevs_operational": 4, 00:15:10.361 "base_bdevs_list": [ 00:15:10.361 { 00:15:10.361 "name": "pt1", 00:15:10.361 "uuid": "00000000-0000-0000-0000-000000000001", 00:15:10.361 "is_configured": true, 00:15:10.361 "data_offset": 2048, 00:15:10.361 "data_size": 63488 00:15:10.361 }, 00:15:10.361 { 00:15:10.361 "name": "pt2", 00:15:10.361 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:10.361 "is_configured": true, 00:15:10.361 "data_offset": 2048, 00:15:10.361 "data_size": 63488 00:15:10.361 }, 00:15:10.361 { 00:15:10.361 "name": "pt3", 00:15:10.361 "uuid": "00000000-0000-0000-0000-000000000003", 00:15:10.361 "is_configured": true, 00:15:10.361 "data_offset": 2048, 00:15:10.361 "data_size": 63488 00:15:10.361 }, 00:15:10.361 { 00:15:10.361 "name": "pt4", 00:15:10.361 "uuid": "00000000-0000-0000-0000-000000000004", 00:15:10.361 "is_configured": true, 00:15:10.361 "data_offset": 2048, 00:15:10.361 "data_size": 63488 00:15:10.361 } 00:15:10.361 ] 00:15:10.361 }' 00:15:10.361 17:36:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:10.361 17:36:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:10.622 17:36:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:15:10.622 17:36:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:15:10.622 17:36:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:15:10.622 17:36:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:15:10.622 17:36:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:15:10.622 17:36:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:15:10.622 17:36:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:15:10.622 17:36:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:15:10.622 17:36:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:10.622 17:36:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:10.622 [2024-11-27 17:36:41.743676] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:10.622 17:36:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:10.622 17:36:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:15:10.622 "name": "raid_bdev1", 00:15:10.622 "aliases": [ 00:15:10.622 "38dff1dc-17b5-4c86-ad2a-f6e7bbae90f6" 00:15:10.622 ], 00:15:10.622 "product_name": "Raid Volume", 00:15:10.622 "block_size": 512, 00:15:10.622 "num_blocks": 190464, 00:15:10.622 "uuid": "38dff1dc-17b5-4c86-ad2a-f6e7bbae90f6", 00:15:10.622 "assigned_rate_limits": { 00:15:10.622 "rw_ios_per_sec": 0, 00:15:10.622 "rw_mbytes_per_sec": 0, 00:15:10.622 "r_mbytes_per_sec": 0, 00:15:10.622 "w_mbytes_per_sec": 0 00:15:10.622 }, 00:15:10.622 "claimed": false, 00:15:10.622 "zoned": false, 00:15:10.622 "supported_io_types": { 00:15:10.622 "read": true, 00:15:10.622 "write": true, 00:15:10.622 "unmap": false, 00:15:10.622 "flush": false, 00:15:10.622 "reset": true, 00:15:10.622 "nvme_admin": false, 00:15:10.622 "nvme_io": false, 00:15:10.622 "nvme_io_md": false, 00:15:10.622 "write_zeroes": true, 00:15:10.622 "zcopy": false, 00:15:10.622 "get_zone_info": false, 00:15:10.622 "zone_management": false, 00:15:10.622 "zone_append": false, 00:15:10.622 "compare": false, 00:15:10.622 "compare_and_write": false, 00:15:10.622 "abort": false, 00:15:10.622 "seek_hole": false, 00:15:10.622 "seek_data": false, 00:15:10.622 "copy": false, 00:15:10.622 "nvme_iov_md": false 00:15:10.622 }, 00:15:10.622 "driver_specific": { 00:15:10.622 "raid": { 00:15:10.622 "uuid": "38dff1dc-17b5-4c86-ad2a-f6e7bbae90f6", 00:15:10.622 "strip_size_kb": 64, 00:15:10.622 "state": "online", 00:15:10.622 "raid_level": "raid5f", 00:15:10.622 "superblock": true, 00:15:10.622 "num_base_bdevs": 4, 00:15:10.622 "num_base_bdevs_discovered": 4, 00:15:10.622 "num_base_bdevs_operational": 4, 00:15:10.622 "base_bdevs_list": [ 00:15:10.622 { 00:15:10.622 "name": "pt1", 00:15:10.622 "uuid": "00000000-0000-0000-0000-000000000001", 00:15:10.622 "is_configured": true, 00:15:10.622 "data_offset": 2048, 00:15:10.622 "data_size": 63488 00:15:10.622 }, 00:15:10.622 { 00:15:10.622 "name": "pt2", 00:15:10.622 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:10.622 "is_configured": true, 00:15:10.623 "data_offset": 2048, 00:15:10.623 "data_size": 63488 00:15:10.623 }, 00:15:10.623 { 00:15:10.623 "name": "pt3", 00:15:10.623 "uuid": "00000000-0000-0000-0000-000000000003", 00:15:10.623 "is_configured": true, 00:15:10.623 "data_offset": 2048, 00:15:10.623 "data_size": 63488 00:15:10.623 }, 00:15:10.623 { 00:15:10.623 "name": "pt4", 00:15:10.623 "uuid": "00000000-0000-0000-0000-000000000004", 00:15:10.623 "is_configured": true, 00:15:10.623 "data_offset": 2048, 00:15:10.623 "data_size": 63488 00:15:10.623 } 00:15:10.623 ] 00:15:10.623 } 00:15:10.623 } 00:15:10.623 }' 00:15:10.623 17:36:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:15:10.883 17:36:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:15:10.883 pt2 00:15:10.883 pt3 00:15:10.883 pt4' 00:15:10.883 17:36:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:10.883 17:36:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:15:10.883 17:36:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:10.883 17:36:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:10.883 17:36:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:15:10.883 17:36:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:10.883 17:36:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:10.883 17:36:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:10.883 17:36:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:15:10.883 17:36:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:15:10.883 17:36:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:10.883 17:36:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:15:10.883 17:36:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:10.883 17:36:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:10.883 17:36:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:10.883 17:36:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:10.884 17:36:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:15:10.884 17:36:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:15:10.884 17:36:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:10.884 17:36:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:15:10.884 17:36:41 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:10.884 17:36:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:10.884 17:36:41 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:10.884 17:36:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:10.884 17:36:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:15:10.884 17:36:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:15:10.884 17:36:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:10.884 17:36:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt4 00:15:10.884 17:36:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:10.884 17:36:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:10.884 17:36:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:10.884 17:36:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:11.144 17:36:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:15:11.144 17:36:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:15:11.144 17:36:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:15:11.144 17:36:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:15:11.144 17:36:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:11.144 17:36:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:11.144 [2024-11-27 17:36:42.087237] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:11.144 17:36:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:11.144 17:36:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=38dff1dc-17b5-4c86-ad2a-f6e7bbae90f6 00:15:11.144 17:36:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@436 -- # '[' -z 38dff1dc-17b5-4c86-ad2a-f6e7bbae90f6 ']' 00:15:11.144 17:36:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:15:11.144 17:36:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:11.144 17:36:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:11.144 [2024-11-27 17:36:42.127092] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:15:11.144 [2024-11-27 17:36:42.127180] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:15:11.144 [2024-11-27 17:36:42.127248] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:11.144 [2024-11-27 17:36:42.127325] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:11.144 [2024-11-27 17:36:42.127344] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:15:11.144 17:36:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:11.144 17:36:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:11.144 17:36:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:11.144 17:36:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:15:11.144 17:36:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:11.144 17:36:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:11.144 17:36:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:15:11.144 17:36:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:15:11.144 17:36:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:15:11.144 17:36:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:15:11.144 17:36:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:11.144 17:36:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:11.144 17:36:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:11.145 17:36:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:15:11.145 17:36:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:15:11.145 17:36:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:11.145 17:36:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:11.145 17:36:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:11.145 17:36:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:15:11.145 17:36:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt3 00:15:11.145 17:36:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:11.145 17:36:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:11.145 17:36:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:11.145 17:36:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:15:11.145 17:36:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt4 00:15:11.145 17:36:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:11.145 17:36:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:11.145 17:36:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:11.145 17:36:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:15:11.145 17:36:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:11.145 17:36:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:15:11.145 17:36:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:11.145 17:36:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:11.145 17:36:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:15:11.145 17:36:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:15:11.145 17:36:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@650 -- # local es=0 00:15:11.145 17:36:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:15:11.145 17:36:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:15:11.145 17:36:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:15:11.145 17:36:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:15:11.145 17:36:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:15:11.145 17:36:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''malloc1 malloc2 malloc3 malloc4'\''' -n raid_bdev1 00:15:11.145 17:36:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:11.145 17:36:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:11.145 [2024-11-27 17:36:42.286882] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:15:11.145 [2024-11-27 17:36:42.288681] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:15:11.145 [2024-11-27 17:36:42.288761] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc3 is claimed 00:15:11.145 [2024-11-27 17:36:42.288811] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc4 is claimed 00:15:11.145 [2024-11-27 17:36:42.288882] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:15:11.145 [2024-11-27 17:36:42.288955] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:15:11.145 [2024-11-27 17:36:42.289041] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc3 00:15:11.145 [2024-11-27 17:36:42.289093] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc4 00:15:11.145 [2024-11-27 17:36:42.289169] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:15:11.145 [2024-11-27 17:36:42.289202] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state configuring 00:15:11.145 request: 00:15:11.145 { 00:15:11.145 "name": "raid_bdev1", 00:15:11.145 "raid_level": "raid5f", 00:15:11.145 "base_bdevs": [ 00:15:11.145 "malloc1", 00:15:11.145 "malloc2", 00:15:11.145 "malloc3", 00:15:11.145 "malloc4" 00:15:11.145 ], 00:15:11.145 "strip_size_kb": 64, 00:15:11.145 "superblock": false, 00:15:11.145 "method": "bdev_raid_create", 00:15:11.145 "req_id": 1 00:15:11.145 } 00:15:11.145 Got JSON-RPC error response 00:15:11.145 response: 00:15:11.145 { 00:15:11.145 "code": -17, 00:15:11.145 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:15:11.145 } 00:15:11.145 17:36:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:15:11.145 17:36:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@653 -- # es=1 00:15:11.145 17:36:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:15:11.145 17:36:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:15:11.145 17:36:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:15:11.145 17:36:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:11.145 17:36:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:15:11.145 17:36:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:11.145 17:36:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:11.145 17:36:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:11.405 17:36:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:15:11.405 17:36:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:15:11.406 17:36:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:15:11.406 17:36:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:11.406 17:36:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:11.406 [2024-11-27 17:36:42.354723] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:15:11.406 [2024-11-27 17:36:42.354767] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:11.406 [2024-11-27 17:36:42.354788] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:15:11.406 [2024-11-27 17:36:42.354796] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:11.406 [2024-11-27 17:36:42.356734] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:11.406 [2024-11-27 17:36:42.356770] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:15:11.406 [2024-11-27 17:36:42.356828] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:15:11.406 [2024-11-27 17:36:42.356863] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:15:11.406 pt1 00:15:11.406 17:36:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:11.406 17:36:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 4 00:15:11.406 17:36:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:11.406 17:36:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:11.406 17:36:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:11.406 17:36:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:11.406 17:36:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:11.406 17:36:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:11.406 17:36:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:11.406 17:36:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:11.406 17:36:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:11.406 17:36:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:11.406 17:36:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:11.406 17:36:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:11.406 17:36:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:11.406 17:36:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:11.406 17:36:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:11.406 "name": "raid_bdev1", 00:15:11.406 "uuid": "38dff1dc-17b5-4c86-ad2a-f6e7bbae90f6", 00:15:11.406 "strip_size_kb": 64, 00:15:11.406 "state": "configuring", 00:15:11.406 "raid_level": "raid5f", 00:15:11.406 "superblock": true, 00:15:11.406 "num_base_bdevs": 4, 00:15:11.406 "num_base_bdevs_discovered": 1, 00:15:11.406 "num_base_bdevs_operational": 4, 00:15:11.406 "base_bdevs_list": [ 00:15:11.406 { 00:15:11.406 "name": "pt1", 00:15:11.406 "uuid": "00000000-0000-0000-0000-000000000001", 00:15:11.406 "is_configured": true, 00:15:11.406 "data_offset": 2048, 00:15:11.406 "data_size": 63488 00:15:11.406 }, 00:15:11.406 { 00:15:11.406 "name": null, 00:15:11.406 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:11.406 "is_configured": false, 00:15:11.406 "data_offset": 2048, 00:15:11.406 "data_size": 63488 00:15:11.406 }, 00:15:11.406 { 00:15:11.406 "name": null, 00:15:11.406 "uuid": "00000000-0000-0000-0000-000000000003", 00:15:11.406 "is_configured": false, 00:15:11.406 "data_offset": 2048, 00:15:11.406 "data_size": 63488 00:15:11.406 }, 00:15:11.406 { 00:15:11.406 "name": null, 00:15:11.406 "uuid": "00000000-0000-0000-0000-000000000004", 00:15:11.406 "is_configured": false, 00:15:11.406 "data_offset": 2048, 00:15:11.406 "data_size": 63488 00:15:11.406 } 00:15:11.406 ] 00:15:11.406 }' 00:15:11.406 17:36:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:11.406 17:36:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:11.666 17:36:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@470 -- # '[' 4 -gt 2 ']' 00:15:11.666 17:36:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@472 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:15:11.666 17:36:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:11.666 17:36:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:11.666 [2024-11-27 17:36:42.834113] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:15:11.666 [2024-11-27 17:36:42.834214] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:11.666 [2024-11-27 17:36:42.834247] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009c80 00:15:11.666 [2024-11-27 17:36:42.834274] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:11.666 [2024-11-27 17:36:42.834581] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:11.666 [2024-11-27 17:36:42.834632] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:15:11.666 [2024-11-27 17:36:42.834709] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:15:11.666 [2024-11-27 17:36:42.834761] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:15:11.666 pt2 00:15:11.666 17:36:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:11.666 17:36:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@473 -- # rpc_cmd bdev_passthru_delete pt2 00:15:11.666 17:36:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:11.666 17:36:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:11.666 [2024-11-27 17:36:42.846126] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: pt2 00:15:11.666 17:36:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:11.666 17:36:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@474 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 4 00:15:11.666 17:36:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:11.666 17:36:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:11.666 17:36:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:11.666 17:36:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:11.666 17:36:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:11.666 17:36:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:11.666 17:36:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:11.666 17:36:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:11.666 17:36:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:11.926 17:36:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:11.926 17:36:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:11.926 17:36:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:11.926 17:36:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:11.926 17:36:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:11.926 17:36:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:11.926 "name": "raid_bdev1", 00:15:11.926 "uuid": "38dff1dc-17b5-4c86-ad2a-f6e7bbae90f6", 00:15:11.926 "strip_size_kb": 64, 00:15:11.926 "state": "configuring", 00:15:11.926 "raid_level": "raid5f", 00:15:11.926 "superblock": true, 00:15:11.926 "num_base_bdevs": 4, 00:15:11.926 "num_base_bdevs_discovered": 1, 00:15:11.926 "num_base_bdevs_operational": 4, 00:15:11.926 "base_bdevs_list": [ 00:15:11.926 { 00:15:11.926 "name": "pt1", 00:15:11.926 "uuid": "00000000-0000-0000-0000-000000000001", 00:15:11.926 "is_configured": true, 00:15:11.926 "data_offset": 2048, 00:15:11.926 "data_size": 63488 00:15:11.926 }, 00:15:11.926 { 00:15:11.926 "name": null, 00:15:11.926 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:11.926 "is_configured": false, 00:15:11.926 "data_offset": 0, 00:15:11.926 "data_size": 63488 00:15:11.926 }, 00:15:11.926 { 00:15:11.926 "name": null, 00:15:11.926 "uuid": "00000000-0000-0000-0000-000000000003", 00:15:11.926 "is_configured": false, 00:15:11.926 "data_offset": 2048, 00:15:11.926 "data_size": 63488 00:15:11.926 }, 00:15:11.926 { 00:15:11.926 "name": null, 00:15:11.926 "uuid": "00000000-0000-0000-0000-000000000004", 00:15:11.926 "is_configured": false, 00:15:11.927 "data_offset": 2048, 00:15:11.927 "data_size": 63488 00:15:11.927 } 00:15:11.927 ] 00:15:11.927 }' 00:15:11.927 17:36:42 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:11.927 17:36:42 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:12.187 17:36:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:15:12.187 17:36:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:15:12.187 17:36:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:15:12.187 17:36:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:12.187 17:36:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:12.187 [2024-11-27 17:36:43.293344] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:15:12.187 [2024-11-27 17:36:43.293448] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:12.187 [2024-11-27 17:36:43.293482] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009f80 00:15:12.187 [2024-11-27 17:36:43.293512] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:12.187 [2024-11-27 17:36:43.293849] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:12.187 [2024-11-27 17:36:43.293902] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:15:12.187 [2024-11-27 17:36:43.293964] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:15:12.187 [2024-11-27 17:36:43.293988] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:15:12.187 pt2 00:15:12.187 17:36:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:12.187 17:36:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:15:12.187 17:36:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:15:12.187 17:36:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:15:12.187 17:36:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:12.187 17:36:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:12.187 [2024-11-27 17:36:43.305290] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:15:12.187 [2024-11-27 17:36:43.305339] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:12.187 [2024-11-27 17:36:43.305354] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:15:12.187 [2024-11-27 17:36:43.305371] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:12.187 [2024-11-27 17:36:43.305646] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:12.187 [2024-11-27 17:36:43.305662] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:15:12.187 [2024-11-27 17:36:43.305712] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:15:12.187 [2024-11-27 17:36:43.305730] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:15:12.187 pt3 00:15:12.187 17:36:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:12.187 17:36:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:15:12.187 17:36:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:15:12.187 17:36:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:15:12.187 17:36:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:12.187 17:36:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:12.187 [2024-11-27 17:36:43.317302] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:15:12.187 [2024-11-27 17:36:43.317350] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:12.187 [2024-11-27 17:36:43.317362] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a580 00:15:12.187 [2024-11-27 17:36:43.317372] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:12.187 [2024-11-27 17:36:43.317624] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:12.187 [2024-11-27 17:36:43.317641] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:15:12.188 [2024-11-27 17:36:43.317685] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt4 00:15:12.188 [2024-11-27 17:36:43.317701] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:15:12.188 [2024-11-27 17:36:43.317798] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:15:12.188 [2024-11-27 17:36:43.317810] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:15:12.188 [2024-11-27 17:36:43.318000] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:15:12.188 [2024-11-27 17:36:43.318446] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:15:12.188 [2024-11-27 17:36:43.318463] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:15:12.188 [2024-11-27 17:36:43.318571] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:12.188 pt4 00:15:12.188 17:36:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:12.188 17:36:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:15:12.188 17:36:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:15:12.188 17:36:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 4 00:15:12.188 17:36:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:12.188 17:36:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:12.188 17:36:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:12.188 17:36:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:12.188 17:36:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:12.188 17:36:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:12.188 17:36:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:12.188 17:36:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:12.188 17:36:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:12.188 17:36:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:12.188 17:36:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:12.188 17:36:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:12.188 17:36:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:12.188 17:36:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:12.188 17:36:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:12.188 "name": "raid_bdev1", 00:15:12.188 "uuid": "38dff1dc-17b5-4c86-ad2a-f6e7bbae90f6", 00:15:12.188 "strip_size_kb": 64, 00:15:12.188 "state": "online", 00:15:12.188 "raid_level": "raid5f", 00:15:12.188 "superblock": true, 00:15:12.188 "num_base_bdevs": 4, 00:15:12.188 "num_base_bdevs_discovered": 4, 00:15:12.188 "num_base_bdevs_operational": 4, 00:15:12.188 "base_bdevs_list": [ 00:15:12.188 { 00:15:12.188 "name": "pt1", 00:15:12.188 "uuid": "00000000-0000-0000-0000-000000000001", 00:15:12.188 "is_configured": true, 00:15:12.188 "data_offset": 2048, 00:15:12.188 "data_size": 63488 00:15:12.188 }, 00:15:12.188 { 00:15:12.188 "name": "pt2", 00:15:12.188 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:12.188 "is_configured": true, 00:15:12.188 "data_offset": 2048, 00:15:12.188 "data_size": 63488 00:15:12.188 }, 00:15:12.188 { 00:15:12.188 "name": "pt3", 00:15:12.188 "uuid": "00000000-0000-0000-0000-000000000003", 00:15:12.188 "is_configured": true, 00:15:12.188 "data_offset": 2048, 00:15:12.188 "data_size": 63488 00:15:12.188 }, 00:15:12.188 { 00:15:12.188 "name": "pt4", 00:15:12.188 "uuid": "00000000-0000-0000-0000-000000000004", 00:15:12.188 "is_configured": true, 00:15:12.188 "data_offset": 2048, 00:15:12.188 "data_size": 63488 00:15:12.188 } 00:15:12.188 ] 00:15:12.188 }' 00:15:12.188 17:36:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:12.188 17:36:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:12.758 17:36:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:15:12.758 17:36:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:15:12.758 17:36:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:15:12.758 17:36:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:15:12.758 17:36:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@184 -- # local name 00:15:12.758 17:36:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:15:12.758 17:36:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:15:12.758 17:36:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:15:12.758 17:36:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:12.759 17:36:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:12.759 [2024-11-27 17:36:43.788639] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:12.759 17:36:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:12.759 17:36:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:15:12.759 "name": "raid_bdev1", 00:15:12.759 "aliases": [ 00:15:12.759 "38dff1dc-17b5-4c86-ad2a-f6e7bbae90f6" 00:15:12.759 ], 00:15:12.759 "product_name": "Raid Volume", 00:15:12.759 "block_size": 512, 00:15:12.759 "num_blocks": 190464, 00:15:12.759 "uuid": "38dff1dc-17b5-4c86-ad2a-f6e7bbae90f6", 00:15:12.759 "assigned_rate_limits": { 00:15:12.759 "rw_ios_per_sec": 0, 00:15:12.759 "rw_mbytes_per_sec": 0, 00:15:12.759 "r_mbytes_per_sec": 0, 00:15:12.759 "w_mbytes_per_sec": 0 00:15:12.759 }, 00:15:12.759 "claimed": false, 00:15:12.759 "zoned": false, 00:15:12.759 "supported_io_types": { 00:15:12.759 "read": true, 00:15:12.759 "write": true, 00:15:12.759 "unmap": false, 00:15:12.759 "flush": false, 00:15:12.759 "reset": true, 00:15:12.759 "nvme_admin": false, 00:15:12.759 "nvme_io": false, 00:15:12.759 "nvme_io_md": false, 00:15:12.759 "write_zeroes": true, 00:15:12.759 "zcopy": false, 00:15:12.759 "get_zone_info": false, 00:15:12.759 "zone_management": false, 00:15:12.759 "zone_append": false, 00:15:12.759 "compare": false, 00:15:12.759 "compare_and_write": false, 00:15:12.759 "abort": false, 00:15:12.759 "seek_hole": false, 00:15:12.759 "seek_data": false, 00:15:12.759 "copy": false, 00:15:12.759 "nvme_iov_md": false 00:15:12.759 }, 00:15:12.759 "driver_specific": { 00:15:12.759 "raid": { 00:15:12.759 "uuid": "38dff1dc-17b5-4c86-ad2a-f6e7bbae90f6", 00:15:12.759 "strip_size_kb": 64, 00:15:12.759 "state": "online", 00:15:12.759 "raid_level": "raid5f", 00:15:12.759 "superblock": true, 00:15:12.759 "num_base_bdevs": 4, 00:15:12.759 "num_base_bdevs_discovered": 4, 00:15:12.759 "num_base_bdevs_operational": 4, 00:15:12.759 "base_bdevs_list": [ 00:15:12.759 { 00:15:12.759 "name": "pt1", 00:15:12.759 "uuid": "00000000-0000-0000-0000-000000000001", 00:15:12.759 "is_configured": true, 00:15:12.759 "data_offset": 2048, 00:15:12.759 "data_size": 63488 00:15:12.759 }, 00:15:12.759 { 00:15:12.759 "name": "pt2", 00:15:12.759 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:12.759 "is_configured": true, 00:15:12.759 "data_offset": 2048, 00:15:12.759 "data_size": 63488 00:15:12.759 }, 00:15:12.759 { 00:15:12.759 "name": "pt3", 00:15:12.759 "uuid": "00000000-0000-0000-0000-000000000003", 00:15:12.759 "is_configured": true, 00:15:12.759 "data_offset": 2048, 00:15:12.759 "data_size": 63488 00:15:12.759 }, 00:15:12.759 { 00:15:12.759 "name": "pt4", 00:15:12.759 "uuid": "00000000-0000-0000-0000-000000000004", 00:15:12.759 "is_configured": true, 00:15:12.759 "data_offset": 2048, 00:15:12.759 "data_size": 63488 00:15:12.759 } 00:15:12.759 ] 00:15:12.759 } 00:15:12.759 } 00:15:12.759 }' 00:15:12.759 17:36:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:15:12.759 17:36:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:15:12.759 pt2 00:15:12.759 pt3 00:15:12.759 pt4' 00:15:12.759 17:36:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:12.759 17:36:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='512 ' 00:15:12.759 17:36:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:12.759 17:36:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:15:12.759 17:36:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:12.759 17:36:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:12.759 17:36:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:12.759 17:36:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:13.020 17:36:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:15:13.020 17:36:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:15:13.020 17:36:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:13.020 17:36:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:15:13.020 17:36:43 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:13.020 17:36:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:13.020 17:36:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:13.020 17:36:43 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:13.020 17:36:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:15:13.020 17:36:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:15:13.020 17:36:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:13.020 17:36:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt3 00:15:13.020 17:36:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:13.020 17:36:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:13.020 17:36:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:13.020 17:36:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:13.020 17:36:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:15:13.020 17:36:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:15:13.020 17:36:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:15:13.020 17:36:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt4 00:15:13.020 17:36:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:15:13.020 17:36:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:13.020 17:36:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:13.020 17:36:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:13.020 17:36:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='512 ' 00:15:13.020 17:36:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@193 -- # [[ 512 == \5\1\2\ \ \ ]] 00:15:13.020 17:36:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:15:13.020 17:36:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:13.020 17:36:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:13.020 17:36:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:15:13.020 [2024-11-27 17:36:44.120098] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:13.020 17:36:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:13.020 17:36:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@487 -- # '[' 38dff1dc-17b5-4c86-ad2a-f6e7bbae90f6 '!=' 38dff1dc-17b5-4c86-ad2a-f6e7bbae90f6 ']' 00:15:13.020 17:36:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@491 -- # has_redundancy raid5f 00:15:13.020 17:36:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@198 -- # case $1 in 00:15:13.020 17:36:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@199 -- # return 0 00:15:13.020 17:36:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@493 -- # rpc_cmd bdev_passthru_delete pt1 00:15:13.020 17:36:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:13.020 17:36:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:13.020 [2024-11-27 17:36:44.167867] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: pt1 00:15:13.020 17:36:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:13.020 17:36:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@496 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:15:13.020 17:36:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:13.020 17:36:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:13.020 17:36:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:13.020 17:36:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:13.020 17:36:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:15:13.020 17:36:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:13.020 17:36:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:13.020 17:36:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:13.020 17:36:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:13.020 17:36:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:13.020 17:36:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:13.020 17:36:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:13.020 17:36:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:13.020 17:36:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:13.280 17:36:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:13.280 "name": "raid_bdev1", 00:15:13.280 "uuid": "38dff1dc-17b5-4c86-ad2a-f6e7bbae90f6", 00:15:13.280 "strip_size_kb": 64, 00:15:13.280 "state": "online", 00:15:13.280 "raid_level": "raid5f", 00:15:13.281 "superblock": true, 00:15:13.281 "num_base_bdevs": 4, 00:15:13.281 "num_base_bdevs_discovered": 3, 00:15:13.281 "num_base_bdevs_operational": 3, 00:15:13.281 "base_bdevs_list": [ 00:15:13.281 { 00:15:13.281 "name": null, 00:15:13.281 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:13.281 "is_configured": false, 00:15:13.281 "data_offset": 0, 00:15:13.281 "data_size": 63488 00:15:13.281 }, 00:15:13.281 { 00:15:13.281 "name": "pt2", 00:15:13.281 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:13.281 "is_configured": true, 00:15:13.281 "data_offset": 2048, 00:15:13.281 "data_size": 63488 00:15:13.281 }, 00:15:13.281 { 00:15:13.281 "name": "pt3", 00:15:13.281 "uuid": "00000000-0000-0000-0000-000000000003", 00:15:13.281 "is_configured": true, 00:15:13.281 "data_offset": 2048, 00:15:13.281 "data_size": 63488 00:15:13.281 }, 00:15:13.281 { 00:15:13.281 "name": "pt4", 00:15:13.281 "uuid": "00000000-0000-0000-0000-000000000004", 00:15:13.281 "is_configured": true, 00:15:13.281 "data_offset": 2048, 00:15:13.281 "data_size": 63488 00:15:13.281 } 00:15:13.281 ] 00:15:13.281 }' 00:15:13.281 17:36:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:13.281 17:36:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:13.541 17:36:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@499 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:15:13.541 17:36:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:13.541 17:36:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:13.541 [2024-11-27 17:36:44.611080] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:15:13.541 [2024-11-27 17:36:44.611162] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:15:13.541 [2024-11-27 17:36:44.611234] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:13.541 [2024-11-27 17:36:44.611310] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:13.541 [2024-11-27 17:36:44.611367] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:15:13.541 17:36:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:13.541 17:36:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@500 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:13.541 17:36:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@500 -- # jq -r '.[]' 00:15:13.541 17:36:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:13.541 17:36:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:13.541 17:36:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:13.541 17:36:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@500 -- # raid_bdev= 00:15:13.541 17:36:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@501 -- # '[' -n '' ']' 00:15:13.541 17:36:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i = 1 )) 00:15:13.541 17:36:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:15:13.541 17:36:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt2 00:15:13.541 17:36:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:13.541 17:36:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:13.541 17:36:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:13.541 17:36:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:15:13.541 17:36:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:15:13.541 17:36:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt3 00:15:13.541 17:36:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:13.541 17:36:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:13.541 17:36:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:13.541 17:36:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:15:13.541 17:36:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:15:13.541 17:36:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt4 00:15:13.541 17:36:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:13.541 17:36:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:13.542 17:36:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:13.542 17:36:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:15:13.542 17:36:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:15:13.542 17:36:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i = 1 )) 00:15:13.542 17:36:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:15:13.542 17:36:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@512 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:15:13.542 17:36:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:13.542 17:36:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:13.542 [2024-11-27 17:36:44.711085] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:15:13.542 [2024-11-27 17:36:44.711133] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:13.542 [2024-11-27 17:36:44.711159] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a880 00:15:13.542 [2024-11-27 17:36:44.711169] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:13.542 [2024-11-27 17:36:44.713190] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:13.542 [2024-11-27 17:36:44.713226] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:15:13.542 [2024-11-27 17:36:44.713283] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:15:13.542 [2024-11-27 17:36:44.713319] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:15:13.542 pt2 00:15:13.542 17:36:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:13.542 17:36:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@515 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 3 00:15:13.542 17:36:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:13.542 17:36:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:13.542 17:36:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:13.542 17:36:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:13.542 17:36:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:15:13.542 17:36:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:13.542 17:36:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:13.542 17:36:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:13.542 17:36:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:13.542 17:36:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:13.542 17:36:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:13.542 17:36:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:13.542 17:36:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:13.802 17:36:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:13.802 17:36:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:13.802 "name": "raid_bdev1", 00:15:13.802 "uuid": "38dff1dc-17b5-4c86-ad2a-f6e7bbae90f6", 00:15:13.802 "strip_size_kb": 64, 00:15:13.802 "state": "configuring", 00:15:13.802 "raid_level": "raid5f", 00:15:13.802 "superblock": true, 00:15:13.802 "num_base_bdevs": 4, 00:15:13.802 "num_base_bdevs_discovered": 1, 00:15:13.802 "num_base_bdevs_operational": 3, 00:15:13.802 "base_bdevs_list": [ 00:15:13.802 { 00:15:13.802 "name": null, 00:15:13.802 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:13.802 "is_configured": false, 00:15:13.802 "data_offset": 2048, 00:15:13.802 "data_size": 63488 00:15:13.802 }, 00:15:13.802 { 00:15:13.802 "name": "pt2", 00:15:13.802 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:13.802 "is_configured": true, 00:15:13.802 "data_offset": 2048, 00:15:13.802 "data_size": 63488 00:15:13.802 }, 00:15:13.802 { 00:15:13.802 "name": null, 00:15:13.802 "uuid": "00000000-0000-0000-0000-000000000003", 00:15:13.802 "is_configured": false, 00:15:13.802 "data_offset": 2048, 00:15:13.802 "data_size": 63488 00:15:13.802 }, 00:15:13.802 { 00:15:13.802 "name": null, 00:15:13.802 "uuid": "00000000-0000-0000-0000-000000000004", 00:15:13.802 "is_configured": false, 00:15:13.802 "data_offset": 2048, 00:15:13.802 "data_size": 63488 00:15:13.802 } 00:15:13.802 ] 00:15:13.802 }' 00:15:13.802 17:36:44 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:13.802 17:36:44 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:14.063 17:36:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i++ )) 00:15:14.063 17:36:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:15:14.063 17:36:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@512 -- # rpc_cmd bdev_passthru_create -b malloc3 -p pt3 -u 00000000-0000-0000-0000-000000000003 00:15:14.063 17:36:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:14.063 17:36:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:14.063 [2024-11-27 17:36:45.170563] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc3 00:15:14.063 [2024-11-27 17:36:45.170658] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:14.063 [2024-11-27 17:36:45.170689] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ae80 00:15:14.063 [2024-11-27 17:36:45.170719] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:14.063 [2024-11-27 17:36:45.171058] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:14.063 [2024-11-27 17:36:45.171115] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt3 00:15:14.063 [2024-11-27 17:36:45.171210] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt3 00:15:14.063 [2024-11-27 17:36:45.171258] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:15:14.063 pt3 00:15:14.063 17:36:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:14.063 17:36:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@515 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 3 00:15:14.063 17:36:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:14.063 17:36:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:14.063 17:36:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:14.063 17:36:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:14.063 17:36:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:15:14.063 17:36:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:14.063 17:36:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:14.063 17:36:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:14.063 17:36:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:14.063 17:36:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:14.063 17:36:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:14.063 17:36:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:14.063 17:36:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:14.063 17:36:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:14.063 17:36:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:14.063 "name": "raid_bdev1", 00:15:14.063 "uuid": "38dff1dc-17b5-4c86-ad2a-f6e7bbae90f6", 00:15:14.063 "strip_size_kb": 64, 00:15:14.063 "state": "configuring", 00:15:14.063 "raid_level": "raid5f", 00:15:14.063 "superblock": true, 00:15:14.063 "num_base_bdevs": 4, 00:15:14.063 "num_base_bdevs_discovered": 2, 00:15:14.063 "num_base_bdevs_operational": 3, 00:15:14.063 "base_bdevs_list": [ 00:15:14.063 { 00:15:14.063 "name": null, 00:15:14.063 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:14.063 "is_configured": false, 00:15:14.063 "data_offset": 2048, 00:15:14.063 "data_size": 63488 00:15:14.063 }, 00:15:14.063 { 00:15:14.063 "name": "pt2", 00:15:14.063 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:14.063 "is_configured": true, 00:15:14.063 "data_offset": 2048, 00:15:14.063 "data_size": 63488 00:15:14.063 }, 00:15:14.063 { 00:15:14.063 "name": "pt3", 00:15:14.063 "uuid": "00000000-0000-0000-0000-000000000003", 00:15:14.063 "is_configured": true, 00:15:14.063 "data_offset": 2048, 00:15:14.063 "data_size": 63488 00:15:14.063 }, 00:15:14.063 { 00:15:14.063 "name": null, 00:15:14.063 "uuid": "00000000-0000-0000-0000-000000000004", 00:15:14.063 "is_configured": false, 00:15:14.063 "data_offset": 2048, 00:15:14.063 "data_size": 63488 00:15:14.063 } 00:15:14.063 ] 00:15:14.063 }' 00:15:14.063 17:36:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:14.064 17:36:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:14.633 17:36:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i++ )) 00:15:14.634 17:36:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:15:14.634 17:36:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@519 -- # i=3 00:15:14.634 17:36:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@520 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:15:14.634 17:36:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:14.634 17:36:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:14.634 [2024-11-27 17:36:45.665724] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:15:14.634 [2024-11-27 17:36:45.665776] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:14.634 [2024-11-27 17:36:45.665795] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b180 00:15:14.634 [2024-11-27 17:36:45.665805] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:14.634 [2024-11-27 17:36:45.666130] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:14.634 [2024-11-27 17:36:45.666163] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:15:14.634 [2024-11-27 17:36:45.666222] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt4 00:15:14.634 [2024-11-27 17:36:45.666242] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:15:14.634 [2024-11-27 17:36:45.666327] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:15:14.634 [2024-11-27 17:36:45.666338] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:15:14.634 [2024-11-27 17:36:45.666545] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000026d0 00:15:14.634 [2024-11-27 17:36:45.667068] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:15:14.634 [2024-11-27 17:36:45.667087] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001c80 00:15:14.634 [2024-11-27 17:36:45.667325] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:14.634 pt4 00:15:14.634 17:36:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:14.634 17:36:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@523 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:15:14.634 17:36:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:14.634 17:36:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:14.634 17:36:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:14.634 17:36:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:14.634 17:36:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:15:14.634 17:36:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:14.634 17:36:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:14.634 17:36:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:14.634 17:36:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:14.634 17:36:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:14.634 17:36:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:14.634 17:36:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:14.634 17:36:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:14.634 17:36:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:14.634 17:36:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:14.634 "name": "raid_bdev1", 00:15:14.634 "uuid": "38dff1dc-17b5-4c86-ad2a-f6e7bbae90f6", 00:15:14.634 "strip_size_kb": 64, 00:15:14.634 "state": "online", 00:15:14.634 "raid_level": "raid5f", 00:15:14.634 "superblock": true, 00:15:14.634 "num_base_bdevs": 4, 00:15:14.634 "num_base_bdevs_discovered": 3, 00:15:14.634 "num_base_bdevs_operational": 3, 00:15:14.634 "base_bdevs_list": [ 00:15:14.634 { 00:15:14.634 "name": null, 00:15:14.634 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:14.634 "is_configured": false, 00:15:14.634 "data_offset": 2048, 00:15:14.634 "data_size": 63488 00:15:14.634 }, 00:15:14.634 { 00:15:14.634 "name": "pt2", 00:15:14.634 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:14.634 "is_configured": true, 00:15:14.634 "data_offset": 2048, 00:15:14.634 "data_size": 63488 00:15:14.634 }, 00:15:14.634 { 00:15:14.634 "name": "pt3", 00:15:14.634 "uuid": "00000000-0000-0000-0000-000000000003", 00:15:14.634 "is_configured": true, 00:15:14.634 "data_offset": 2048, 00:15:14.634 "data_size": 63488 00:15:14.634 }, 00:15:14.634 { 00:15:14.634 "name": "pt4", 00:15:14.634 "uuid": "00000000-0000-0000-0000-000000000004", 00:15:14.634 "is_configured": true, 00:15:14.634 "data_offset": 2048, 00:15:14.634 "data_size": 63488 00:15:14.634 } 00:15:14.634 ] 00:15:14.634 }' 00:15:14.634 17:36:45 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:14.634 17:36:45 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:14.894 17:36:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@526 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:15:15.154 17:36:46 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:15.154 17:36:46 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:15.154 [2024-11-27 17:36:46.088996] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:15:15.154 [2024-11-27 17:36:46.089070] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:15:15.154 [2024-11-27 17:36:46.089147] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:15.154 [2024-11-27 17:36:46.089226] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:15.155 [2024-11-27 17:36:46.089293] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name raid_bdev1, state offline 00:15:15.155 17:36:46 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:15.155 17:36:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@527 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:15.155 17:36:46 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:15.155 17:36:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@527 -- # jq -r '.[]' 00:15:15.155 17:36:46 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:15.155 17:36:46 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:15.155 17:36:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@527 -- # raid_bdev= 00:15:15.155 17:36:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@528 -- # '[' -n '' ']' 00:15:15.155 17:36:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@532 -- # '[' 4 -gt 2 ']' 00:15:15.155 17:36:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@534 -- # i=3 00:15:15.155 17:36:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@535 -- # rpc_cmd bdev_passthru_delete pt4 00:15:15.155 17:36:46 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:15.155 17:36:46 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:15.155 17:36:46 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:15.155 17:36:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@540 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:15:15.155 17:36:46 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:15.155 17:36:46 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:15.155 [2024-11-27 17:36:46.164874] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:15:15.155 [2024-11-27 17:36:46.164921] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:15.155 [2024-11-27 17:36:46.164937] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b480 00:15:15.155 [2024-11-27 17:36:46.164945] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:15.155 [2024-11-27 17:36:46.167044] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:15.155 [2024-11-27 17:36:46.167114] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:15:15.155 [2024-11-27 17:36:46.167189] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:15:15.155 [2024-11-27 17:36:46.167228] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:15:15.155 [2024-11-27 17:36:46.167329] bdev_raid.c:3675:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev pt2 (4) greater than existing raid bdev raid_bdev1 (2) 00:15:15.155 [2024-11-27 17:36:46.167342] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:15:15.155 [2024-11-27 17:36:46.167374] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002000 name raid_bdev1, state configuring 00:15:15.155 [2024-11-27 17:36:46.167412] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:15:15.155 [2024-11-27 17:36:46.167501] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt3 is claimed 00:15:15.155 pt1 00:15:15.155 17:36:46 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:15.155 17:36:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@542 -- # '[' 4 -gt 2 ']' 00:15:15.155 17:36:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@545 -- # verify_raid_bdev_state raid_bdev1 configuring raid5f 64 3 00:15:15.155 17:36:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:15.155 17:36:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:15:15.155 17:36:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:15.155 17:36:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:15.155 17:36:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:15:15.155 17:36:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:15.155 17:36:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:15.155 17:36:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:15.155 17:36:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:15.155 17:36:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:15.155 17:36:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:15.155 17:36:46 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:15.155 17:36:46 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:15.155 17:36:46 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:15.155 17:36:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:15.155 "name": "raid_bdev1", 00:15:15.155 "uuid": "38dff1dc-17b5-4c86-ad2a-f6e7bbae90f6", 00:15:15.155 "strip_size_kb": 64, 00:15:15.155 "state": "configuring", 00:15:15.155 "raid_level": "raid5f", 00:15:15.155 "superblock": true, 00:15:15.155 "num_base_bdevs": 4, 00:15:15.155 "num_base_bdevs_discovered": 2, 00:15:15.155 "num_base_bdevs_operational": 3, 00:15:15.155 "base_bdevs_list": [ 00:15:15.155 { 00:15:15.155 "name": null, 00:15:15.155 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:15.155 "is_configured": false, 00:15:15.155 "data_offset": 2048, 00:15:15.155 "data_size": 63488 00:15:15.155 }, 00:15:15.155 { 00:15:15.155 "name": "pt2", 00:15:15.155 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:15.155 "is_configured": true, 00:15:15.155 "data_offset": 2048, 00:15:15.155 "data_size": 63488 00:15:15.155 }, 00:15:15.155 { 00:15:15.155 "name": "pt3", 00:15:15.155 "uuid": "00000000-0000-0000-0000-000000000003", 00:15:15.155 "is_configured": true, 00:15:15.155 "data_offset": 2048, 00:15:15.155 "data_size": 63488 00:15:15.155 }, 00:15:15.155 { 00:15:15.155 "name": null, 00:15:15.155 "uuid": "00000000-0000-0000-0000-000000000004", 00:15:15.155 "is_configured": false, 00:15:15.155 "data_offset": 2048, 00:15:15.155 "data_size": 63488 00:15:15.155 } 00:15:15.155 ] 00:15:15.155 }' 00:15:15.155 17:36:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:15.155 17:36:46 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:15.726 17:36:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@546 -- # rpc_cmd bdev_raid_get_bdevs configuring 00:15:15.726 17:36:46 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:15.726 17:36:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@546 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:15:15.726 17:36:46 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:15.726 17:36:46 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:15.726 17:36:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@546 -- # [[ false == \f\a\l\s\e ]] 00:15:15.726 17:36:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@549 -- # rpc_cmd bdev_passthru_create -b malloc4 -p pt4 -u 00000000-0000-0000-0000-000000000004 00:15:15.726 17:36:46 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:15.726 17:36:46 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:15.726 [2024-11-27 17:36:46.672031] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc4 00:15:15.726 [2024-11-27 17:36:46.672129] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:15.726 [2024-11-27 17:36:46.672172] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ba80 00:15:15.726 [2024-11-27 17:36:46.672202] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:15.726 [2024-11-27 17:36:46.672542] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:15.726 [2024-11-27 17:36:46.672604] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt4 00:15:15.726 [2024-11-27 17:36:46.672689] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt4 00:15:15.726 [2024-11-27 17:36:46.672749] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt4 is claimed 00:15:15.726 [2024-11-27 17:36:46.672871] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000002380 00:15:15.726 [2024-11-27 17:36:46.672909] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:15:15.726 [2024-11-27 17:36:46.673160] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000027a0 00:15:15.726 [2024-11-27 17:36:46.673707] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000002380 00:15:15.726 [2024-11-27 17:36:46.673756] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000002380 00:15:15.726 [2024-11-27 17:36:46.673962] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:15.726 pt4 00:15:15.726 17:36:46 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:15.726 17:36:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@554 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:15:15.726 17:36:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:15.726 17:36:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:15.726 17:36:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:15.726 17:36:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:15.726 17:36:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:15:15.726 17:36:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:15.726 17:36:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:15.726 17:36:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:15.726 17:36:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:15.726 17:36:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:15.726 17:36:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:15.726 17:36:46 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:15.726 17:36:46 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:15.726 17:36:46 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:15.726 17:36:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:15.726 "name": "raid_bdev1", 00:15:15.726 "uuid": "38dff1dc-17b5-4c86-ad2a-f6e7bbae90f6", 00:15:15.726 "strip_size_kb": 64, 00:15:15.726 "state": "online", 00:15:15.726 "raid_level": "raid5f", 00:15:15.726 "superblock": true, 00:15:15.726 "num_base_bdevs": 4, 00:15:15.726 "num_base_bdevs_discovered": 3, 00:15:15.726 "num_base_bdevs_operational": 3, 00:15:15.726 "base_bdevs_list": [ 00:15:15.726 { 00:15:15.726 "name": null, 00:15:15.726 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:15.726 "is_configured": false, 00:15:15.726 "data_offset": 2048, 00:15:15.726 "data_size": 63488 00:15:15.726 }, 00:15:15.726 { 00:15:15.726 "name": "pt2", 00:15:15.726 "uuid": "00000000-0000-0000-0000-000000000002", 00:15:15.726 "is_configured": true, 00:15:15.726 "data_offset": 2048, 00:15:15.726 "data_size": 63488 00:15:15.726 }, 00:15:15.726 { 00:15:15.726 "name": "pt3", 00:15:15.726 "uuid": "00000000-0000-0000-0000-000000000003", 00:15:15.726 "is_configured": true, 00:15:15.726 "data_offset": 2048, 00:15:15.726 "data_size": 63488 00:15:15.726 }, 00:15:15.726 { 00:15:15.726 "name": "pt4", 00:15:15.726 "uuid": "00000000-0000-0000-0000-000000000004", 00:15:15.726 "is_configured": true, 00:15:15.726 "data_offset": 2048, 00:15:15.726 "data_size": 63488 00:15:15.726 } 00:15:15.726 ] 00:15:15.726 }' 00:15:15.726 17:36:46 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:15.726 17:36:46 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:15.986 17:36:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@555 -- # rpc_cmd bdev_raid_get_bdevs online 00:15:15.986 17:36:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@555 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:15:15.986 17:36:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:15.986 17:36:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:15.986 17:36:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:15.986 17:36:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@555 -- # [[ false == \f\a\l\s\e ]] 00:15:15.986 17:36:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@558 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:15:15.986 17:36:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@558 -- # jq -r '.[] | .uuid' 00:15:15.986 17:36:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:15.986 17:36:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:16.247 [2024-11-27 17:36:47.179408] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:16.247 17:36:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:16.247 17:36:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@558 -- # '[' 38dff1dc-17b5-4c86-ad2a-f6e7bbae90f6 '!=' 38dff1dc-17b5-4c86-ad2a-f6e7bbae90f6 ']' 00:15:16.247 17:36:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@563 -- # killprocess 94404 00:15:16.247 17:36:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@950 -- # '[' -z 94404 ']' 00:15:16.247 17:36:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@954 -- # kill -0 94404 00:15:16.247 17:36:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@955 -- # uname 00:15:16.247 17:36:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:15:16.247 17:36:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 94404 00:15:16.247 17:36:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:15:16.247 17:36:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:15:16.247 17:36:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 94404' 00:15:16.247 killing process with pid 94404 00:15:16.247 17:36:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@969 -- # kill 94404 00:15:16.247 [2024-11-27 17:36:47.259291] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:15:16.247 [2024-11-27 17:36:47.259353] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:16.247 [2024-11-27 17:36:47.259415] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:16.247 [2024-11-27 17:36:47.259424] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002380 name raid_bdev1, state offline 00:15:16.247 17:36:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@974 -- # wait 94404 00:15:16.247 [2024-11-27 17:36:47.302493] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:15:16.507 17:36:47 bdev_raid.raid5f_superblock_test -- bdev/bdev_raid.sh@565 -- # return 0 00:15:16.507 ************************************ 00:15:16.507 END TEST raid5f_superblock_test 00:15:16.507 ************************************ 00:15:16.507 00:15:16.507 real 0m7.264s 00:15:16.507 user 0m12.277s 00:15:16.507 sys 0m1.582s 00:15:16.508 17:36:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:15:16.508 17:36:47 bdev_raid.raid5f_superblock_test -- common/autotest_common.sh@10 -- # set +x 00:15:16.508 17:36:47 bdev_raid -- bdev/bdev_raid.sh@989 -- # '[' true = true ']' 00:15:16.508 17:36:47 bdev_raid -- bdev/bdev_raid.sh@990 -- # run_test raid5f_rebuild_test raid_rebuild_test raid5f 4 false false true 00:15:16.508 17:36:47 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 7 -le 1 ']' 00:15:16.508 17:36:47 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:15:16.508 17:36:47 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:15:16.508 ************************************ 00:15:16.508 START TEST raid5f_rebuild_test 00:15:16.508 ************************************ 00:15:16.508 17:36:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@1125 -- # raid_rebuild_test raid5f 4 false false true 00:15:16.508 17:36:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@569 -- # local raid_level=raid5f 00:15:16.508 17:36:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=4 00:15:16.508 17:36:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@571 -- # local superblock=false 00:15:16.508 17:36:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:15:16.508 17:36:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@573 -- # local verify=true 00:15:16.508 17:36:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:15:16.508 17:36:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:15:16.508 17:36:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:15:16.508 17:36:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:15:16.508 17:36:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:15:16.508 17:36:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:15:16.508 17:36:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:15:16.508 17:36:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:15:16.508 17:36:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev3 00:15:16.508 17:36:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:15:16.508 17:36:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:15:16.508 17:36:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # echo BaseBdev4 00:15:16.508 17:36:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:15:16.508 17:36:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:15:16.508 17:36:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:15:16.508 17:36:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:15:16.508 17:36:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:15:16.508 17:36:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@576 -- # local strip_size 00:15:16.508 17:36:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@577 -- # local create_arg 00:15:16.508 17:36:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:15:16.508 17:36:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@579 -- # local data_offset 00:15:16.508 17:36:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@581 -- # '[' raid5f '!=' raid1 ']' 00:15:16.508 17:36:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@582 -- # '[' false = true ']' 00:15:16.508 17:36:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@586 -- # strip_size=64 00:15:16.508 17:36:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@587 -- # create_arg+=' -z 64' 00:15:16.508 17:36:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@592 -- # '[' false = true ']' 00:15:16.508 17:36:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@597 -- # raid_pid=94879 00:15:16.508 17:36:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:15:16.508 17:36:47 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@598 -- # waitforlisten 94879 00:15:16.508 17:36:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@831 -- # '[' -z 94879 ']' 00:15:16.508 17:36:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:15:16.508 17:36:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@836 -- # local max_retries=100 00:15:16.508 17:36:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:15:16.508 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:15:16.508 17:36:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@840 -- # xtrace_disable 00:15:16.508 17:36:47 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:16.768 [2024-11-27 17:36:47.728525] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:15:16.768 I/O size of 3145728 is greater than zero copy threshold (65536). 00:15:16.768 Zero copy mechanism will not be used. 00:15:16.768 [2024-11-27 17:36:47.728770] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid94879 ] 00:15:16.768 [2024-11-27 17:36:47.874803] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:15:16.768 [2024-11-27 17:36:47.919793] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:15:17.029 [2024-11-27 17:36:47.962441] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:17.029 [2024-11-27 17:36:47.962485] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:17.601 17:36:48 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:15:17.601 17:36:48 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@864 -- # return 0 00:15:17.601 17:36:48 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:15:17.601 17:36:48 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:15:17.601 17:36:48 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:17.601 17:36:48 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:17.601 BaseBdev1_malloc 00:15:17.601 17:36:48 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:17.601 17:36:48 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:15:17.601 17:36:48 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:17.601 17:36:48 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:17.601 [2024-11-27 17:36:48.564793] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:15:17.601 [2024-11-27 17:36:48.564857] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:17.601 [2024-11-27 17:36:48.564880] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:15:17.601 [2024-11-27 17:36:48.564894] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:17.601 [2024-11-27 17:36:48.566908] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:17.601 [2024-11-27 17:36:48.566947] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:15:17.601 BaseBdev1 00:15:17.601 17:36:48 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:17.601 17:36:48 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:15:17.601 17:36:48 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:15:17.601 17:36:48 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:17.601 17:36:48 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:17.601 BaseBdev2_malloc 00:15:17.601 17:36:48 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:17.601 17:36:48 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:15:17.601 17:36:48 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:17.601 17:36:48 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:17.601 [2024-11-27 17:36:48.606953] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:15:17.601 [2024-11-27 17:36:48.607072] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:17.601 [2024-11-27 17:36:48.607119] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:15:17.601 [2024-11-27 17:36:48.607170] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:17.601 [2024-11-27 17:36:48.610864] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:17.601 [2024-11-27 17:36:48.610918] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:15:17.601 BaseBdev2 00:15:17.601 17:36:48 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:17.601 17:36:48 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:15:17.601 17:36:48 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:15:17.601 17:36:48 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:17.601 17:36:48 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:17.601 BaseBdev3_malloc 00:15:17.601 17:36:48 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:17.601 17:36:48 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev3_malloc -p BaseBdev3 00:15:17.601 17:36:48 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:17.601 17:36:48 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:17.601 [2024-11-27 17:36:48.636304] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev3_malloc 00:15:17.601 [2024-11-27 17:36:48.636358] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:17.601 [2024-11-27 17:36:48.636384] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:15:17.601 [2024-11-27 17:36:48.636393] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:17.601 [2024-11-27 17:36:48.638300] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:17.601 [2024-11-27 17:36:48.638334] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:15:17.601 BaseBdev3 00:15:17.601 17:36:48 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:17.601 17:36:48 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:15:17.601 17:36:48 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:15:17.601 17:36:48 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:17.601 17:36:48 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:17.601 BaseBdev4_malloc 00:15:17.601 17:36:48 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:17.601 17:36:48 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev4_malloc -p BaseBdev4 00:15:17.601 17:36:48 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:17.601 17:36:48 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:17.601 [2024-11-27 17:36:48.664720] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev4_malloc 00:15:17.601 [2024-11-27 17:36:48.664770] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:17.602 [2024-11-27 17:36:48.664790] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:15:17.602 [2024-11-27 17:36:48.664799] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:17.602 [2024-11-27 17:36:48.666716] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:17.602 [2024-11-27 17:36:48.666753] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:15:17.602 BaseBdev4 00:15:17.602 17:36:48 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:17.602 17:36:48 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:15:17.602 17:36:48 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:17.602 17:36:48 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:17.602 spare_malloc 00:15:17.602 17:36:48 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:17.602 17:36:48 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:15:17.602 17:36:48 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:17.602 17:36:48 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:17.602 spare_delay 00:15:17.602 17:36:48 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:17.602 17:36:48 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:15:17.602 17:36:48 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:17.602 17:36:48 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:17.602 [2024-11-27 17:36:48.705036] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:15:17.602 [2024-11-27 17:36:48.705155] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:17.602 [2024-11-27 17:36:48.705177] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009c80 00:15:17.602 [2024-11-27 17:36:48.705185] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:17.602 [2024-11-27 17:36:48.707124] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:17.602 [2024-11-27 17:36:48.707164] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:15:17.602 spare 00:15:17.602 17:36:48 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:17.602 17:36:48 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -z 64 -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 00:15:17.602 17:36:48 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:17.602 17:36:48 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:17.602 [2024-11-27 17:36:48.717082] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:15:17.602 [2024-11-27 17:36:48.718795] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:15:17.602 [2024-11-27 17:36:48.718854] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:15:17.602 [2024-11-27 17:36:48.718900] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:15:17.602 [2024-11-27 17:36:48.718978] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:15:17.602 [2024-11-27 17:36:48.718986] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 196608, blocklen 512 00:15:17.602 [2024-11-27 17:36:48.719271] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:15:17.602 [2024-11-27 17:36:48.719703] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:15:17.602 [2024-11-27 17:36:48.719726] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:15:17.602 [2024-11-27 17:36:48.719831] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:17.602 17:36:48 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:17.602 17:36:48 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 4 00:15:17.602 17:36:48 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:17.602 17:36:48 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:17.602 17:36:48 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:17.602 17:36:48 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:17.602 17:36:48 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:17.602 17:36:48 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:17.602 17:36:48 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:17.602 17:36:48 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:17.602 17:36:48 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:17.602 17:36:48 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:17.602 17:36:48 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:17.602 17:36:48 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:17.602 17:36:48 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:17.602 17:36:48 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:17.602 17:36:48 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:17.602 "name": "raid_bdev1", 00:15:17.602 "uuid": "9e3909f6-4d74-482f-a312-6db1a0a9be25", 00:15:17.602 "strip_size_kb": 64, 00:15:17.602 "state": "online", 00:15:17.602 "raid_level": "raid5f", 00:15:17.602 "superblock": false, 00:15:17.602 "num_base_bdevs": 4, 00:15:17.602 "num_base_bdevs_discovered": 4, 00:15:17.602 "num_base_bdevs_operational": 4, 00:15:17.602 "base_bdevs_list": [ 00:15:17.602 { 00:15:17.602 "name": "BaseBdev1", 00:15:17.602 "uuid": "0f62f81f-dbf2-5f25-8eeb-443c7b01964f", 00:15:17.602 "is_configured": true, 00:15:17.602 "data_offset": 0, 00:15:17.602 "data_size": 65536 00:15:17.602 }, 00:15:17.602 { 00:15:17.602 "name": "BaseBdev2", 00:15:17.602 "uuid": "7fa1e11b-e649-5f82-a5e6-6eceb73f5b9a", 00:15:17.602 "is_configured": true, 00:15:17.602 "data_offset": 0, 00:15:17.602 "data_size": 65536 00:15:17.602 }, 00:15:17.602 { 00:15:17.602 "name": "BaseBdev3", 00:15:17.602 "uuid": "366ce04e-459b-5530-9516-bd873875900e", 00:15:17.602 "is_configured": true, 00:15:17.602 "data_offset": 0, 00:15:17.602 "data_size": 65536 00:15:17.602 }, 00:15:17.602 { 00:15:17.602 "name": "BaseBdev4", 00:15:17.602 "uuid": "49fb52a5-1369-5e2a-8db8-85f6be1c5554", 00:15:17.602 "is_configured": true, 00:15:17.602 "data_offset": 0, 00:15:17.602 "data_size": 65536 00:15:17.602 } 00:15:17.602 ] 00:15:17.602 }' 00:15:17.602 17:36:48 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:17.602 17:36:48 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:18.178 17:36:49 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:15:18.178 17:36:49 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:15:18.178 17:36:49 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:18.178 17:36:49 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:18.178 [2024-11-27 17:36:49.160993] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:18.178 17:36:49 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:18.178 17:36:49 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=196608 00:15:18.178 17:36:49 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:15:18.178 17:36:49 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:18.178 17:36:49 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:18.178 17:36:49 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:18.178 17:36:49 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:18.178 17:36:49 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@619 -- # data_offset=0 00:15:18.178 17:36:49 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:15:18.178 17:36:49 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:15:18.178 17:36:49 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:15:18.178 17:36:49 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:15:18.178 17:36:49 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:15:18.178 17:36:49 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:15:18.178 17:36:49 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@10 -- # local bdev_list 00:15:18.178 17:36:49 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:15:18.178 17:36:49 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@11 -- # local nbd_list 00:15:18.178 17:36:49 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@12 -- # local i 00:15:18.178 17:36:49 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:15:18.178 17:36:49 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:15:18.178 17:36:49 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:15:18.438 [2024-11-27 17:36:49.404460] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000027a0 00:15:18.438 /dev/nbd0 00:15:18.438 17:36:49 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:15:18.438 17:36:49 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:15:18.438 17:36:49 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:15:18.438 17:36:49 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@869 -- # local i 00:15:18.438 17:36:49 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:15:18.438 17:36:49 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:15:18.438 17:36:49 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:15:18.438 17:36:49 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@873 -- # break 00:15:18.438 17:36:49 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:15:18.438 17:36:49 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:15:18.438 17:36:49 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:15:18.438 1+0 records in 00:15:18.438 1+0 records out 00:15:18.438 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000526351 s, 7.8 MB/s 00:15:18.438 17:36:49 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:18.438 17:36:49 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@886 -- # size=4096 00:15:18.438 17:36:49 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:18.438 17:36:49 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:15:18.438 17:36:49 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@889 -- # return 0 00:15:18.439 17:36:49 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:15:18.439 17:36:49 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:15:18.439 17:36:49 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@629 -- # '[' raid5f = raid5f ']' 00:15:18.439 17:36:49 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@630 -- # write_unit_size=384 00:15:18.439 17:36:49 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@631 -- # echo 192 00:15:18.439 17:36:49 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=196608 count=512 oflag=direct 00:15:19.012 512+0 records in 00:15:19.012 512+0 records out 00:15:19.012 100663296 bytes (101 MB, 96 MiB) copied, 0.513093 s, 196 MB/s 00:15:19.012 17:36:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:15:19.012 17:36:50 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:15:19.012 17:36:50 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:15:19.012 17:36:50 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@50 -- # local nbd_list 00:15:19.012 17:36:50 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@51 -- # local i 00:15:19.012 17:36:50 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:15:19.012 17:36:50 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:15:19.272 17:36:50 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:15:19.272 [2024-11-27 17:36:50.209695] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:19.272 17:36:50 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:15:19.272 17:36:50 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:15:19.272 17:36:50 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:15:19.272 17:36:50 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:15:19.272 17:36:50 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:15:19.272 17:36:50 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:15:19.272 17:36:50 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:15:19.272 17:36:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:15:19.272 17:36:50 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:19.272 17:36:50 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:19.272 [2024-11-27 17:36:50.225724] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:15:19.272 17:36:50 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:19.272 17:36:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:15:19.272 17:36:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:19.272 17:36:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:19.272 17:36:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:19.272 17:36:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:19.272 17:36:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:15:19.272 17:36:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:19.272 17:36:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:19.272 17:36:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:19.272 17:36:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:19.272 17:36:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:19.272 17:36:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:19.272 17:36:50 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:19.272 17:36:50 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:19.272 17:36:50 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:19.272 17:36:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:19.272 "name": "raid_bdev1", 00:15:19.272 "uuid": "9e3909f6-4d74-482f-a312-6db1a0a9be25", 00:15:19.272 "strip_size_kb": 64, 00:15:19.272 "state": "online", 00:15:19.272 "raid_level": "raid5f", 00:15:19.272 "superblock": false, 00:15:19.272 "num_base_bdevs": 4, 00:15:19.272 "num_base_bdevs_discovered": 3, 00:15:19.272 "num_base_bdevs_operational": 3, 00:15:19.272 "base_bdevs_list": [ 00:15:19.272 { 00:15:19.272 "name": null, 00:15:19.272 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:19.272 "is_configured": false, 00:15:19.272 "data_offset": 0, 00:15:19.272 "data_size": 65536 00:15:19.272 }, 00:15:19.273 { 00:15:19.273 "name": "BaseBdev2", 00:15:19.273 "uuid": "7fa1e11b-e649-5f82-a5e6-6eceb73f5b9a", 00:15:19.273 "is_configured": true, 00:15:19.273 "data_offset": 0, 00:15:19.273 "data_size": 65536 00:15:19.273 }, 00:15:19.273 { 00:15:19.273 "name": "BaseBdev3", 00:15:19.273 "uuid": "366ce04e-459b-5530-9516-bd873875900e", 00:15:19.273 "is_configured": true, 00:15:19.273 "data_offset": 0, 00:15:19.273 "data_size": 65536 00:15:19.273 }, 00:15:19.273 { 00:15:19.273 "name": "BaseBdev4", 00:15:19.273 "uuid": "49fb52a5-1369-5e2a-8db8-85f6be1c5554", 00:15:19.273 "is_configured": true, 00:15:19.273 "data_offset": 0, 00:15:19.273 "data_size": 65536 00:15:19.273 } 00:15:19.273 ] 00:15:19.273 }' 00:15:19.273 17:36:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:19.273 17:36:50 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:19.532 17:36:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:15:19.532 17:36:50 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:19.532 17:36:50 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:19.532 [2024-11-27 17:36:50.696936] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:19.532 [2024-11-27 17:36:50.700267] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000027da0 00:15:19.532 [2024-11-27 17:36:50.702347] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:15:19.532 17:36:50 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:19.532 17:36:50 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@647 -- # sleep 1 00:15:20.911 17:36:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:20.911 17:36:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:20.911 17:36:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:20.911 17:36:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:20.911 17:36:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:20.911 17:36:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:20.911 17:36:51 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:20.911 17:36:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:20.911 17:36:51 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:20.911 17:36:51 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:20.911 17:36:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:20.911 "name": "raid_bdev1", 00:15:20.911 "uuid": "9e3909f6-4d74-482f-a312-6db1a0a9be25", 00:15:20.911 "strip_size_kb": 64, 00:15:20.911 "state": "online", 00:15:20.911 "raid_level": "raid5f", 00:15:20.911 "superblock": false, 00:15:20.911 "num_base_bdevs": 4, 00:15:20.911 "num_base_bdevs_discovered": 4, 00:15:20.911 "num_base_bdevs_operational": 4, 00:15:20.911 "process": { 00:15:20.911 "type": "rebuild", 00:15:20.911 "target": "spare", 00:15:20.911 "progress": { 00:15:20.911 "blocks": 19200, 00:15:20.912 "percent": 9 00:15:20.912 } 00:15:20.912 }, 00:15:20.912 "base_bdevs_list": [ 00:15:20.912 { 00:15:20.912 "name": "spare", 00:15:20.912 "uuid": "499e8caa-bfa4-524f-958e-5b0bcc73cb3b", 00:15:20.912 "is_configured": true, 00:15:20.912 "data_offset": 0, 00:15:20.912 "data_size": 65536 00:15:20.912 }, 00:15:20.912 { 00:15:20.912 "name": "BaseBdev2", 00:15:20.912 "uuid": "7fa1e11b-e649-5f82-a5e6-6eceb73f5b9a", 00:15:20.912 "is_configured": true, 00:15:20.912 "data_offset": 0, 00:15:20.912 "data_size": 65536 00:15:20.912 }, 00:15:20.912 { 00:15:20.912 "name": "BaseBdev3", 00:15:20.912 "uuid": "366ce04e-459b-5530-9516-bd873875900e", 00:15:20.912 "is_configured": true, 00:15:20.912 "data_offset": 0, 00:15:20.912 "data_size": 65536 00:15:20.912 }, 00:15:20.912 { 00:15:20.912 "name": "BaseBdev4", 00:15:20.912 "uuid": "49fb52a5-1369-5e2a-8db8-85f6be1c5554", 00:15:20.912 "is_configured": true, 00:15:20.912 "data_offset": 0, 00:15:20.912 "data_size": 65536 00:15:20.912 } 00:15:20.912 ] 00:15:20.912 }' 00:15:20.912 17:36:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:20.912 17:36:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:20.912 17:36:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:20.912 17:36:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:20.912 17:36:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:15:20.912 17:36:51 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:20.912 17:36:51 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:20.912 [2024-11-27 17:36:51.852982] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:20.912 [2024-11-27 17:36:51.907643] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:15:20.912 [2024-11-27 17:36:51.907764] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:20.912 [2024-11-27 17:36:51.907787] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:20.912 [2024-11-27 17:36:51.907795] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:15:20.912 17:36:51 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:20.912 17:36:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:15:20.912 17:36:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:20.912 17:36:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:20.912 17:36:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:20.912 17:36:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:20.912 17:36:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:15:20.912 17:36:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:20.912 17:36:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:20.912 17:36:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:20.912 17:36:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:20.912 17:36:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:20.912 17:36:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:20.912 17:36:51 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:20.912 17:36:51 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:20.912 17:36:51 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:20.912 17:36:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:20.912 "name": "raid_bdev1", 00:15:20.912 "uuid": "9e3909f6-4d74-482f-a312-6db1a0a9be25", 00:15:20.912 "strip_size_kb": 64, 00:15:20.912 "state": "online", 00:15:20.912 "raid_level": "raid5f", 00:15:20.912 "superblock": false, 00:15:20.912 "num_base_bdevs": 4, 00:15:20.912 "num_base_bdevs_discovered": 3, 00:15:20.912 "num_base_bdevs_operational": 3, 00:15:20.912 "base_bdevs_list": [ 00:15:20.912 { 00:15:20.912 "name": null, 00:15:20.912 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:20.912 "is_configured": false, 00:15:20.912 "data_offset": 0, 00:15:20.912 "data_size": 65536 00:15:20.912 }, 00:15:20.912 { 00:15:20.912 "name": "BaseBdev2", 00:15:20.912 "uuid": "7fa1e11b-e649-5f82-a5e6-6eceb73f5b9a", 00:15:20.912 "is_configured": true, 00:15:20.912 "data_offset": 0, 00:15:20.912 "data_size": 65536 00:15:20.912 }, 00:15:20.912 { 00:15:20.912 "name": "BaseBdev3", 00:15:20.912 "uuid": "366ce04e-459b-5530-9516-bd873875900e", 00:15:20.912 "is_configured": true, 00:15:20.912 "data_offset": 0, 00:15:20.912 "data_size": 65536 00:15:20.912 }, 00:15:20.912 { 00:15:20.912 "name": "BaseBdev4", 00:15:20.912 "uuid": "49fb52a5-1369-5e2a-8db8-85f6be1c5554", 00:15:20.912 "is_configured": true, 00:15:20.912 "data_offset": 0, 00:15:20.912 "data_size": 65536 00:15:20.912 } 00:15:20.912 ] 00:15:20.912 }' 00:15:20.912 17:36:51 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:20.912 17:36:51 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:21.480 17:36:52 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:21.480 17:36:52 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:21.480 17:36:52 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:21.480 17:36:52 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:21.480 17:36:52 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:21.480 17:36:52 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:21.480 17:36:52 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:21.480 17:36:52 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:21.480 17:36:52 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:21.480 17:36:52 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:21.480 17:36:52 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:21.480 "name": "raid_bdev1", 00:15:21.480 "uuid": "9e3909f6-4d74-482f-a312-6db1a0a9be25", 00:15:21.480 "strip_size_kb": 64, 00:15:21.480 "state": "online", 00:15:21.480 "raid_level": "raid5f", 00:15:21.480 "superblock": false, 00:15:21.480 "num_base_bdevs": 4, 00:15:21.480 "num_base_bdevs_discovered": 3, 00:15:21.480 "num_base_bdevs_operational": 3, 00:15:21.480 "base_bdevs_list": [ 00:15:21.480 { 00:15:21.480 "name": null, 00:15:21.480 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:21.480 "is_configured": false, 00:15:21.480 "data_offset": 0, 00:15:21.480 "data_size": 65536 00:15:21.480 }, 00:15:21.480 { 00:15:21.480 "name": "BaseBdev2", 00:15:21.480 "uuid": "7fa1e11b-e649-5f82-a5e6-6eceb73f5b9a", 00:15:21.480 "is_configured": true, 00:15:21.480 "data_offset": 0, 00:15:21.480 "data_size": 65536 00:15:21.480 }, 00:15:21.480 { 00:15:21.480 "name": "BaseBdev3", 00:15:21.480 "uuid": "366ce04e-459b-5530-9516-bd873875900e", 00:15:21.480 "is_configured": true, 00:15:21.480 "data_offset": 0, 00:15:21.480 "data_size": 65536 00:15:21.480 }, 00:15:21.480 { 00:15:21.480 "name": "BaseBdev4", 00:15:21.480 "uuid": "49fb52a5-1369-5e2a-8db8-85f6be1c5554", 00:15:21.480 "is_configured": true, 00:15:21.480 "data_offset": 0, 00:15:21.480 "data_size": 65536 00:15:21.480 } 00:15:21.480 ] 00:15:21.480 }' 00:15:21.480 17:36:52 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:21.480 17:36:52 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:21.480 17:36:52 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:21.480 17:36:52 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:15:21.480 17:36:52 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:15:21.480 17:36:52 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:21.480 17:36:52 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:21.480 [2024-11-27 17:36:52.539898] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:21.481 [2024-11-27 17:36:52.542925] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000027e70 00:15:21.481 [2024-11-27 17:36:52.545035] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:15:21.481 17:36:52 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:21.481 17:36:52 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@663 -- # sleep 1 00:15:22.418 17:36:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:22.418 17:36:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:22.418 17:36:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:22.418 17:36:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:22.418 17:36:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:22.418 17:36:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:22.418 17:36:53 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:22.418 17:36:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:22.418 17:36:53 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:22.418 17:36:53 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:22.418 17:36:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:22.418 "name": "raid_bdev1", 00:15:22.418 "uuid": "9e3909f6-4d74-482f-a312-6db1a0a9be25", 00:15:22.418 "strip_size_kb": 64, 00:15:22.418 "state": "online", 00:15:22.418 "raid_level": "raid5f", 00:15:22.418 "superblock": false, 00:15:22.418 "num_base_bdevs": 4, 00:15:22.418 "num_base_bdevs_discovered": 4, 00:15:22.418 "num_base_bdevs_operational": 4, 00:15:22.418 "process": { 00:15:22.418 "type": "rebuild", 00:15:22.418 "target": "spare", 00:15:22.418 "progress": { 00:15:22.418 "blocks": 19200, 00:15:22.418 "percent": 9 00:15:22.418 } 00:15:22.418 }, 00:15:22.418 "base_bdevs_list": [ 00:15:22.418 { 00:15:22.418 "name": "spare", 00:15:22.418 "uuid": "499e8caa-bfa4-524f-958e-5b0bcc73cb3b", 00:15:22.418 "is_configured": true, 00:15:22.418 "data_offset": 0, 00:15:22.418 "data_size": 65536 00:15:22.418 }, 00:15:22.418 { 00:15:22.418 "name": "BaseBdev2", 00:15:22.418 "uuid": "7fa1e11b-e649-5f82-a5e6-6eceb73f5b9a", 00:15:22.418 "is_configured": true, 00:15:22.418 "data_offset": 0, 00:15:22.418 "data_size": 65536 00:15:22.418 }, 00:15:22.418 { 00:15:22.418 "name": "BaseBdev3", 00:15:22.418 "uuid": "366ce04e-459b-5530-9516-bd873875900e", 00:15:22.418 "is_configured": true, 00:15:22.418 "data_offset": 0, 00:15:22.418 "data_size": 65536 00:15:22.418 }, 00:15:22.418 { 00:15:22.418 "name": "BaseBdev4", 00:15:22.418 "uuid": "49fb52a5-1369-5e2a-8db8-85f6be1c5554", 00:15:22.418 "is_configured": true, 00:15:22.418 "data_offset": 0, 00:15:22.418 "data_size": 65536 00:15:22.418 } 00:15:22.418 ] 00:15:22.418 }' 00:15:22.418 17:36:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:22.678 17:36:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:22.678 17:36:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:22.678 17:36:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:22.678 17:36:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@666 -- # '[' false = true ']' 00:15:22.678 17:36:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=4 00:15:22.678 17:36:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@693 -- # '[' raid5f = raid1 ']' 00:15:22.678 17:36:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@706 -- # local timeout=521 00:15:22.678 17:36:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:22.678 17:36:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:22.678 17:36:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:22.678 17:36:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:22.678 17:36:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:22.678 17:36:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:22.678 17:36:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:22.678 17:36:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:22.678 17:36:53 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:22.678 17:36:53 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:22.678 17:36:53 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:22.678 17:36:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:22.678 "name": "raid_bdev1", 00:15:22.678 "uuid": "9e3909f6-4d74-482f-a312-6db1a0a9be25", 00:15:22.678 "strip_size_kb": 64, 00:15:22.678 "state": "online", 00:15:22.678 "raid_level": "raid5f", 00:15:22.678 "superblock": false, 00:15:22.678 "num_base_bdevs": 4, 00:15:22.678 "num_base_bdevs_discovered": 4, 00:15:22.678 "num_base_bdevs_operational": 4, 00:15:22.678 "process": { 00:15:22.679 "type": "rebuild", 00:15:22.679 "target": "spare", 00:15:22.679 "progress": { 00:15:22.679 "blocks": 21120, 00:15:22.679 "percent": 10 00:15:22.679 } 00:15:22.679 }, 00:15:22.679 "base_bdevs_list": [ 00:15:22.679 { 00:15:22.679 "name": "spare", 00:15:22.679 "uuid": "499e8caa-bfa4-524f-958e-5b0bcc73cb3b", 00:15:22.679 "is_configured": true, 00:15:22.679 "data_offset": 0, 00:15:22.679 "data_size": 65536 00:15:22.679 }, 00:15:22.679 { 00:15:22.679 "name": "BaseBdev2", 00:15:22.679 "uuid": "7fa1e11b-e649-5f82-a5e6-6eceb73f5b9a", 00:15:22.679 "is_configured": true, 00:15:22.679 "data_offset": 0, 00:15:22.679 "data_size": 65536 00:15:22.679 }, 00:15:22.679 { 00:15:22.679 "name": "BaseBdev3", 00:15:22.679 "uuid": "366ce04e-459b-5530-9516-bd873875900e", 00:15:22.679 "is_configured": true, 00:15:22.679 "data_offset": 0, 00:15:22.679 "data_size": 65536 00:15:22.679 }, 00:15:22.679 { 00:15:22.679 "name": "BaseBdev4", 00:15:22.679 "uuid": "49fb52a5-1369-5e2a-8db8-85f6be1c5554", 00:15:22.679 "is_configured": true, 00:15:22.679 "data_offset": 0, 00:15:22.679 "data_size": 65536 00:15:22.679 } 00:15:22.679 ] 00:15:22.679 }' 00:15:22.679 17:36:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:22.679 17:36:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:22.679 17:36:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:22.679 17:36:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:22.679 17:36:53 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:24.069 17:36:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:24.069 17:36:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:24.069 17:36:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:24.069 17:36:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:24.069 17:36:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:24.069 17:36:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:24.069 17:36:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:24.069 17:36:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:24.069 17:36:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:24.069 17:36:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:24.069 17:36:54 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:24.069 17:36:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:24.069 "name": "raid_bdev1", 00:15:24.069 "uuid": "9e3909f6-4d74-482f-a312-6db1a0a9be25", 00:15:24.069 "strip_size_kb": 64, 00:15:24.069 "state": "online", 00:15:24.069 "raid_level": "raid5f", 00:15:24.069 "superblock": false, 00:15:24.069 "num_base_bdevs": 4, 00:15:24.069 "num_base_bdevs_discovered": 4, 00:15:24.069 "num_base_bdevs_operational": 4, 00:15:24.069 "process": { 00:15:24.069 "type": "rebuild", 00:15:24.069 "target": "spare", 00:15:24.069 "progress": { 00:15:24.069 "blocks": 42240, 00:15:24.069 "percent": 21 00:15:24.069 } 00:15:24.069 }, 00:15:24.069 "base_bdevs_list": [ 00:15:24.069 { 00:15:24.069 "name": "spare", 00:15:24.069 "uuid": "499e8caa-bfa4-524f-958e-5b0bcc73cb3b", 00:15:24.069 "is_configured": true, 00:15:24.069 "data_offset": 0, 00:15:24.069 "data_size": 65536 00:15:24.069 }, 00:15:24.069 { 00:15:24.069 "name": "BaseBdev2", 00:15:24.069 "uuid": "7fa1e11b-e649-5f82-a5e6-6eceb73f5b9a", 00:15:24.069 "is_configured": true, 00:15:24.069 "data_offset": 0, 00:15:24.069 "data_size": 65536 00:15:24.069 }, 00:15:24.069 { 00:15:24.069 "name": "BaseBdev3", 00:15:24.069 "uuid": "366ce04e-459b-5530-9516-bd873875900e", 00:15:24.069 "is_configured": true, 00:15:24.069 "data_offset": 0, 00:15:24.069 "data_size": 65536 00:15:24.069 }, 00:15:24.069 { 00:15:24.069 "name": "BaseBdev4", 00:15:24.069 "uuid": "49fb52a5-1369-5e2a-8db8-85f6be1c5554", 00:15:24.069 "is_configured": true, 00:15:24.069 "data_offset": 0, 00:15:24.069 "data_size": 65536 00:15:24.069 } 00:15:24.069 ] 00:15:24.069 }' 00:15:24.069 17:36:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:24.069 17:36:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:24.069 17:36:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:24.069 17:36:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:24.069 17:36:54 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:25.036 17:36:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:25.036 17:36:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:25.036 17:36:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:25.036 17:36:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:25.036 17:36:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:25.036 17:36:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:25.036 17:36:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:25.036 17:36:55 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:25.036 17:36:55 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:25.036 17:36:55 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:25.036 17:36:56 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:25.036 17:36:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:25.036 "name": "raid_bdev1", 00:15:25.036 "uuid": "9e3909f6-4d74-482f-a312-6db1a0a9be25", 00:15:25.036 "strip_size_kb": 64, 00:15:25.036 "state": "online", 00:15:25.036 "raid_level": "raid5f", 00:15:25.036 "superblock": false, 00:15:25.036 "num_base_bdevs": 4, 00:15:25.036 "num_base_bdevs_discovered": 4, 00:15:25.036 "num_base_bdevs_operational": 4, 00:15:25.036 "process": { 00:15:25.036 "type": "rebuild", 00:15:25.036 "target": "spare", 00:15:25.036 "progress": { 00:15:25.036 "blocks": 65280, 00:15:25.036 "percent": 33 00:15:25.036 } 00:15:25.036 }, 00:15:25.036 "base_bdevs_list": [ 00:15:25.036 { 00:15:25.036 "name": "spare", 00:15:25.036 "uuid": "499e8caa-bfa4-524f-958e-5b0bcc73cb3b", 00:15:25.036 "is_configured": true, 00:15:25.036 "data_offset": 0, 00:15:25.036 "data_size": 65536 00:15:25.036 }, 00:15:25.036 { 00:15:25.036 "name": "BaseBdev2", 00:15:25.036 "uuid": "7fa1e11b-e649-5f82-a5e6-6eceb73f5b9a", 00:15:25.036 "is_configured": true, 00:15:25.036 "data_offset": 0, 00:15:25.036 "data_size": 65536 00:15:25.036 }, 00:15:25.036 { 00:15:25.036 "name": "BaseBdev3", 00:15:25.036 "uuid": "366ce04e-459b-5530-9516-bd873875900e", 00:15:25.036 "is_configured": true, 00:15:25.036 "data_offset": 0, 00:15:25.036 "data_size": 65536 00:15:25.036 }, 00:15:25.036 { 00:15:25.036 "name": "BaseBdev4", 00:15:25.036 "uuid": "49fb52a5-1369-5e2a-8db8-85f6be1c5554", 00:15:25.036 "is_configured": true, 00:15:25.036 "data_offset": 0, 00:15:25.036 "data_size": 65536 00:15:25.036 } 00:15:25.036 ] 00:15:25.036 }' 00:15:25.036 17:36:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:25.036 17:36:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:25.036 17:36:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:25.036 17:36:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:25.036 17:36:56 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:25.976 17:36:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:25.976 17:36:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:25.976 17:36:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:25.976 17:36:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:25.976 17:36:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:25.976 17:36:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:25.976 17:36:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:25.976 17:36:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:25.976 17:36:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:25.976 17:36:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:25.976 17:36:57 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:26.236 17:36:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:26.236 "name": "raid_bdev1", 00:15:26.236 "uuid": "9e3909f6-4d74-482f-a312-6db1a0a9be25", 00:15:26.236 "strip_size_kb": 64, 00:15:26.236 "state": "online", 00:15:26.236 "raid_level": "raid5f", 00:15:26.236 "superblock": false, 00:15:26.236 "num_base_bdevs": 4, 00:15:26.236 "num_base_bdevs_discovered": 4, 00:15:26.236 "num_base_bdevs_operational": 4, 00:15:26.236 "process": { 00:15:26.236 "type": "rebuild", 00:15:26.236 "target": "spare", 00:15:26.236 "progress": { 00:15:26.236 "blocks": 86400, 00:15:26.236 "percent": 43 00:15:26.236 } 00:15:26.236 }, 00:15:26.236 "base_bdevs_list": [ 00:15:26.236 { 00:15:26.236 "name": "spare", 00:15:26.236 "uuid": "499e8caa-bfa4-524f-958e-5b0bcc73cb3b", 00:15:26.236 "is_configured": true, 00:15:26.236 "data_offset": 0, 00:15:26.236 "data_size": 65536 00:15:26.236 }, 00:15:26.236 { 00:15:26.236 "name": "BaseBdev2", 00:15:26.236 "uuid": "7fa1e11b-e649-5f82-a5e6-6eceb73f5b9a", 00:15:26.236 "is_configured": true, 00:15:26.236 "data_offset": 0, 00:15:26.236 "data_size": 65536 00:15:26.236 }, 00:15:26.236 { 00:15:26.236 "name": "BaseBdev3", 00:15:26.236 "uuid": "366ce04e-459b-5530-9516-bd873875900e", 00:15:26.236 "is_configured": true, 00:15:26.236 "data_offset": 0, 00:15:26.236 "data_size": 65536 00:15:26.236 }, 00:15:26.236 { 00:15:26.236 "name": "BaseBdev4", 00:15:26.236 "uuid": "49fb52a5-1369-5e2a-8db8-85f6be1c5554", 00:15:26.236 "is_configured": true, 00:15:26.236 "data_offset": 0, 00:15:26.236 "data_size": 65536 00:15:26.236 } 00:15:26.236 ] 00:15:26.236 }' 00:15:26.236 17:36:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:26.236 17:36:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:26.236 17:36:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:26.236 17:36:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:26.236 17:36:57 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:27.176 17:36:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:27.176 17:36:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:27.176 17:36:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:27.176 17:36:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:27.176 17:36:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:27.176 17:36:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:27.176 17:36:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:27.176 17:36:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:27.176 17:36:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:27.176 17:36:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:27.176 17:36:58 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:27.176 17:36:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:27.176 "name": "raid_bdev1", 00:15:27.176 "uuid": "9e3909f6-4d74-482f-a312-6db1a0a9be25", 00:15:27.176 "strip_size_kb": 64, 00:15:27.176 "state": "online", 00:15:27.176 "raid_level": "raid5f", 00:15:27.176 "superblock": false, 00:15:27.176 "num_base_bdevs": 4, 00:15:27.176 "num_base_bdevs_discovered": 4, 00:15:27.176 "num_base_bdevs_operational": 4, 00:15:27.176 "process": { 00:15:27.176 "type": "rebuild", 00:15:27.176 "target": "spare", 00:15:27.176 "progress": { 00:15:27.176 "blocks": 109440, 00:15:27.176 "percent": 55 00:15:27.176 } 00:15:27.176 }, 00:15:27.176 "base_bdevs_list": [ 00:15:27.176 { 00:15:27.176 "name": "spare", 00:15:27.176 "uuid": "499e8caa-bfa4-524f-958e-5b0bcc73cb3b", 00:15:27.176 "is_configured": true, 00:15:27.176 "data_offset": 0, 00:15:27.176 "data_size": 65536 00:15:27.176 }, 00:15:27.176 { 00:15:27.176 "name": "BaseBdev2", 00:15:27.176 "uuid": "7fa1e11b-e649-5f82-a5e6-6eceb73f5b9a", 00:15:27.176 "is_configured": true, 00:15:27.176 "data_offset": 0, 00:15:27.176 "data_size": 65536 00:15:27.176 }, 00:15:27.176 { 00:15:27.176 "name": "BaseBdev3", 00:15:27.176 "uuid": "366ce04e-459b-5530-9516-bd873875900e", 00:15:27.176 "is_configured": true, 00:15:27.176 "data_offset": 0, 00:15:27.176 "data_size": 65536 00:15:27.176 }, 00:15:27.176 { 00:15:27.176 "name": "BaseBdev4", 00:15:27.176 "uuid": "49fb52a5-1369-5e2a-8db8-85f6be1c5554", 00:15:27.176 "is_configured": true, 00:15:27.176 "data_offset": 0, 00:15:27.176 "data_size": 65536 00:15:27.176 } 00:15:27.176 ] 00:15:27.176 }' 00:15:27.176 17:36:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:27.436 17:36:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:27.436 17:36:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:27.436 17:36:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:27.436 17:36:58 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:28.375 17:36:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:28.375 17:36:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:28.375 17:36:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:28.375 17:36:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:28.375 17:36:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:28.375 17:36:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:28.375 17:36:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:28.375 17:36:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:28.375 17:36:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:28.375 17:36:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:28.375 17:36:59 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:28.375 17:36:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:28.375 "name": "raid_bdev1", 00:15:28.375 "uuid": "9e3909f6-4d74-482f-a312-6db1a0a9be25", 00:15:28.375 "strip_size_kb": 64, 00:15:28.375 "state": "online", 00:15:28.375 "raid_level": "raid5f", 00:15:28.375 "superblock": false, 00:15:28.375 "num_base_bdevs": 4, 00:15:28.375 "num_base_bdevs_discovered": 4, 00:15:28.375 "num_base_bdevs_operational": 4, 00:15:28.375 "process": { 00:15:28.375 "type": "rebuild", 00:15:28.375 "target": "spare", 00:15:28.375 "progress": { 00:15:28.375 "blocks": 130560, 00:15:28.375 "percent": 66 00:15:28.375 } 00:15:28.376 }, 00:15:28.376 "base_bdevs_list": [ 00:15:28.376 { 00:15:28.376 "name": "spare", 00:15:28.376 "uuid": "499e8caa-bfa4-524f-958e-5b0bcc73cb3b", 00:15:28.376 "is_configured": true, 00:15:28.376 "data_offset": 0, 00:15:28.376 "data_size": 65536 00:15:28.376 }, 00:15:28.376 { 00:15:28.376 "name": "BaseBdev2", 00:15:28.376 "uuid": "7fa1e11b-e649-5f82-a5e6-6eceb73f5b9a", 00:15:28.376 "is_configured": true, 00:15:28.376 "data_offset": 0, 00:15:28.376 "data_size": 65536 00:15:28.376 }, 00:15:28.376 { 00:15:28.376 "name": "BaseBdev3", 00:15:28.376 "uuid": "366ce04e-459b-5530-9516-bd873875900e", 00:15:28.376 "is_configured": true, 00:15:28.376 "data_offset": 0, 00:15:28.376 "data_size": 65536 00:15:28.376 }, 00:15:28.376 { 00:15:28.376 "name": "BaseBdev4", 00:15:28.376 "uuid": "49fb52a5-1369-5e2a-8db8-85f6be1c5554", 00:15:28.376 "is_configured": true, 00:15:28.376 "data_offset": 0, 00:15:28.376 "data_size": 65536 00:15:28.376 } 00:15:28.376 ] 00:15:28.376 }' 00:15:28.376 17:36:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:28.376 17:36:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:28.376 17:36:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:28.376 17:36:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:28.376 17:36:59 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:29.759 17:37:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:29.759 17:37:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:29.759 17:37:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:29.759 17:37:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:29.759 17:37:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:29.759 17:37:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:29.759 17:37:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:29.759 17:37:00 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:29.759 17:37:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:29.759 17:37:00 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:29.759 17:37:00 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:29.759 17:37:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:29.759 "name": "raid_bdev1", 00:15:29.759 "uuid": "9e3909f6-4d74-482f-a312-6db1a0a9be25", 00:15:29.759 "strip_size_kb": 64, 00:15:29.759 "state": "online", 00:15:29.759 "raid_level": "raid5f", 00:15:29.759 "superblock": false, 00:15:29.759 "num_base_bdevs": 4, 00:15:29.759 "num_base_bdevs_discovered": 4, 00:15:29.759 "num_base_bdevs_operational": 4, 00:15:29.759 "process": { 00:15:29.759 "type": "rebuild", 00:15:29.759 "target": "spare", 00:15:29.759 "progress": { 00:15:29.759 "blocks": 151680, 00:15:29.759 "percent": 77 00:15:29.759 } 00:15:29.759 }, 00:15:29.759 "base_bdevs_list": [ 00:15:29.759 { 00:15:29.759 "name": "spare", 00:15:29.759 "uuid": "499e8caa-bfa4-524f-958e-5b0bcc73cb3b", 00:15:29.759 "is_configured": true, 00:15:29.759 "data_offset": 0, 00:15:29.759 "data_size": 65536 00:15:29.759 }, 00:15:29.759 { 00:15:29.759 "name": "BaseBdev2", 00:15:29.759 "uuid": "7fa1e11b-e649-5f82-a5e6-6eceb73f5b9a", 00:15:29.759 "is_configured": true, 00:15:29.759 "data_offset": 0, 00:15:29.759 "data_size": 65536 00:15:29.759 }, 00:15:29.759 { 00:15:29.759 "name": "BaseBdev3", 00:15:29.759 "uuid": "366ce04e-459b-5530-9516-bd873875900e", 00:15:29.759 "is_configured": true, 00:15:29.759 "data_offset": 0, 00:15:29.759 "data_size": 65536 00:15:29.759 }, 00:15:29.759 { 00:15:29.759 "name": "BaseBdev4", 00:15:29.759 "uuid": "49fb52a5-1369-5e2a-8db8-85f6be1c5554", 00:15:29.759 "is_configured": true, 00:15:29.759 "data_offset": 0, 00:15:29.759 "data_size": 65536 00:15:29.759 } 00:15:29.759 ] 00:15:29.759 }' 00:15:29.759 17:37:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:29.759 17:37:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:29.759 17:37:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:29.759 17:37:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:29.759 17:37:00 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:30.701 17:37:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:30.701 17:37:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:30.701 17:37:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:30.701 17:37:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:30.701 17:37:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:30.701 17:37:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:30.701 17:37:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:30.701 17:37:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:30.701 17:37:01 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:30.701 17:37:01 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:30.701 17:37:01 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:30.701 17:37:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:30.701 "name": "raid_bdev1", 00:15:30.701 "uuid": "9e3909f6-4d74-482f-a312-6db1a0a9be25", 00:15:30.701 "strip_size_kb": 64, 00:15:30.701 "state": "online", 00:15:30.701 "raid_level": "raid5f", 00:15:30.701 "superblock": false, 00:15:30.701 "num_base_bdevs": 4, 00:15:30.701 "num_base_bdevs_discovered": 4, 00:15:30.701 "num_base_bdevs_operational": 4, 00:15:30.701 "process": { 00:15:30.701 "type": "rebuild", 00:15:30.701 "target": "spare", 00:15:30.701 "progress": { 00:15:30.701 "blocks": 174720, 00:15:30.701 "percent": 88 00:15:30.701 } 00:15:30.701 }, 00:15:30.701 "base_bdevs_list": [ 00:15:30.701 { 00:15:30.701 "name": "spare", 00:15:30.701 "uuid": "499e8caa-bfa4-524f-958e-5b0bcc73cb3b", 00:15:30.701 "is_configured": true, 00:15:30.701 "data_offset": 0, 00:15:30.701 "data_size": 65536 00:15:30.701 }, 00:15:30.701 { 00:15:30.701 "name": "BaseBdev2", 00:15:30.701 "uuid": "7fa1e11b-e649-5f82-a5e6-6eceb73f5b9a", 00:15:30.701 "is_configured": true, 00:15:30.701 "data_offset": 0, 00:15:30.701 "data_size": 65536 00:15:30.701 }, 00:15:30.701 { 00:15:30.701 "name": "BaseBdev3", 00:15:30.701 "uuid": "366ce04e-459b-5530-9516-bd873875900e", 00:15:30.701 "is_configured": true, 00:15:30.701 "data_offset": 0, 00:15:30.701 "data_size": 65536 00:15:30.701 }, 00:15:30.701 { 00:15:30.701 "name": "BaseBdev4", 00:15:30.701 "uuid": "49fb52a5-1369-5e2a-8db8-85f6be1c5554", 00:15:30.701 "is_configured": true, 00:15:30.701 "data_offset": 0, 00:15:30.701 "data_size": 65536 00:15:30.701 } 00:15:30.701 ] 00:15:30.701 }' 00:15:30.701 17:37:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:30.701 17:37:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:30.701 17:37:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:30.701 17:37:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:30.701 17:37:01 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:32.084 17:37:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:32.084 17:37:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:32.084 17:37:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:32.084 17:37:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:32.084 17:37:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:32.084 17:37:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:32.084 17:37:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:32.084 17:37:02 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:32.084 17:37:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:32.084 17:37:02 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:32.084 [2024-11-27 17:37:02.884577] bdev_raid.c:2896:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:15:32.084 [2024-11-27 17:37:02.884639] bdev_raid.c:2558:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:15:32.085 [2024-11-27 17:37:02.884679] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:32.085 17:37:02 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:32.085 17:37:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:32.085 "name": "raid_bdev1", 00:15:32.085 "uuid": "9e3909f6-4d74-482f-a312-6db1a0a9be25", 00:15:32.085 "strip_size_kb": 64, 00:15:32.085 "state": "online", 00:15:32.085 "raid_level": "raid5f", 00:15:32.085 "superblock": false, 00:15:32.085 "num_base_bdevs": 4, 00:15:32.085 "num_base_bdevs_discovered": 4, 00:15:32.085 "num_base_bdevs_operational": 4, 00:15:32.085 "base_bdevs_list": [ 00:15:32.085 { 00:15:32.085 "name": "spare", 00:15:32.085 "uuid": "499e8caa-bfa4-524f-958e-5b0bcc73cb3b", 00:15:32.085 "is_configured": true, 00:15:32.085 "data_offset": 0, 00:15:32.085 "data_size": 65536 00:15:32.085 }, 00:15:32.085 { 00:15:32.085 "name": "BaseBdev2", 00:15:32.085 "uuid": "7fa1e11b-e649-5f82-a5e6-6eceb73f5b9a", 00:15:32.085 "is_configured": true, 00:15:32.085 "data_offset": 0, 00:15:32.085 "data_size": 65536 00:15:32.085 }, 00:15:32.085 { 00:15:32.085 "name": "BaseBdev3", 00:15:32.085 "uuid": "366ce04e-459b-5530-9516-bd873875900e", 00:15:32.085 "is_configured": true, 00:15:32.085 "data_offset": 0, 00:15:32.085 "data_size": 65536 00:15:32.085 }, 00:15:32.085 { 00:15:32.085 "name": "BaseBdev4", 00:15:32.085 "uuid": "49fb52a5-1369-5e2a-8db8-85f6be1c5554", 00:15:32.085 "is_configured": true, 00:15:32.085 "data_offset": 0, 00:15:32.085 "data_size": 65536 00:15:32.085 } 00:15:32.085 ] 00:15:32.085 }' 00:15:32.085 17:37:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:32.085 17:37:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:15:32.085 17:37:02 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:32.085 17:37:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:15:32.085 17:37:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@709 -- # break 00:15:32.085 17:37:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:32.085 17:37:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:32.085 17:37:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:32.085 17:37:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:32.085 17:37:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:32.085 17:37:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:32.085 17:37:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:32.085 17:37:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:32.085 17:37:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:32.085 17:37:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:32.085 17:37:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:32.085 "name": "raid_bdev1", 00:15:32.085 "uuid": "9e3909f6-4d74-482f-a312-6db1a0a9be25", 00:15:32.085 "strip_size_kb": 64, 00:15:32.085 "state": "online", 00:15:32.085 "raid_level": "raid5f", 00:15:32.085 "superblock": false, 00:15:32.085 "num_base_bdevs": 4, 00:15:32.085 "num_base_bdevs_discovered": 4, 00:15:32.085 "num_base_bdevs_operational": 4, 00:15:32.085 "base_bdevs_list": [ 00:15:32.085 { 00:15:32.085 "name": "spare", 00:15:32.085 "uuid": "499e8caa-bfa4-524f-958e-5b0bcc73cb3b", 00:15:32.085 "is_configured": true, 00:15:32.085 "data_offset": 0, 00:15:32.085 "data_size": 65536 00:15:32.085 }, 00:15:32.085 { 00:15:32.085 "name": "BaseBdev2", 00:15:32.085 "uuid": "7fa1e11b-e649-5f82-a5e6-6eceb73f5b9a", 00:15:32.085 "is_configured": true, 00:15:32.085 "data_offset": 0, 00:15:32.085 "data_size": 65536 00:15:32.085 }, 00:15:32.085 { 00:15:32.085 "name": "BaseBdev3", 00:15:32.085 "uuid": "366ce04e-459b-5530-9516-bd873875900e", 00:15:32.085 "is_configured": true, 00:15:32.085 "data_offset": 0, 00:15:32.085 "data_size": 65536 00:15:32.085 }, 00:15:32.085 { 00:15:32.085 "name": "BaseBdev4", 00:15:32.085 "uuid": "49fb52a5-1369-5e2a-8db8-85f6be1c5554", 00:15:32.085 "is_configured": true, 00:15:32.085 "data_offset": 0, 00:15:32.085 "data_size": 65536 00:15:32.085 } 00:15:32.085 ] 00:15:32.085 }' 00:15:32.085 17:37:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:32.085 17:37:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:32.085 17:37:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:32.085 17:37:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:15:32.085 17:37:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 4 00:15:32.085 17:37:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:32.085 17:37:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:32.085 17:37:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:32.085 17:37:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:32.085 17:37:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:32.085 17:37:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:32.085 17:37:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:32.085 17:37:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:32.085 17:37:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:32.085 17:37:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:32.085 17:37:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:32.085 17:37:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:32.085 17:37:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:32.085 17:37:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:32.085 17:37:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:32.085 "name": "raid_bdev1", 00:15:32.085 "uuid": "9e3909f6-4d74-482f-a312-6db1a0a9be25", 00:15:32.085 "strip_size_kb": 64, 00:15:32.085 "state": "online", 00:15:32.085 "raid_level": "raid5f", 00:15:32.085 "superblock": false, 00:15:32.085 "num_base_bdevs": 4, 00:15:32.085 "num_base_bdevs_discovered": 4, 00:15:32.085 "num_base_bdevs_operational": 4, 00:15:32.085 "base_bdevs_list": [ 00:15:32.085 { 00:15:32.085 "name": "spare", 00:15:32.085 "uuid": "499e8caa-bfa4-524f-958e-5b0bcc73cb3b", 00:15:32.085 "is_configured": true, 00:15:32.085 "data_offset": 0, 00:15:32.085 "data_size": 65536 00:15:32.085 }, 00:15:32.085 { 00:15:32.085 "name": "BaseBdev2", 00:15:32.085 "uuid": "7fa1e11b-e649-5f82-a5e6-6eceb73f5b9a", 00:15:32.085 "is_configured": true, 00:15:32.085 "data_offset": 0, 00:15:32.085 "data_size": 65536 00:15:32.085 }, 00:15:32.085 { 00:15:32.085 "name": "BaseBdev3", 00:15:32.085 "uuid": "366ce04e-459b-5530-9516-bd873875900e", 00:15:32.085 "is_configured": true, 00:15:32.085 "data_offset": 0, 00:15:32.085 "data_size": 65536 00:15:32.085 }, 00:15:32.085 { 00:15:32.086 "name": "BaseBdev4", 00:15:32.086 "uuid": "49fb52a5-1369-5e2a-8db8-85f6be1c5554", 00:15:32.086 "is_configured": true, 00:15:32.086 "data_offset": 0, 00:15:32.086 "data_size": 65536 00:15:32.086 } 00:15:32.086 ] 00:15:32.086 }' 00:15:32.086 17:37:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:32.086 17:37:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:32.655 17:37:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:15:32.655 17:37:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:32.655 17:37:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:32.655 [2024-11-27 17:37:03.656369] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:15:32.655 [2024-11-27 17:37:03.656400] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:15:32.655 [2024-11-27 17:37:03.656471] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:32.655 [2024-11-27 17:37:03.656546] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:32.655 [2024-11-27 17:37:03.656556] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:15:32.655 17:37:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:32.655 17:37:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:32.655 17:37:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:32.655 17:37:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:32.655 17:37:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@720 -- # jq length 00:15:32.655 17:37:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:32.655 17:37:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:15:32.655 17:37:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:15:32.655 17:37:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:15:32.655 17:37:03 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:15:32.655 17:37:03 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:15:32.655 17:37:03 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:15:32.655 17:37:03 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@10 -- # local bdev_list 00:15:32.655 17:37:03 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:15:32.655 17:37:03 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@11 -- # local nbd_list 00:15:32.655 17:37:03 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@12 -- # local i 00:15:32.655 17:37:03 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:15:32.655 17:37:03 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:15:32.655 17:37:03 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:15:32.915 /dev/nbd0 00:15:32.915 17:37:03 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:15:32.915 17:37:03 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:15:32.915 17:37:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:15:32.915 17:37:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@869 -- # local i 00:15:32.915 17:37:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:15:32.915 17:37:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:15:32.915 17:37:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:15:32.915 17:37:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@873 -- # break 00:15:32.915 17:37:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:15:32.915 17:37:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:15:32.915 17:37:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:15:32.915 1+0 records in 00:15:32.915 1+0 records out 00:15:32.915 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000592629 s, 6.9 MB/s 00:15:32.915 17:37:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:32.915 17:37:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@886 -- # size=4096 00:15:32.915 17:37:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:32.915 17:37:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:15:32.915 17:37:03 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@889 -- # return 0 00:15:32.915 17:37:03 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:15:32.915 17:37:03 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:15:32.915 17:37:03 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:15:33.174 /dev/nbd1 00:15:33.174 17:37:04 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:15:33.175 17:37:04 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:15:33.175 17:37:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@868 -- # local nbd_name=nbd1 00:15:33.175 17:37:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@869 -- # local i 00:15:33.175 17:37:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:15:33.175 17:37:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:15:33.175 17:37:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@872 -- # grep -q -w nbd1 /proc/partitions 00:15:33.175 17:37:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@873 -- # break 00:15:33.175 17:37:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:15:33.175 17:37:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:15:33.175 17:37:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@885 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:15:33.175 1+0 records in 00:15:33.175 1+0 records out 00:15:33.175 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000435725 s, 9.4 MB/s 00:15:33.175 17:37:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:33.175 17:37:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@886 -- # size=4096 00:15:33.175 17:37:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:33.175 17:37:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:15:33.175 17:37:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@889 -- # return 0 00:15:33.175 17:37:04 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:15:33.175 17:37:04 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:15:33.175 17:37:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@738 -- # cmp -i 0 /dev/nbd0 /dev/nbd1 00:15:33.175 17:37:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:15:33.175 17:37:04 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:15:33.175 17:37:04 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:15:33.175 17:37:04 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@50 -- # local nbd_list 00:15:33.175 17:37:04 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@51 -- # local i 00:15:33.175 17:37:04 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:15:33.175 17:37:04 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:15:33.435 17:37:04 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:15:33.435 17:37:04 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:15:33.435 17:37:04 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:15:33.435 17:37:04 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:15:33.435 17:37:04 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:15:33.435 17:37:04 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:15:33.435 17:37:04 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:15:33.435 17:37:04 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:15:33.435 17:37:04 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:15:33.435 17:37:04 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:15:33.695 17:37:04 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:15:33.695 17:37:04 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:15:33.695 17:37:04 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:15:33.695 17:37:04 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:15:33.695 17:37:04 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:15:33.695 17:37:04 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:15:33.695 17:37:04 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@41 -- # break 00:15:33.695 17:37:04 bdev_raid.raid5f_rebuild_test -- bdev/nbd_common.sh@45 -- # return 0 00:15:33.695 17:37:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@743 -- # '[' false = true ']' 00:15:33.695 17:37:04 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@784 -- # killprocess 94879 00:15:33.695 17:37:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@950 -- # '[' -z 94879 ']' 00:15:33.695 17:37:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@954 -- # kill -0 94879 00:15:33.695 17:37:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@955 -- # uname 00:15:33.695 17:37:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:15:33.695 17:37:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 94879 00:15:33.695 17:37:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:15:33.695 killing process with pid 94879 00:15:33.695 Received shutdown signal, test time was about 60.000000 seconds 00:15:33.695 00:15:33.695 Latency(us) 00:15:33.695 [2024-11-27T17:37:04.887Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:15:33.695 [2024-11-27T17:37:04.887Z] =================================================================================================================== 00:15:33.695 [2024-11-27T17:37:04.887Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:15:33.695 17:37:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:15:33.695 17:37:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@968 -- # echo 'killing process with pid 94879' 00:15:33.695 17:37:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@969 -- # kill 94879 00:15:33.695 [2024-11-27 17:37:04.788567] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:15:33.695 17:37:04 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@974 -- # wait 94879 00:15:33.695 [2024-11-27 17:37:04.838226] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:15:33.956 17:37:05 bdev_raid.raid5f_rebuild_test -- bdev/bdev_raid.sh@786 -- # return 0 00:15:33.956 00:15:33.956 real 0m17.435s 00:15:33.956 user 0m21.174s 00:15:33.956 sys 0m2.413s 00:15:33.956 ************************************ 00:15:33.956 END TEST raid5f_rebuild_test 00:15:33.956 ************************************ 00:15:33.956 17:37:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:15:33.956 17:37:05 bdev_raid.raid5f_rebuild_test -- common/autotest_common.sh@10 -- # set +x 00:15:33.956 17:37:05 bdev_raid -- bdev/bdev_raid.sh@991 -- # run_test raid5f_rebuild_test_sb raid_rebuild_test raid5f 4 true false true 00:15:33.956 17:37:05 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 7 -le 1 ']' 00:15:33.956 17:37:05 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:15:33.956 17:37:05 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:15:33.956 ************************************ 00:15:33.956 START TEST raid5f_rebuild_test_sb 00:15:33.956 ************************************ 00:15:33.956 17:37:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@1125 -- # raid_rebuild_test raid5f 4 true false true 00:15:33.956 17:37:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@569 -- # local raid_level=raid5f 00:15:33.956 17:37:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=4 00:15:33.956 17:37:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:15:33.956 17:37:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:15:33.956 17:37:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@573 -- # local verify=true 00:15:34.217 17:37:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:15:34.217 17:37:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:15:34.217 17:37:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:15:34.217 17:37:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:15:34.217 17:37:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:15:34.217 17:37:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:15:34.217 17:37:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:15:34.217 17:37:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:15:34.217 17:37:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev3 00:15:34.217 17:37:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:15:34.217 17:37:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:15:34.217 17:37:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # echo BaseBdev4 00:15:34.217 17:37:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:15:34.217 17:37:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:15:34.217 17:37:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2' 'BaseBdev3' 'BaseBdev4') 00:15:34.217 17:37:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:15:34.217 17:37:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:15:34.217 17:37:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@576 -- # local strip_size 00:15:34.217 17:37:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@577 -- # local create_arg 00:15:34.217 17:37:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:15:34.217 17:37:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@579 -- # local data_offset 00:15:34.217 17:37:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@581 -- # '[' raid5f '!=' raid1 ']' 00:15:34.217 17:37:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@582 -- # '[' false = true ']' 00:15:34.217 17:37:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@586 -- # strip_size=64 00:15:34.217 17:37:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@587 -- # create_arg+=' -z 64' 00:15:34.217 17:37:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:15:34.217 17:37:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:15:34.217 17:37:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@597 -- # raid_pid=95364 00:15:34.217 17:37:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:15:34.217 17:37:05 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@598 -- # waitforlisten 95364 00:15:34.217 17:37:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@831 -- # '[' -z 95364 ']' 00:15:34.217 17:37:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:15:34.217 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:15:34.217 17:37:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@836 -- # local max_retries=100 00:15:34.217 17:37:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:15:34.217 17:37:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@840 -- # xtrace_disable 00:15:34.217 17:37:05 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:34.217 [2024-11-27 17:37:05.268364] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:15:34.217 [2024-11-27 17:37:05.268671] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid95364 ] 00:15:34.217 I/O size of 3145728 is greater than zero copy threshold (65536). 00:15:34.217 Zero copy mechanism will not be used. 00:15:34.477 [2024-11-27 17:37:05.419800] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:15:34.477 [2024-11-27 17:37:05.468242] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:15:34.477 [2024-11-27 17:37:05.512136] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:34.478 [2024-11-27 17:37:05.512266] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:35.048 17:37:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:15:35.048 17:37:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@864 -- # return 0 00:15:35.048 17:37:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:15:35.048 17:37:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev1_malloc 00:15:35.048 17:37:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:35.048 17:37:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:35.048 BaseBdev1_malloc 00:15:35.048 17:37:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:35.048 17:37:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:15:35.048 17:37:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:35.048 17:37:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:35.048 [2024-11-27 17:37:06.090284] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:15:35.048 [2024-11-27 17:37:06.090338] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:35.048 [2024-11-27 17:37:06.090360] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:15:35.048 [2024-11-27 17:37:06.090373] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:35.048 [2024-11-27 17:37:06.092412] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:35.048 [2024-11-27 17:37:06.092511] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:15:35.048 BaseBdev1 00:15:35.048 17:37:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:35.048 17:37:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:15:35.048 17:37:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev2_malloc 00:15:35.048 17:37:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:35.048 17:37:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:35.048 BaseBdev2_malloc 00:15:35.048 17:37:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:35.048 17:37:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:15:35.048 17:37:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:35.048 17:37:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:35.048 [2024-11-27 17:37:06.139280] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:15:35.048 [2024-11-27 17:37:06.139378] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:35.048 [2024-11-27 17:37:06.139423] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:15:35.048 [2024-11-27 17:37:06.139444] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:35.048 [2024-11-27 17:37:06.144259] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:35.048 [2024-11-27 17:37:06.144420] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:15:35.048 BaseBdev2 00:15:35.048 17:37:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:35.048 17:37:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:15:35.048 17:37:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev3_malloc 00:15:35.048 17:37:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:35.048 17:37:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:35.048 BaseBdev3_malloc 00:15:35.048 17:37:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:35.048 17:37:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev3_malloc -p BaseBdev3 00:15:35.048 17:37:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:35.048 17:37:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:35.048 [2024-11-27 17:37:06.170596] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev3_malloc 00:15:35.048 [2024-11-27 17:37:06.170651] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:35.048 [2024-11-27 17:37:06.170676] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:15:35.048 [2024-11-27 17:37:06.170685] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:35.048 [2024-11-27 17:37:06.172730] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:35.048 [2024-11-27 17:37:06.172767] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev3 00:15:35.048 BaseBdev3 00:15:35.048 17:37:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:35.048 17:37:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:15:35.048 17:37:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 512 -b BaseBdev4_malloc 00:15:35.048 17:37:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:35.048 17:37:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:35.048 BaseBdev4_malloc 00:15:35.048 17:37:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:35.048 17:37:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev4_malloc -p BaseBdev4 00:15:35.048 17:37:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:35.048 17:37:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:35.048 [2024-11-27 17:37:06.199495] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev4_malloc 00:15:35.048 [2024-11-27 17:37:06.199542] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:35.048 [2024-11-27 17:37:06.199567] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008a80 00:15:35.048 [2024-11-27 17:37:06.199576] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:35.048 [2024-11-27 17:37:06.201583] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:35.048 [2024-11-27 17:37:06.201619] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev4 00:15:35.048 BaseBdev4 00:15:35.048 17:37:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:35.048 17:37:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 512 -b spare_malloc 00:15:35.048 17:37:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:35.048 17:37:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:35.048 spare_malloc 00:15:35.048 17:37:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:35.048 17:37:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:15:35.048 17:37:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:35.048 17:37:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:35.048 spare_delay 00:15:35.048 17:37:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:35.048 17:37:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:15:35.048 17:37:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:35.048 17:37:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:35.309 [2024-11-27 17:37:06.240057] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:15:35.309 [2024-11-27 17:37:06.240102] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:35.309 [2024-11-27 17:37:06.240121] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009c80 00:15:35.309 [2024-11-27 17:37:06.240130] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:35.309 [2024-11-27 17:37:06.242161] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:35.309 [2024-11-27 17:37:06.242193] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:15:35.309 spare 00:15:35.309 17:37:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:35.309 17:37:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -z 64 -s -r raid5f -b ''\''BaseBdev1 BaseBdev2 BaseBdev3 BaseBdev4'\''' -n raid_bdev1 00:15:35.309 17:37:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:35.309 17:37:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:35.309 [2024-11-27 17:37:06.252116] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:15:35.309 [2024-11-27 17:37:06.253819] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:15:35.309 [2024-11-27 17:37:06.253870] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:15:35.309 [2024-11-27 17:37:06.253913] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:15:35.309 [2024-11-27 17:37:06.254067] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:15:35.309 [2024-11-27 17:37:06.254077] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:15:35.309 [2024-11-27 17:37:06.254379] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:15:35.309 [2024-11-27 17:37:06.254847] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:15:35.309 [2024-11-27 17:37:06.254898] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:15:35.309 [2024-11-27 17:37:06.255059] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:35.309 17:37:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:35.309 17:37:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 4 00:15:35.309 17:37:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:35.309 17:37:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:35.309 17:37:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:35.309 17:37:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:35.309 17:37:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:35.309 17:37:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:35.309 17:37:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:35.309 17:37:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:35.309 17:37:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:35.309 17:37:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:35.309 17:37:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:35.309 17:37:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:35.309 17:37:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:35.309 17:37:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:35.309 17:37:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:35.309 "name": "raid_bdev1", 00:15:35.309 "uuid": "084a4405-2241-475a-b267-f790b58fd56c", 00:15:35.309 "strip_size_kb": 64, 00:15:35.309 "state": "online", 00:15:35.309 "raid_level": "raid5f", 00:15:35.309 "superblock": true, 00:15:35.309 "num_base_bdevs": 4, 00:15:35.309 "num_base_bdevs_discovered": 4, 00:15:35.309 "num_base_bdevs_operational": 4, 00:15:35.309 "base_bdevs_list": [ 00:15:35.309 { 00:15:35.309 "name": "BaseBdev1", 00:15:35.309 "uuid": "8e98fc4e-86be-572b-889e-d64222c5aa3e", 00:15:35.309 "is_configured": true, 00:15:35.309 "data_offset": 2048, 00:15:35.309 "data_size": 63488 00:15:35.309 }, 00:15:35.309 { 00:15:35.309 "name": "BaseBdev2", 00:15:35.309 "uuid": "4d1cb354-a1cd-51f2-af06-696a7e2e1e92", 00:15:35.309 "is_configured": true, 00:15:35.310 "data_offset": 2048, 00:15:35.310 "data_size": 63488 00:15:35.310 }, 00:15:35.310 { 00:15:35.310 "name": "BaseBdev3", 00:15:35.310 "uuid": "caa820a2-49f8-5195-9dbe-f1fe056af663", 00:15:35.310 "is_configured": true, 00:15:35.310 "data_offset": 2048, 00:15:35.310 "data_size": 63488 00:15:35.310 }, 00:15:35.310 { 00:15:35.310 "name": "BaseBdev4", 00:15:35.310 "uuid": "887e80d8-ac51-55e0-a496-c3d1d2171574", 00:15:35.310 "is_configured": true, 00:15:35.310 "data_offset": 2048, 00:15:35.310 "data_size": 63488 00:15:35.310 } 00:15:35.310 ] 00:15:35.310 }' 00:15:35.310 17:37:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:35.310 17:37:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:35.570 17:37:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:15:35.570 17:37:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:35.570 17:37:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:35.570 17:37:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:15:35.830 [2024-11-27 17:37:06.764075] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:15:35.830 17:37:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:35.831 17:37:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=190464 00:15:35.831 17:37:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:35.831 17:37:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:35.831 17:37:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:35.831 17:37:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:15:35.831 17:37:06 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:35.831 17:37:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@619 -- # data_offset=2048 00:15:35.831 17:37:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:15:35.831 17:37:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:15:35.831 17:37:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:15:35.831 17:37:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:15:35.831 17:37:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:15:35.831 17:37:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:15:35.831 17:37:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # local bdev_list 00:15:35.831 17:37:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:15:35.831 17:37:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # local nbd_list 00:15:35.831 17:37:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@12 -- # local i 00:15:35.831 17:37:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:15:35.831 17:37:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:15:35.831 17:37:06 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:15:36.090 [2024-11-27 17:37:07.035444] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000027a0 00:15:36.090 /dev/nbd0 00:15:36.090 17:37:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:15:36.090 17:37:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:15:36.090 17:37:07 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:15:36.090 17:37:07 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@869 -- # local i 00:15:36.090 17:37:07 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:15:36.090 17:37:07 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:15:36.090 17:37:07 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:15:36.090 17:37:07 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@873 -- # break 00:15:36.090 17:37:07 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:15:36.090 17:37:07 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:15:36.090 17:37:07 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:15:36.090 1+0 records in 00:15:36.090 1+0 records out 00:15:36.090 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000548876 s, 7.5 MB/s 00:15:36.090 17:37:07 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:36.090 17:37:07 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@886 -- # size=4096 00:15:36.090 17:37:07 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:36.090 17:37:07 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:15:36.090 17:37:07 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@889 -- # return 0 00:15:36.090 17:37:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:15:36.090 17:37:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:15:36.090 17:37:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@629 -- # '[' raid5f = raid5f ']' 00:15:36.090 17:37:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@630 -- # write_unit_size=384 00:15:36.090 17:37:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@631 -- # echo 192 00:15:36.090 17:37:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=196608 count=496 oflag=direct 00:15:36.660 496+0 records in 00:15:36.660 496+0 records out 00:15:36.660 97517568 bytes (98 MB, 93 MiB) copied, 0.533422 s, 183 MB/s 00:15:36.660 17:37:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:15:36.660 17:37:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:15:36.660 17:37:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:15:36.660 17:37:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # local nbd_list 00:15:36.660 17:37:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@51 -- # local i 00:15:36.660 17:37:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:15:36.660 17:37:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:15:36.660 17:37:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:15:36.920 [2024-11-27 17:37:07.850973] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:36.920 17:37:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:15:36.920 17:37:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:15:36.920 17:37:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:15:36.920 17:37:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:15:36.920 17:37:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:15:36.920 17:37:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:15:36.920 17:37:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:15:36.920 17:37:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:15:36.920 17:37:07 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:36.920 17:37:07 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:36.920 [2024-11-27 17:37:07.867024] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:15:36.920 17:37:07 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:36.920 17:37:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:15:36.920 17:37:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:36.920 17:37:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:36.920 17:37:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:36.920 17:37:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:36.920 17:37:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:15:36.920 17:37:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:36.920 17:37:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:36.920 17:37:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:36.920 17:37:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:36.920 17:37:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:36.920 17:37:07 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:36.920 17:37:07 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:36.920 17:37:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:36.920 17:37:07 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:36.920 17:37:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:36.920 "name": "raid_bdev1", 00:15:36.921 "uuid": "084a4405-2241-475a-b267-f790b58fd56c", 00:15:36.921 "strip_size_kb": 64, 00:15:36.921 "state": "online", 00:15:36.921 "raid_level": "raid5f", 00:15:36.921 "superblock": true, 00:15:36.921 "num_base_bdevs": 4, 00:15:36.921 "num_base_bdevs_discovered": 3, 00:15:36.921 "num_base_bdevs_operational": 3, 00:15:36.921 "base_bdevs_list": [ 00:15:36.921 { 00:15:36.921 "name": null, 00:15:36.921 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:36.921 "is_configured": false, 00:15:36.921 "data_offset": 0, 00:15:36.921 "data_size": 63488 00:15:36.921 }, 00:15:36.921 { 00:15:36.921 "name": "BaseBdev2", 00:15:36.921 "uuid": "4d1cb354-a1cd-51f2-af06-696a7e2e1e92", 00:15:36.921 "is_configured": true, 00:15:36.921 "data_offset": 2048, 00:15:36.921 "data_size": 63488 00:15:36.921 }, 00:15:36.921 { 00:15:36.921 "name": "BaseBdev3", 00:15:36.921 "uuid": "caa820a2-49f8-5195-9dbe-f1fe056af663", 00:15:36.921 "is_configured": true, 00:15:36.921 "data_offset": 2048, 00:15:36.921 "data_size": 63488 00:15:36.921 }, 00:15:36.921 { 00:15:36.921 "name": "BaseBdev4", 00:15:36.921 "uuid": "887e80d8-ac51-55e0-a496-c3d1d2171574", 00:15:36.921 "is_configured": true, 00:15:36.921 "data_offset": 2048, 00:15:36.921 "data_size": 63488 00:15:36.921 } 00:15:36.921 ] 00:15:36.921 }' 00:15:36.921 17:37:07 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:36.921 17:37:07 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:37.180 17:37:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:15:37.180 17:37:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:37.180 17:37:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:37.180 [2024-11-27 17:37:08.294266] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:37.180 [2024-11-27 17:37:08.297550] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000270a0 00:15:37.180 [2024-11-27 17:37:08.299683] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:15:37.180 17:37:08 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:37.180 17:37:08 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@647 -- # sleep 1 00:15:38.120 17:37:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:38.120 17:37:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:38.120 17:37:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:38.120 17:37:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:38.120 17:37:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:38.380 17:37:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:38.380 17:37:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:38.380 17:37:09 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:38.380 17:37:09 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:38.380 17:37:09 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:38.380 17:37:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:38.380 "name": "raid_bdev1", 00:15:38.380 "uuid": "084a4405-2241-475a-b267-f790b58fd56c", 00:15:38.380 "strip_size_kb": 64, 00:15:38.380 "state": "online", 00:15:38.380 "raid_level": "raid5f", 00:15:38.380 "superblock": true, 00:15:38.380 "num_base_bdevs": 4, 00:15:38.380 "num_base_bdevs_discovered": 4, 00:15:38.380 "num_base_bdevs_operational": 4, 00:15:38.380 "process": { 00:15:38.380 "type": "rebuild", 00:15:38.380 "target": "spare", 00:15:38.380 "progress": { 00:15:38.380 "blocks": 19200, 00:15:38.380 "percent": 10 00:15:38.380 } 00:15:38.380 }, 00:15:38.380 "base_bdevs_list": [ 00:15:38.380 { 00:15:38.380 "name": "spare", 00:15:38.380 "uuid": "b97b13ff-990b-5307-ba00-202b7d15c61b", 00:15:38.380 "is_configured": true, 00:15:38.380 "data_offset": 2048, 00:15:38.380 "data_size": 63488 00:15:38.380 }, 00:15:38.380 { 00:15:38.380 "name": "BaseBdev2", 00:15:38.380 "uuid": "4d1cb354-a1cd-51f2-af06-696a7e2e1e92", 00:15:38.380 "is_configured": true, 00:15:38.380 "data_offset": 2048, 00:15:38.380 "data_size": 63488 00:15:38.380 }, 00:15:38.380 { 00:15:38.380 "name": "BaseBdev3", 00:15:38.380 "uuid": "caa820a2-49f8-5195-9dbe-f1fe056af663", 00:15:38.380 "is_configured": true, 00:15:38.380 "data_offset": 2048, 00:15:38.380 "data_size": 63488 00:15:38.380 }, 00:15:38.380 { 00:15:38.380 "name": "BaseBdev4", 00:15:38.380 "uuid": "887e80d8-ac51-55e0-a496-c3d1d2171574", 00:15:38.380 "is_configured": true, 00:15:38.380 "data_offset": 2048, 00:15:38.380 "data_size": 63488 00:15:38.380 } 00:15:38.380 ] 00:15:38.380 }' 00:15:38.380 17:37:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:38.380 17:37:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:38.380 17:37:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:38.380 17:37:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:38.380 17:37:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:15:38.380 17:37:09 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:38.380 17:37:09 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:38.380 [2024-11-27 17:37:09.450236] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:38.380 [2024-11-27 17:37:09.504944] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:15:38.380 [2024-11-27 17:37:09.505043] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:38.380 [2024-11-27 17:37:09.505066] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:38.380 [2024-11-27 17:37:09.505076] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:15:38.380 17:37:09 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:38.380 17:37:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:15:38.380 17:37:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:38.380 17:37:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:38.380 17:37:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:38.380 17:37:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:38.380 17:37:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:15:38.380 17:37:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:38.380 17:37:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:38.380 17:37:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:38.380 17:37:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:38.380 17:37:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:38.380 17:37:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:38.380 17:37:09 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:38.380 17:37:09 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:38.380 17:37:09 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:38.639 17:37:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:38.640 "name": "raid_bdev1", 00:15:38.640 "uuid": "084a4405-2241-475a-b267-f790b58fd56c", 00:15:38.640 "strip_size_kb": 64, 00:15:38.640 "state": "online", 00:15:38.640 "raid_level": "raid5f", 00:15:38.640 "superblock": true, 00:15:38.640 "num_base_bdevs": 4, 00:15:38.640 "num_base_bdevs_discovered": 3, 00:15:38.640 "num_base_bdevs_operational": 3, 00:15:38.640 "base_bdevs_list": [ 00:15:38.640 { 00:15:38.640 "name": null, 00:15:38.640 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:38.640 "is_configured": false, 00:15:38.640 "data_offset": 0, 00:15:38.640 "data_size": 63488 00:15:38.640 }, 00:15:38.640 { 00:15:38.640 "name": "BaseBdev2", 00:15:38.640 "uuid": "4d1cb354-a1cd-51f2-af06-696a7e2e1e92", 00:15:38.640 "is_configured": true, 00:15:38.640 "data_offset": 2048, 00:15:38.640 "data_size": 63488 00:15:38.640 }, 00:15:38.640 { 00:15:38.640 "name": "BaseBdev3", 00:15:38.640 "uuid": "caa820a2-49f8-5195-9dbe-f1fe056af663", 00:15:38.640 "is_configured": true, 00:15:38.640 "data_offset": 2048, 00:15:38.640 "data_size": 63488 00:15:38.640 }, 00:15:38.640 { 00:15:38.640 "name": "BaseBdev4", 00:15:38.640 "uuid": "887e80d8-ac51-55e0-a496-c3d1d2171574", 00:15:38.640 "is_configured": true, 00:15:38.640 "data_offset": 2048, 00:15:38.640 "data_size": 63488 00:15:38.640 } 00:15:38.640 ] 00:15:38.640 }' 00:15:38.640 17:37:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:38.640 17:37:09 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:38.899 17:37:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:38.899 17:37:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:38.899 17:37:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:38.899 17:37:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:38.899 17:37:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:38.899 17:37:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:38.900 17:37:09 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:38.900 17:37:09 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:38.900 17:37:09 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:38.900 17:37:10 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:38.900 17:37:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:38.900 "name": "raid_bdev1", 00:15:38.900 "uuid": "084a4405-2241-475a-b267-f790b58fd56c", 00:15:38.900 "strip_size_kb": 64, 00:15:38.900 "state": "online", 00:15:38.900 "raid_level": "raid5f", 00:15:38.900 "superblock": true, 00:15:38.900 "num_base_bdevs": 4, 00:15:38.900 "num_base_bdevs_discovered": 3, 00:15:38.900 "num_base_bdevs_operational": 3, 00:15:38.900 "base_bdevs_list": [ 00:15:38.900 { 00:15:38.900 "name": null, 00:15:38.900 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:38.900 "is_configured": false, 00:15:38.900 "data_offset": 0, 00:15:38.900 "data_size": 63488 00:15:38.900 }, 00:15:38.900 { 00:15:38.900 "name": "BaseBdev2", 00:15:38.900 "uuid": "4d1cb354-a1cd-51f2-af06-696a7e2e1e92", 00:15:38.900 "is_configured": true, 00:15:38.900 "data_offset": 2048, 00:15:38.900 "data_size": 63488 00:15:38.900 }, 00:15:38.900 { 00:15:38.900 "name": "BaseBdev3", 00:15:38.900 "uuid": "caa820a2-49f8-5195-9dbe-f1fe056af663", 00:15:38.900 "is_configured": true, 00:15:38.900 "data_offset": 2048, 00:15:38.900 "data_size": 63488 00:15:38.900 }, 00:15:38.900 { 00:15:38.900 "name": "BaseBdev4", 00:15:38.900 "uuid": "887e80d8-ac51-55e0-a496-c3d1d2171574", 00:15:38.900 "is_configured": true, 00:15:38.900 "data_offset": 2048, 00:15:38.900 "data_size": 63488 00:15:38.900 } 00:15:38.900 ] 00:15:38.900 }' 00:15:38.900 17:37:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:38.900 17:37:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:38.900 17:37:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:39.160 17:37:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:15:39.160 17:37:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:15:39.160 17:37:10 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:39.160 17:37:10 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:39.160 [2024-11-27 17:37:10.129202] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:39.160 [2024-11-27 17:37:10.132101] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000027170 00:15:39.160 [2024-11-27 17:37:10.134214] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:15:39.160 17:37:10 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:39.160 17:37:10 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@663 -- # sleep 1 00:15:40.100 17:37:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:40.100 17:37:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:40.100 17:37:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:40.100 17:37:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:40.100 17:37:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:40.100 17:37:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:40.100 17:37:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:40.101 17:37:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:40.101 17:37:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:40.101 17:37:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:40.101 17:37:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:40.101 "name": "raid_bdev1", 00:15:40.101 "uuid": "084a4405-2241-475a-b267-f790b58fd56c", 00:15:40.101 "strip_size_kb": 64, 00:15:40.101 "state": "online", 00:15:40.101 "raid_level": "raid5f", 00:15:40.101 "superblock": true, 00:15:40.101 "num_base_bdevs": 4, 00:15:40.101 "num_base_bdevs_discovered": 4, 00:15:40.101 "num_base_bdevs_operational": 4, 00:15:40.101 "process": { 00:15:40.101 "type": "rebuild", 00:15:40.101 "target": "spare", 00:15:40.101 "progress": { 00:15:40.101 "blocks": 19200, 00:15:40.101 "percent": 10 00:15:40.101 } 00:15:40.101 }, 00:15:40.101 "base_bdevs_list": [ 00:15:40.101 { 00:15:40.101 "name": "spare", 00:15:40.101 "uuid": "b97b13ff-990b-5307-ba00-202b7d15c61b", 00:15:40.101 "is_configured": true, 00:15:40.101 "data_offset": 2048, 00:15:40.101 "data_size": 63488 00:15:40.101 }, 00:15:40.101 { 00:15:40.101 "name": "BaseBdev2", 00:15:40.101 "uuid": "4d1cb354-a1cd-51f2-af06-696a7e2e1e92", 00:15:40.101 "is_configured": true, 00:15:40.101 "data_offset": 2048, 00:15:40.101 "data_size": 63488 00:15:40.101 }, 00:15:40.101 { 00:15:40.101 "name": "BaseBdev3", 00:15:40.101 "uuid": "caa820a2-49f8-5195-9dbe-f1fe056af663", 00:15:40.101 "is_configured": true, 00:15:40.101 "data_offset": 2048, 00:15:40.101 "data_size": 63488 00:15:40.101 }, 00:15:40.101 { 00:15:40.101 "name": "BaseBdev4", 00:15:40.101 "uuid": "887e80d8-ac51-55e0-a496-c3d1d2171574", 00:15:40.101 "is_configured": true, 00:15:40.101 "data_offset": 2048, 00:15:40.101 "data_size": 63488 00:15:40.101 } 00:15:40.101 ] 00:15:40.101 }' 00:15:40.101 17:37:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:40.101 17:37:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:40.101 17:37:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:40.101 17:37:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:40.101 17:37:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:15:40.101 17:37:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:15:40.101 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:15:40.101 17:37:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=4 00:15:40.101 17:37:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@693 -- # '[' raid5f = raid1 ']' 00:15:40.101 17:37:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@706 -- # local timeout=539 00:15:40.101 17:37:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:40.101 17:37:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:40.101 17:37:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:40.101 17:37:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:40.101 17:37:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:40.101 17:37:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:40.101 17:37:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:40.101 17:37:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:40.101 17:37:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:40.101 17:37:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:40.361 17:37:11 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:40.361 17:37:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:40.361 "name": "raid_bdev1", 00:15:40.361 "uuid": "084a4405-2241-475a-b267-f790b58fd56c", 00:15:40.361 "strip_size_kb": 64, 00:15:40.361 "state": "online", 00:15:40.361 "raid_level": "raid5f", 00:15:40.361 "superblock": true, 00:15:40.361 "num_base_bdevs": 4, 00:15:40.361 "num_base_bdevs_discovered": 4, 00:15:40.361 "num_base_bdevs_operational": 4, 00:15:40.361 "process": { 00:15:40.361 "type": "rebuild", 00:15:40.361 "target": "spare", 00:15:40.361 "progress": { 00:15:40.361 "blocks": 21120, 00:15:40.361 "percent": 11 00:15:40.361 } 00:15:40.361 }, 00:15:40.361 "base_bdevs_list": [ 00:15:40.361 { 00:15:40.361 "name": "spare", 00:15:40.361 "uuid": "b97b13ff-990b-5307-ba00-202b7d15c61b", 00:15:40.361 "is_configured": true, 00:15:40.361 "data_offset": 2048, 00:15:40.361 "data_size": 63488 00:15:40.361 }, 00:15:40.361 { 00:15:40.361 "name": "BaseBdev2", 00:15:40.361 "uuid": "4d1cb354-a1cd-51f2-af06-696a7e2e1e92", 00:15:40.361 "is_configured": true, 00:15:40.361 "data_offset": 2048, 00:15:40.361 "data_size": 63488 00:15:40.361 }, 00:15:40.361 { 00:15:40.361 "name": "BaseBdev3", 00:15:40.361 "uuid": "caa820a2-49f8-5195-9dbe-f1fe056af663", 00:15:40.361 "is_configured": true, 00:15:40.361 "data_offset": 2048, 00:15:40.361 "data_size": 63488 00:15:40.361 }, 00:15:40.361 { 00:15:40.361 "name": "BaseBdev4", 00:15:40.361 "uuid": "887e80d8-ac51-55e0-a496-c3d1d2171574", 00:15:40.361 "is_configured": true, 00:15:40.361 "data_offset": 2048, 00:15:40.361 "data_size": 63488 00:15:40.361 } 00:15:40.361 ] 00:15:40.361 }' 00:15:40.361 17:37:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:40.361 17:37:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:40.361 17:37:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:40.361 17:37:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:40.361 17:37:11 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:41.300 17:37:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:41.300 17:37:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:41.300 17:37:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:41.300 17:37:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:41.300 17:37:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:41.300 17:37:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:41.300 17:37:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:41.300 17:37:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:41.301 17:37:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:41.301 17:37:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:41.301 17:37:12 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:41.301 17:37:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:41.301 "name": "raid_bdev1", 00:15:41.301 "uuid": "084a4405-2241-475a-b267-f790b58fd56c", 00:15:41.301 "strip_size_kb": 64, 00:15:41.301 "state": "online", 00:15:41.301 "raid_level": "raid5f", 00:15:41.301 "superblock": true, 00:15:41.301 "num_base_bdevs": 4, 00:15:41.301 "num_base_bdevs_discovered": 4, 00:15:41.301 "num_base_bdevs_operational": 4, 00:15:41.301 "process": { 00:15:41.301 "type": "rebuild", 00:15:41.301 "target": "spare", 00:15:41.301 "progress": { 00:15:41.301 "blocks": 44160, 00:15:41.301 "percent": 23 00:15:41.301 } 00:15:41.301 }, 00:15:41.301 "base_bdevs_list": [ 00:15:41.301 { 00:15:41.301 "name": "spare", 00:15:41.301 "uuid": "b97b13ff-990b-5307-ba00-202b7d15c61b", 00:15:41.301 "is_configured": true, 00:15:41.301 "data_offset": 2048, 00:15:41.301 "data_size": 63488 00:15:41.301 }, 00:15:41.301 { 00:15:41.301 "name": "BaseBdev2", 00:15:41.301 "uuid": "4d1cb354-a1cd-51f2-af06-696a7e2e1e92", 00:15:41.301 "is_configured": true, 00:15:41.301 "data_offset": 2048, 00:15:41.301 "data_size": 63488 00:15:41.301 }, 00:15:41.301 { 00:15:41.301 "name": "BaseBdev3", 00:15:41.301 "uuid": "caa820a2-49f8-5195-9dbe-f1fe056af663", 00:15:41.301 "is_configured": true, 00:15:41.301 "data_offset": 2048, 00:15:41.301 "data_size": 63488 00:15:41.301 }, 00:15:41.301 { 00:15:41.301 "name": "BaseBdev4", 00:15:41.301 "uuid": "887e80d8-ac51-55e0-a496-c3d1d2171574", 00:15:41.301 "is_configured": true, 00:15:41.301 "data_offset": 2048, 00:15:41.301 "data_size": 63488 00:15:41.301 } 00:15:41.301 ] 00:15:41.301 }' 00:15:41.561 17:37:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:41.561 17:37:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:41.561 17:37:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:41.561 17:37:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:41.561 17:37:12 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:42.502 17:37:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:42.502 17:37:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:42.502 17:37:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:42.502 17:37:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:42.502 17:37:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:42.502 17:37:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:42.502 17:37:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:42.502 17:37:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:42.502 17:37:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:42.502 17:37:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:42.502 17:37:13 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:42.502 17:37:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:42.502 "name": "raid_bdev1", 00:15:42.502 "uuid": "084a4405-2241-475a-b267-f790b58fd56c", 00:15:42.502 "strip_size_kb": 64, 00:15:42.502 "state": "online", 00:15:42.502 "raid_level": "raid5f", 00:15:42.502 "superblock": true, 00:15:42.502 "num_base_bdevs": 4, 00:15:42.502 "num_base_bdevs_discovered": 4, 00:15:42.502 "num_base_bdevs_operational": 4, 00:15:42.502 "process": { 00:15:42.502 "type": "rebuild", 00:15:42.502 "target": "spare", 00:15:42.502 "progress": { 00:15:42.502 "blocks": 65280, 00:15:42.502 "percent": 34 00:15:42.502 } 00:15:42.502 }, 00:15:42.502 "base_bdevs_list": [ 00:15:42.502 { 00:15:42.502 "name": "spare", 00:15:42.502 "uuid": "b97b13ff-990b-5307-ba00-202b7d15c61b", 00:15:42.502 "is_configured": true, 00:15:42.502 "data_offset": 2048, 00:15:42.502 "data_size": 63488 00:15:42.502 }, 00:15:42.502 { 00:15:42.502 "name": "BaseBdev2", 00:15:42.502 "uuid": "4d1cb354-a1cd-51f2-af06-696a7e2e1e92", 00:15:42.502 "is_configured": true, 00:15:42.502 "data_offset": 2048, 00:15:42.502 "data_size": 63488 00:15:42.502 }, 00:15:42.502 { 00:15:42.502 "name": "BaseBdev3", 00:15:42.502 "uuid": "caa820a2-49f8-5195-9dbe-f1fe056af663", 00:15:42.502 "is_configured": true, 00:15:42.502 "data_offset": 2048, 00:15:42.502 "data_size": 63488 00:15:42.503 }, 00:15:42.503 { 00:15:42.503 "name": "BaseBdev4", 00:15:42.503 "uuid": "887e80d8-ac51-55e0-a496-c3d1d2171574", 00:15:42.503 "is_configured": true, 00:15:42.503 "data_offset": 2048, 00:15:42.503 "data_size": 63488 00:15:42.503 } 00:15:42.503 ] 00:15:42.503 }' 00:15:42.503 17:37:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:42.763 17:37:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:42.763 17:37:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:42.763 17:37:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:42.763 17:37:13 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:43.702 17:37:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:43.702 17:37:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:43.702 17:37:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:43.702 17:37:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:43.702 17:37:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:43.702 17:37:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:43.702 17:37:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:43.702 17:37:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:43.702 17:37:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:43.702 17:37:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:43.702 17:37:14 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:43.702 17:37:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:43.702 "name": "raid_bdev1", 00:15:43.703 "uuid": "084a4405-2241-475a-b267-f790b58fd56c", 00:15:43.703 "strip_size_kb": 64, 00:15:43.703 "state": "online", 00:15:43.703 "raid_level": "raid5f", 00:15:43.703 "superblock": true, 00:15:43.703 "num_base_bdevs": 4, 00:15:43.703 "num_base_bdevs_discovered": 4, 00:15:43.703 "num_base_bdevs_operational": 4, 00:15:43.703 "process": { 00:15:43.703 "type": "rebuild", 00:15:43.703 "target": "spare", 00:15:43.703 "progress": { 00:15:43.703 "blocks": 88320, 00:15:43.703 "percent": 46 00:15:43.703 } 00:15:43.703 }, 00:15:43.703 "base_bdevs_list": [ 00:15:43.703 { 00:15:43.703 "name": "spare", 00:15:43.703 "uuid": "b97b13ff-990b-5307-ba00-202b7d15c61b", 00:15:43.703 "is_configured": true, 00:15:43.703 "data_offset": 2048, 00:15:43.703 "data_size": 63488 00:15:43.703 }, 00:15:43.703 { 00:15:43.703 "name": "BaseBdev2", 00:15:43.703 "uuid": "4d1cb354-a1cd-51f2-af06-696a7e2e1e92", 00:15:43.703 "is_configured": true, 00:15:43.703 "data_offset": 2048, 00:15:43.703 "data_size": 63488 00:15:43.703 }, 00:15:43.703 { 00:15:43.703 "name": "BaseBdev3", 00:15:43.703 "uuid": "caa820a2-49f8-5195-9dbe-f1fe056af663", 00:15:43.703 "is_configured": true, 00:15:43.703 "data_offset": 2048, 00:15:43.703 "data_size": 63488 00:15:43.703 }, 00:15:43.703 { 00:15:43.703 "name": "BaseBdev4", 00:15:43.703 "uuid": "887e80d8-ac51-55e0-a496-c3d1d2171574", 00:15:43.703 "is_configured": true, 00:15:43.703 "data_offset": 2048, 00:15:43.703 "data_size": 63488 00:15:43.703 } 00:15:43.703 ] 00:15:43.703 }' 00:15:43.703 17:37:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:43.703 17:37:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:43.703 17:37:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:43.703 17:37:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:43.703 17:37:14 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:45.085 17:37:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:45.085 17:37:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:45.085 17:37:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:45.085 17:37:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:45.085 17:37:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:45.085 17:37:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:45.085 17:37:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:45.085 17:37:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:45.085 17:37:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:45.085 17:37:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:45.085 17:37:15 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:45.085 17:37:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:45.085 "name": "raid_bdev1", 00:15:45.085 "uuid": "084a4405-2241-475a-b267-f790b58fd56c", 00:15:45.085 "strip_size_kb": 64, 00:15:45.085 "state": "online", 00:15:45.085 "raid_level": "raid5f", 00:15:45.085 "superblock": true, 00:15:45.085 "num_base_bdevs": 4, 00:15:45.085 "num_base_bdevs_discovered": 4, 00:15:45.085 "num_base_bdevs_operational": 4, 00:15:45.086 "process": { 00:15:45.086 "type": "rebuild", 00:15:45.086 "target": "spare", 00:15:45.086 "progress": { 00:15:45.086 "blocks": 109440, 00:15:45.086 "percent": 57 00:15:45.086 } 00:15:45.086 }, 00:15:45.086 "base_bdevs_list": [ 00:15:45.086 { 00:15:45.086 "name": "spare", 00:15:45.086 "uuid": "b97b13ff-990b-5307-ba00-202b7d15c61b", 00:15:45.086 "is_configured": true, 00:15:45.086 "data_offset": 2048, 00:15:45.086 "data_size": 63488 00:15:45.086 }, 00:15:45.086 { 00:15:45.086 "name": "BaseBdev2", 00:15:45.086 "uuid": "4d1cb354-a1cd-51f2-af06-696a7e2e1e92", 00:15:45.086 "is_configured": true, 00:15:45.086 "data_offset": 2048, 00:15:45.086 "data_size": 63488 00:15:45.086 }, 00:15:45.086 { 00:15:45.086 "name": "BaseBdev3", 00:15:45.086 "uuid": "caa820a2-49f8-5195-9dbe-f1fe056af663", 00:15:45.086 "is_configured": true, 00:15:45.086 "data_offset": 2048, 00:15:45.086 "data_size": 63488 00:15:45.086 }, 00:15:45.086 { 00:15:45.086 "name": "BaseBdev4", 00:15:45.086 "uuid": "887e80d8-ac51-55e0-a496-c3d1d2171574", 00:15:45.086 "is_configured": true, 00:15:45.086 "data_offset": 2048, 00:15:45.086 "data_size": 63488 00:15:45.086 } 00:15:45.086 ] 00:15:45.086 }' 00:15:45.086 17:37:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:45.086 17:37:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:45.086 17:37:15 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:45.086 17:37:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:45.086 17:37:16 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:46.026 17:37:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:46.026 17:37:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:46.026 17:37:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:46.026 17:37:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:46.026 17:37:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:46.027 17:37:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:46.027 17:37:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:46.027 17:37:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:46.027 17:37:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:46.027 17:37:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:46.027 17:37:17 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:46.027 17:37:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:46.027 "name": "raid_bdev1", 00:15:46.027 "uuid": "084a4405-2241-475a-b267-f790b58fd56c", 00:15:46.027 "strip_size_kb": 64, 00:15:46.027 "state": "online", 00:15:46.027 "raid_level": "raid5f", 00:15:46.027 "superblock": true, 00:15:46.027 "num_base_bdevs": 4, 00:15:46.027 "num_base_bdevs_discovered": 4, 00:15:46.027 "num_base_bdevs_operational": 4, 00:15:46.027 "process": { 00:15:46.027 "type": "rebuild", 00:15:46.027 "target": "spare", 00:15:46.027 "progress": { 00:15:46.027 "blocks": 130560, 00:15:46.027 "percent": 68 00:15:46.027 } 00:15:46.027 }, 00:15:46.027 "base_bdevs_list": [ 00:15:46.027 { 00:15:46.027 "name": "spare", 00:15:46.027 "uuid": "b97b13ff-990b-5307-ba00-202b7d15c61b", 00:15:46.027 "is_configured": true, 00:15:46.027 "data_offset": 2048, 00:15:46.027 "data_size": 63488 00:15:46.027 }, 00:15:46.027 { 00:15:46.027 "name": "BaseBdev2", 00:15:46.027 "uuid": "4d1cb354-a1cd-51f2-af06-696a7e2e1e92", 00:15:46.027 "is_configured": true, 00:15:46.027 "data_offset": 2048, 00:15:46.027 "data_size": 63488 00:15:46.027 }, 00:15:46.027 { 00:15:46.027 "name": "BaseBdev3", 00:15:46.027 "uuid": "caa820a2-49f8-5195-9dbe-f1fe056af663", 00:15:46.027 "is_configured": true, 00:15:46.027 "data_offset": 2048, 00:15:46.027 "data_size": 63488 00:15:46.027 }, 00:15:46.027 { 00:15:46.027 "name": "BaseBdev4", 00:15:46.027 "uuid": "887e80d8-ac51-55e0-a496-c3d1d2171574", 00:15:46.027 "is_configured": true, 00:15:46.027 "data_offset": 2048, 00:15:46.027 "data_size": 63488 00:15:46.027 } 00:15:46.027 ] 00:15:46.027 }' 00:15:46.027 17:37:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:46.027 17:37:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:46.027 17:37:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:46.027 17:37:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:46.027 17:37:17 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:47.409 17:37:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:47.409 17:37:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:47.409 17:37:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:47.409 17:37:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:47.409 17:37:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:47.409 17:37:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:47.409 17:37:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:47.409 17:37:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:47.409 17:37:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:47.409 17:37:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:47.409 17:37:18 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:47.409 17:37:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:47.409 "name": "raid_bdev1", 00:15:47.409 "uuid": "084a4405-2241-475a-b267-f790b58fd56c", 00:15:47.409 "strip_size_kb": 64, 00:15:47.409 "state": "online", 00:15:47.409 "raid_level": "raid5f", 00:15:47.409 "superblock": true, 00:15:47.409 "num_base_bdevs": 4, 00:15:47.409 "num_base_bdevs_discovered": 4, 00:15:47.409 "num_base_bdevs_operational": 4, 00:15:47.409 "process": { 00:15:47.409 "type": "rebuild", 00:15:47.409 "target": "spare", 00:15:47.409 "progress": { 00:15:47.409 "blocks": 153600, 00:15:47.409 "percent": 80 00:15:47.409 } 00:15:47.409 }, 00:15:47.409 "base_bdevs_list": [ 00:15:47.409 { 00:15:47.409 "name": "spare", 00:15:47.409 "uuid": "b97b13ff-990b-5307-ba00-202b7d15c61b", 00:15:47.409 "is_configured": true, 00:15:47.409 "data_offset": 2048, 00:15:47.409 "data_size": 63488 00:15:47.409 }, 00:15:47.409 { 00:15:47.409 "name": "BaseBdev2", 00:15:47.409 "uuid": "4d1cb354-a1cd-51f2-af06-696a7e2e1e92", 00:15:47.409 "is_configured": true, 00:15:47.409 "data_offset": 2048, 00:15:47.409 "data_size": 63488 00:15:47.409 }, 00:15:47.409 { 00:15:47.409 "name": "BaseBdev3", 00:15:47.409 "uuid": "caa820a2-49f8-5195-9dbe-f1fe056af663", 00:15:47.409 "is_configured": true, 00:15:47.409 "data_offset": 2048, 00:15:47.409 "data_size": 63488 00:15:47.409 }, 00:15:47.409 { 00:15:47.409 "name": "BaseBdev4", 00:15:47.409 "uuid": "887e80d8-ac51-55e0-a496-c3d1d2171574", 00:15:47.409 "is_configured": true, 00:15:47.409 "data_offset": 2048, 00:15:47.409 "data_size": 63488 00:15:47.409 } 00:15:47.409 ] 00:15:47.409 }' 00:15:47.409 17:37:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:47.409 17:37:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:47.409 17:37:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:47.409 17:37:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:47.409 17:37:18 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:48.378 17:37:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:48.378 17:37:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:48.378 17:37:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:48.378 17:37:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:48.378 17:37:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:48.378 17:37:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:48.378 17:37:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:48.378 17:37:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:48.378 17:37:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:48.378 17:37:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:48.378 17:37:19 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:48.378 17:37:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:48.378 "name": "raid_bdev1", 00:15:48.378 "uuid": "084a4405-2241-475a-b267-f790b58fd56c", 00:15:48.378 "strip_size_kb": 64, 00:15:48.378 "state": "online", 00:15:48.378 "raid_level": "raid5f", 00:15:48.378 "superblock": true, 00:15:48.378 "num_base_bdevs": 4, 00:15:48.378 "num_base_bdevs_discovered": 4, 00:15:48.378 "num_base_bdevs_operational": 4, 00:15:48.378 "process": { 00:15:48.378 "type": "rebuild", 00:15:48.378 "target": "spare", 00:15:48.378 "progress": { 00:15:48.378 "blocks": 174720, 00:15:48.378 "percent": 91 00:15:48.378 } 00:15:48.378 }, 00:15:48.378 "base_bdevs_list": [ 00:15:48.378 { 00:15:48.378 "name": "spare", 00:15:48.378 "uuid": "b97b13ff-990b-5307-ba00-202b7d15c61b", 00:15:48.378 "is_configured": true, 00:15:48.378 "data_offset": 2048, 00:15:48.378 "data_size": 63488 00:15:48.378 }, 00:15:48.378 { 00:15:48.378 "name": "BaseBdev2", 00:15:48.378 "uuid": "4d1cb354-a1cd-51f2-af06-696a7e2e1e92", 00:15:48.378 "is_configured": true, 00:15:48.378 "data_offset": 2048, 00:15:48.378 "data_size": 63488 00:15:48.378 }, 00:15:48.378 { 00:15:48.378 "name": "BaseBdev3", 00:15:48.378 "uuid": "caa820a2-49f8-5195-9dbe-f1fe056af663", 00:15:48.378 "is_configured": true, 00:15:48.378 "data_offset": 2048, 00:15:48.378 "data_size": 63488 00:15:48.378 }, 00:15:48.378 { 00:15:48.378 "name": "BaseBdev4", 00:15:48.378 "uuid": "887e80d8-ac51-55e0-a496-c3d1d2171574", 00:15:48.378 "is_configured": true, 00:15:48.378 "data_offset": 2048, 00:15:48.378 "data_size": 63488 00:15:48.378 } 00:15:48.378 ] 00:15:48.378 }' 00:15:48.378 17:37:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:48.378 17:37:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:48.378 17:37:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:48.378 17:37:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:48.378 17:37:19 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@711 -- # sleep 1 00:15:49.318 [2024-11-27 17:37:20.172255] bdev_raid.c:2896:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:15:49.318 [2024-11-27 17:37:20.172322] bdev_raid.c:2558:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:15:49.318 [2024-11-27 17:37:20.172458] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:49.318 17:37:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:15:49.318 17:37:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:49.318 17:37:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:49.318 17:37:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:49.318 17:37:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:49.318 17:37:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:49.318 17:37:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:49.318 17:37:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:49.318 17:37:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:49.318 17:37:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:49.318 17:37:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:49.579 17:37:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:49.579 "name": "raid_bdev1", 00:15:49.579 "uuid": "084a4405-2241-475a-b267-f790b58fd56c", 00:15:49.579 "strip_size_kb": 64, 00:15:49.579 "state": "online", 00:15:49.579 "raid_level": "raid5f", 00:15:49.579 "superblock": true, 00:15:49.579 "num_base_bdevs": 4, 00:15:49.579 "num_base_bdevs_discovered": 4, 00:15:49.579 "num_base_bdevs_operational": 4, 00:15:49.579 "base_bdevs_list": [ 00:15:49.579 { 00:15:49.579 "name": "spare", 00:15:49.579 "uuid": "b97b13ff-990b-5307-ba00-202b7d15c61b", 00:15:49.579 "is_configured": true, 00:15:49.579 "data_offset": 2048, 00:15:49.579 "data_size": 63488 00:15:49.579 }, 00:15:49.579 { 00:15:49.579 "name": "BaseBdev2", 00:15:49.579 "uuid": "4d1cb354-a1cd-51f2-af06-696a7e2e1e92", 00:15:49.579 "is_configured": true, 00:15:49.579 "data_offset": 2048, 00:15:49.579 "data_size": 63488 00:15:49.579 }, 00:15:49.579 { 00:15:49.579 "name": "BaseBdev3", 00:15:49.579 "uuid": "caa820a2-49f8-5195-9dbe-f1fe056af663", 00:15:49.579 "is_configured": true, 00:15:49.579 "data_offset": 2048, 00:15:49.579 "data_size": 63488 00:15:49.579 }, 00:15:49.579 { 00:15:49.579 "name": "BaseBdev4", 00:15:49.579 "uuid": "887e80d8-ac51-55e0-a496-c3d1d2171574", 00:15:49.579 "is_configured": true, 00:15:49.579 "data_offset": 2048, 00:15:49.579 "data_size": 63488 00:15:49.579 } 00:15:49.579 ] 00:15:49.579 }' 00:15:49.579 17:37:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:49.579 17:37:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:15:49.579 17:37:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:49.579 17:37:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:15:49.579 17:37:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@709 -- # break 00:15:49.579 17:37:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:49.579 17:37:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:49.579 17:37:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:49.579 17:37:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:49.579 17:37:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:49.579 17:37:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:49.579 17:37:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:49.579 17:37:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:49.579 17:37:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:49.579 17:37:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:49.579 17:37:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:49.579 "name": "raid_bdev1", 00:15:49.579 "uuid": "084a4405-2241-475a-b267-f790b58fd56c", 00:15:49.579 "strip_size_kb": 64, 00:15:49.579 "state": "online", 00:15:49.579 "raid_level": "raid5f", 00:15:49.579 "superblock": true, 00:15:49.579 "num_base_bdevs": 4, 00:15:49.579 "num_base_bdevs_discovered": 4, 00:15:49.579 "num_base_bdevs_operational": 4, 00:15:49.579 "base_bdevs_list": [ 00:15:49.579 { 00:15:49.579 "name": "spare", 00:15:49.579 "uuid": "b97b13ff-990b-5307-ba00-202b7d15c61b", 00:15:49.579 "is_configured": true, 00:15:49.579 "data_offset": 2048, 00:15:49.579 "data_size": 63488 00:15:49.579 }, 00:15:49.579 { 00:15:49.579 "name": "BaseBdev2", 00:15:49.579 "uuid": "4d1cb354-a1cd-51f2-af06-696a7e2e1e92", 00:15:49.579 "is_configured": true, 00:15:49.579 "data_offset": 2048, 00:15:49.579 "data_size": 63488 00:15:49.579 }, 00:15:49.579 { 00:15:49.579 "name": "BaseBdev3", 00:15:49.579 "uuid": "caa820a2-49f8-5195-9dbe-f1fe056af663", 00:15:49.579 "is_configured": true, 00:15:49.579 "data_offset": 2048, 00:15:49.579 "data_size": 63488 00:15:49.579 }, 00:15:49.579 { 00:15:49.579 "name": "BaseBdev4", 00:15:49.579 "uuid": "887e80d8-ac51-55e0-a496-c3d1d2171574", 00:15:49.579 "is_configured": true, 00:15:49.579 "data_offset": 2048, 00:15:49.579 "data_size": 63488 00:15:49.579 } 00:15:49.579 ] 00:15:49.579 }' 00:15:49.579 17:37:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:49.579 17:37:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:49.579 17:37:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:49.579 17:37:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:15:49.579 17:37:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 4 00:15:49.579 17:37:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:49.579 17:37:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:49.579 17:37:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:49.579 17:37:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:49.579 17:37:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:49.579 17:37:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:49.579 17:37:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:49.839 17:37:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:49.839 17:37:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:49.839 17:37:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:49.839 17:37:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:49.839 17:37:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:49.839 17:37:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:49.839 17:37:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:49.839 17:37:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:49.839 "name": "raid_bdev1", 00:15:49.839 "uuid": "084a4405-2241-475a-b267-f790b58fd56c", 00:15:49.839 "strip_size_kb": 64, 00:15:49.839 "state": "online", 00:15:49.839 "raid_level": "raid5f", 00:15:49.839 "superblock": true, 00:15:49.839 "num_base_bdevs": 4, 00:15:49.839 "num_base_bdevs_discovered": 4, 00:15:49.839 "num_base_bdevs_operational": 4, 00:15:49.839 "base_bdevs_list": [ 00:15:49.839 { 00:15:49.839 "name": "spare", 00:15:49.839 "uuid": "b97b13ff-990b-5307-ba00-202b7d15c61b", 00:15:49.839 "is_configured": true, 00:15:49.839 "data_offset": 2048, 00:15:49.839 "data_size": 63488 00:15:49.839 }, 00:15:49.839 { 00:15:49.839 "name": "BaseBdev2", 00:15:49.839 "uuid": "4d1cb354-a1cd-51f2-af06-696a7e2e1e92", 00:15:49.839 "is_configured": true, 00:15:49.839 "data_offset": 2048, 00:15:49.839 "data_size": 63488 00:15:49.839 }, 00:15:49.839 { 00:15:49.839 "name": "BaseBdev3", 00:15:49.839 "uuid": "caa820a2-49f8-5195-9dbe-f1fe056af663", 00:15:49.839 "is_configured": true, 00:15:49.839 "data_offset": 2048, 00:15:49.839 "data_size": 63488 00:15:49.839 }, 00:15:49.839 { 00:15:49.839 "name": "BaseBdev4", 00:15:49.839 "uuid": "887e80d8-ac51-55e0-a496-c3d1d2171574", 00:15:49.839 "is_configured": true, 00:15:49.839 "data_offset": 2048, 00:15:49.839 "data_size": 63488 00:15:49.839 } 00:15:49.839 ] 00:15:49.839 }' 00:15:49.839 17:37:20 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:49.839 17:37:20 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:50.099 17:37:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:15:50.099 17:37:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:50.099 17:37:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:50.099 [2024-11-27 17:37:21.243561] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:15:50.099 [2024-11-27 17:37:21.243633] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:15:50.099 [2024-11-27 17:37:21.243727] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:50.099 [2024-11-27 17:37:21.243843] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:50.099 [2024-11-27 17:37:21.243899] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:15:50.099 17:37:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:50.099 17:37:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # jq length 00:15:50.099 17:37:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:50.099 17:37:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:50.099 17:37:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:50.099 17:37:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:50.359 17:37:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:15:50.359 17:37:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:15:50.359 17:37:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:15:50.359 17:37:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:15:50.359 17:37:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:15:50.359 17:37:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:15:50.359 17:37:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@10 -- # local bdev_list 00:15:50.359 17:37:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:15:50.359 17:37:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@11 -- # local nbd_list 00:15:50.359 17:37:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@12 -- # local i 00:15:50.359 17:37:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:15:50.359 17:37:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:15:50.359 17:37:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:15:50.359 /dev/nbd0 00:15:50.359 17:37:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:15:50.359 17:37:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:15:50.359 17:37:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:15:50.359 17:37:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@869 -- # local i 00:15:50.359 17:37:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:15:50.359 17:37:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:15:50.359 17:37:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:15:50.359 17:37:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@873 -- # break 00:15:50.359 17:37:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:15:50.359 17:37:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:15:50.359 17:37:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:15:50.359 1+0 records in 00:15:50.359 1+0 records out 00:15:50.359 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.00043032 s, 9.5 MB/s 00:15:50.359 17:37:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:50.359 17:37:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@886 -- # size=4096 00:15:50.359 17:37:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:50.359 17:37:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:15:50.359 17:37:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@889 -- # return 0 00:15:50.359 17:37:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:15:50.359 17:37:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:15:50.359 17:37:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:15:50.620 /dev/nbd1 00:15:50.620 17:37:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:15:50.620 17:37:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:15:50.620 17:37:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@868 -- # local nbd_name=nbd1 00:15:50.620 17:37:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@869 -- # local i 00:15:50.620 17:37:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:15:50.620 17:37:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:15:50.620 17:37:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@872 -- # grep -q -w nbd1 /proc/partitions 00:15:50.620 17:37:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@873 -- # break 00:15:50.620 17:37:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:15:50.620 17:37:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:15:50.620 17:37:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@885 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:15:50.620 1+0 records in 00:15:50.620 1+0 records out 00:15:50.620 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.00038681 s, 10.6 MB/s 00:15:50.620 17:37:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:50.620 17:37:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@886 -- # size=4096 00:15:50.620 17:37:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:15:50.620 17:37:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:15:50.620 17:37:21 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@889 -- # return 0 00:15:50.620 17:37:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:15:50.620 17:37:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:15:50.620 17:37:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@738 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:15:50.880 17:37:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:15:50.881 17:37:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:15:50.881 17:37:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:15:50.881 17:37:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@50 -- # local nbd_list 00:15:50.881 17:37:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@51 -- # local i 00:15:50.881 17:37:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:15:50.881 17:37:21 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:15:50.881 17:37:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:15:50.881 17:37:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:15:50.881 17:37:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:15:50.881 17:37:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:15:50.881 17:37:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:15:50.881 17:37:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:15:51.141 17:37:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:15:51.141 17:37:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:15:51.141 17:37:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:15:51.141 17:37:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:15:51.141 17:37:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:15:51.141 17:37:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:15:51.141 17:37:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:15:51.141 17:37:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:15:51.141 17:37:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:15:51.141 17:37:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:15:51.141 17:37:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@41 -- # break 00:15:51.141 17:37:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/nbd_common.sh@45 -- # return 0 00:15:51.141 17:37:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:15:51.141 17:37:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:15:51.141 17:37:22 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:51.141 17:37:22 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:51.141 17:37:22 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:51.141 17:37:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:15:51.141 17:37:22 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:51.141 17:37:22 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:51.141 [2024-11-27 17:37:22.327155] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:15:51.141 [2024-11-27 17:37:22.327216] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:51.141 [2024-11-27 17:37:22.327239] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000ae80 00:15:51.141 [2024-11-27 17:37:22.327251] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:51.141 [2024-11-27 17:37:22.329444] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:51.141 [2024-11-27 17:37:22.329485] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:15:51.141 [2024-11-27 17:37:22.329565] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:15:51.141 [2024-11-27 17:37:22.329612] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:51.141 [2024-11-27 17:37:22.329718] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:15:51.141 [2024-11-27 17:37:22.329824] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev3 is claimed 00:15:51.141 [2024-11-27 17:37:22.329915] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev4 is claimed 00:15:51.402 spare 00:15:51.402 17:37:22 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:51.402 17:37:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:15:51.402 17:37:22 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:51.402 17:37:22 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:51.402 [2024-11-27 17:37:22.429804] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001580 00:15:51.402 [2024-11-27 17:37:22.429830] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 190464, blocklen 512 00:15:51.402 [2024-11-27 17:37:22.430086] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000045820 00:15:51.402 [2024-11-27 17:37:22.430558] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001580 00:15:51.402 [2024-11-27 17:37:22.430586] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001580 00:15:51.402 [2024-11-27 17:37:22.430716] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:51.402 17:37:22 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:51.402 17:37:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 4 00:15:51.402 17:37:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:51.402 17:37:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:51.402 17:37:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:51.402 17:37:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:51.402 17:37:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=4 00:15:51.402 17:37:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:51.402 17:37:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:51.402 17:37:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:51.402 17:37:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:51.402 17:37:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:51.402 17:37:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:51.402 17:37:22 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:51.402 17:37:22 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:51.402 17:37:22 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:51.402 17:37:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:51.402 "name": "raid_bdev1", 00:15:51.402 "uuid": "084a4405-2241-475a-b267-f790b58fd56c", 00:15:51.402 "strip_size_kb": 64, 00:15:51.402 "state": "online", 00:15:51.402 "raid_level": "raid5f", 00:15:51.402 "superblock": true, 00:15:51.402 "num_base_bdevs": 4, 00:15:51.402 "num_base_bdevs_discovered": 4, 00:15:51.402 "num_base_bdevs_operational": 4, 00:15:51.402 "base_bdevs_list": [ 00:15:51.402 { 00:15:51.402 "name": "spare", 00:15:51.402 "uuid": "b97b13ff-990b-5307-ba00-202b7d15c61b", 00:15:51.402 "is_configured": true, 00:15:51.402 "data_offset": 2048, 00:15:51.402 "data_size": 63488 00:15:51.402 }, 00:15:51.402 { 00:15:51.402 "name": "BaseBdev2", 00:15:51.402 "uuid": "4d1cb354-a1cd-51f2-af06-696a7e2e1e92", 00:15:51.402 "is_configured": true, 00:15:51.402 "data_offset": 2048, 00:15:51.402 "data_size": 63488 00:15:51.402 }, 00:15:51.402 { 00:15:51.402 "name": "BaseBdev3", 00:15:51.402 "uuid": "caa820a2-49f8-5195-9dbe-f1fe056af663", 00:15:51.402 "is_configured": true, 00:15:51.402 "data_offset": 2048, 00:15:51.402 "data_size": 63488 00:15:51.402 }, 00:15:51.402 { 00:15:51.402 "name": "BaseBdev4", 00:15:51.402 "uuid": "887e80d8-ac51-55e0-a496-c3d1d2171574", 00:15:51.402 "is_configured": true, 00:15:51.402 "data_offset": 2048, 00:15:51.402 "data_size": 63488 00:15:51.402 } 00:15:51.402 ] 00:15:51.402 }' 00:15:51.402 17:37:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:51.402 17:37:22 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:51.972 17:37:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:51.973 17:37:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:51.973 17:37:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:51.973 17:37:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:51.973 17:37:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:51.973 17:37:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:51.973 17:37:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:51.973 17:37:22 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:51.973 17:37:22 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:51.973 17:37:22 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:51.973 17:37:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:51.973 "name": "raid_bdev1", 00:15:51.973 "uuid": "084a4405-2241-475a-b267-f790b58fd56c", 00:15:51.973 "strip_size_kb": 64, 00:15:51.973 "state": "online", 00:15:51.973 "raid_level": "raid5f", 00:15:51.973 "superblock": true, 00:15:51.973 "num_base_bdevs": 4, 00:15:51.973 "num_base_bdevs_discovered": 4, 00:15:51.973 "num_base_bdevs_operational": 4, 00:15:51.973 "base_bdevs_list": [ 00:15:51.973 { 00:15:51.973 "name": "spare", 00:15:51.973 "uuid": "b97b13ff-990b-5307-ba00-202b7d15c61b", 00:15:51.973 "is_configured": true, 00:15:51.973 "data_offset": 2048, 00:15:51.973 "data_size": 63488 00:15:51.973 }, 00:15:51.973 { 00:15:51.973 "name": "BaseBdev2", 00:15:51.973 "uuid": "4d1cb354-a1cd-51f2-af06-696a7e2e1e92", 00:15:51.973 "is_configured": true, 00:15:51.973 "data_offset": 2048, 00:15:51.973 "data_size": 63488 00:15:51.973 }, 00:15:51.973 { 00:15:51.973 "name": "BaseBdev3", 00:15:51.973 "uuid": "caa820a2-49f8-5195-9dbe-f1fe056af663", 00:15:51.973 "is_configured": true, 00:15:51.973 "data_offset": 2048, 00:15:51.973 "data_size": 63488 00:15:51.973 }, 00:15:51.973 { 00:15:51.973 "name": "BaseBdev4", 00:15:51.973 "uuid": "887e80d8-ac51-55e0-a496-c3d1d2171574", 00:15:51.973 "is_configured": true, 00:15:51.973 "data_offset": 2048, 00:15:51.973 "data_size": 63488 00:15:51.973 } 00:15:51.973 ] 00:15:51.973 }' 00:15:51.973 17:37:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:51.973 17:37:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:51.973 17:37:22 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:51.973 17:37:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:15:51.973 17:37:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:51.973 17:37:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:15:51.973 17:37:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:51.973 17:37:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:51.973 17:37:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:51.973 17:37:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:15:51.973 17:37:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:15:51.973 17:37:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:51.973 17:37:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:51.973 [2024-11-27 17:37:23.103167] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:51.973 17:37:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:51.973 17:37:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:15:51.973 17:37:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:51.973 17:37:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:51.973 17:37:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:51.973 17:37:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:51.973 17:37:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:15:51.973 17:37:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:51.973 17:37:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:51.973 17:37:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:51.973 17:37:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:51.973 17:37:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:51.973 17:37:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:51.973 17:37:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:51.973 17:37:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:51.973 17:37:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:51.973 17:37:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:51.973 "name": "raid_bdev1", 00:15:51.973 "uuid": "084a4405-2241-475a-b267-f790b58fd56c", 00:15:51.973 "strip_size_kb": 64, 00:15:51.973 "state": "online", 00:15:51.973 "raid_level": "raid5f", 00:15:51.973 "superblock": true, 00:15:51.973 "num_base_bdevs": 4, 00:15:51.973 "num_base_bdevs_discovered": 3, 00:15:51.973 "num_base_bdevs_operational": 3, 00:15:51.973 "base_bdevs_list": [ 00:15:51.973 { 00:15:51.973 "name": null, 00:15:51.973 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:51.973 "is_configured": false, 00:15:51.973 "data_offset": 0, 00:15:51.973 "data_size": 63488 00:15:51.973 }, 00:15:51.973 { 00:15:51.973 "name": "BaseBdev2", 00:15:51.973 "uuid": "4d1cb354-a1cd-51f2-af06-696a7e2e1e92", 00:15:51.973 "is_configured": true, 00:15:51.973 "data_offset": 2048, 00:15:51.973 "data_size": 63488 00:15:51.973 }, 00:15:51.973 { 00:15:51.973 "name": "BaseBdev3", 00:15:51.973 "uuid": "caa820a2-49f8-5195-9dbe-f1fe056af663", 00:15:51.973 "is_configured": true, 00:15:51.973 "data_offset": 2048, 00:15:51.973 "data_size": 63488 00:15:51.973 }, 00:15:51.973 { 00:15:51.973 "name": "BaseBdev4", 00:15:51.973 "uuid": "887e80d8-ac51-55e0-a496-c3d1d2171574", 00:15:51.973 "is_configured": true, 00:15:51.973 "data_offset": 2048, 00:15:51.973 "data_size": 63488 00:15:51.973 } 00:15:51.973 ] 00:15:51.973 }' 00:15:51.973 17:37:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:51.973 17:37:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:52.544 17:37:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:15:52.544 17:37:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:52.544 17:37:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:52.544 [2024-11-27 17:37:23.558373] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:52.544 [2024-11-27 17:37:23.558543] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:15:52.544 [2024-11-27 17:37:23.558557] bdev_raid.c:3748:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:15:52.544 [2024-11-27 17:37:23.558613] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:52.544 [2024-11-27 17:37:23.561838] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000458f0 00:15:52.544 [2024-11-27 17:37:23.563985] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:15:52.544 17:37:23 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:52.544 17:37:23 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@757 -- # sleep 1 00:15:53.486 17:37:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:53.486 17:37:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:53.486 17:37:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:53.486 17:37:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:53.486 17:37:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:53.486 17:37:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:53.486 17:37:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:53.486 17:37:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:53.486 17:37:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:53.486 17:37:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:53.486 17:37:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:53.486 "name": "raid_bdev1", 00:15:53.486 "uuid": "084a4405-2241-475a-b267-f790b58fd56c", 00:15:53.486 "strip_size_kb": 64, 00:15:53.486 "state": "online", 00:15:53.486 "raid_level": "raid5f", 00:15:53.486 "superblock": true, 00:15:53.486 "num_base_bdevs": 4, 00:15:53.486 "num_base_bdevs_discovered": 4, 00:15:53.486 "num_base_bdevs_operational": 4, 00:15:53.486 "process": { 00:15:53.486 "type": "rebuild", 00:15:53.486 "target": "spare", 00:15:53.486 "progress": { 00:15:53.486 "blocks": 19200, 00:15:53.486 "percent": 10 00:15:53.486 } 00:15:53.486 }, 00:15:53.486 "base_bdevs_list": [ 00:15:53.486 { 00:15:53.486 "name": "spare", 00:15:53.486 "uuid": "b97b13ff-990b-5307-ba00-202b7d15c61b", 00:15:53.486 "is_configured": true, 00:15:53.486 "data_offset": 2048, 00:15:53.486 "data_size": 63488 00:15:53.486 }, 00:15:53.486 { 00:15:53.486 "name": "BaseBdev2", 00:15:53.486 "uuid": "4d1cb354-a1cd-51f2-af06-696a7e2e1e92", 00:15:53.486 "is_configured": true, 00:15:53.486 "data_offset": 2048, 00:15:53.486 "data_size": 63488 00:15:53.486 }, 00:15:53.486 { 00:15:53.486 "name": "BaseBdev3", 00:15:53.486 "uuid": "caa820a2-49f8-5195-9dbe-f1fe056af663", 00:15:53.486 "is_configured": true, 00:15:53.486 "data_offset": 2048, 00:15:53.486 "data_size": 63488 00:15:53.486 }, 00:15:53.486 { 00:15:53.486 "name": "BaseBdev4", 00:15:53.486 "uuid": "887e80d8-ac51-55e0-a496-c3d1d2171574", 00:15:53.486 "is_configured": true, 00:15:53.486 "data_offset": 2048, 00:15:53.486 "data_size": 63488 00:15:53.486 } 00:15:53.486 ] 00:15:53.486 }' 00:15:53.486 17:37:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:53.486 17:37:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:53.486 17:37:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:53.746 17:37:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:53.746 17:37:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:15:53.746 17:37:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:53.746 17:37:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:53.746 [2024-11-27 17:37:24.722903] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:53.746 [2024-11-27 17:37:24.769386] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:15:53.746 [2024-11-27 17:37:24.769443] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:53.746 [2024-11-27 17:37:24.769464] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:53.746 [2024-11-27 17:37:24.769473] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:15:53.746 17:37:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:53.746 17:37:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:15:53.746 17:37:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:53.746 17:37:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:53.746 17:37:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:53.746 17:37:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:53.746 17:37:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:15:53.746 17:37:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:53.746 17:37:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:53.746 17:37:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:53.746 17:37:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:53.746 17:37:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:53.746 17:37:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:53.746 17:37:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:53.746 17:37:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:53.746 17:37:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:53.746 17:37:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:53.746 "name": "raid_bdev1", 00:15:53.746 "uuid": "084a4405-2241-475a-b267-f790b58fd56c", 00:15:53.746 "strip_size_kb": 64, 00:15:53.746 "state": "online", 00:15:53.746 "raid_level": "raid5f", 00:15:53.746 "superblock": true, 00:15:53.746 "num_base_bdevs": 4, 00:15:53.746 "num_base_bdevs_discovered": 3, 00:15:53.746 "num_base_bdevs_operational": 3, 00:15:53.746 "base_bdevs_list": [ 00:15:53.746 { 00:15:53.746 "name": null, 00:15:53.746 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:53.746 "is_configured": false, 00:15:53.746 "data_offset": 0, 00:15:53.746 "data_size": 63488 00:15:53.746 }, 00:15:53.746 { 00:15:53.746 "name": "BaseBdev2", 00:15:53.746 "uuid": "4d1cb354-a1cd-51f2-af06-696a7e2e1e92", 00:15:53.746 "is_configured": true, 00:15:53.746 "data_offset": 2048, 00:15:53.746 "data_size": 63488 00:15:53.746 }, 00:15:53.746 { 00:15:53.746 "name": "BaseBdev3", 00:15:53.746 "uuid": "caa820a2-49f8-5195-9dbe-f1fe056af663", 00:15:53.746 "is_configured": true, 00:15:53.746 "data_offset": 2048, 00:15:53.746 "data_size": 63488 00:15:53.746 }, 00:15:53.746 { 00:15:53.746 "name": "BaseBdev4", 00:15:53.746 "uuid": "887e80d8-ac51-55e0-a496-c3d1d2171574", 00:15:53.746 "is_configured": true, 00:15:53.746 "data_offset": 2048, 00:15:53.746 "data_size": 63488 00:15:53.746 } 00:15:53.746 ] 00:15:53.746 }' 00:15:53.746 17:37:24 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:53.746 17:37:24 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:54.316 17:37:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:15:54.316 17:37:25 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:54.316 17:37:25 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:54.316 [2024-11-27 17:37:25.205478] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:15:54.316 [2024-11-27 17:37:25.205539] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:54.316 [2024-11-27 17:37:25.205565] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000b780 00:15:54.316 [2024-11-27 17:37:25.205574] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:54.316 [2024-11-27 17:37:25.206005] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:54.316 [2024-11-27 17:37:25.206030] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:15:54.316 [2024-11-27 17:37:25.206121] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:15:54.316 [2024-11-27 17:37:25.206148] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:15:54.316 [2024-11-27 17:37:25.206164] bdev_raid.c:3748:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:15:54.316 [2024-11-27 17:37:25.206188] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:15:54.316 [2024-11-27 17:37:25.209483] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000459c0 00:15:54.316 spare 00:15:54.316 17:37:25 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:54.316 17:37:25 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@764 -- # sleep 1 00:15:54.316 [2024-11-27 17:37:25.211584] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:15:55.257 17:37:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:15:55.257 17:37:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:55.257 17:37:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:15:55.257 17:37:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=spare 00:15:55.257 17:37:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:55.257 17:37:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:55.257 17:37:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:55.257 17:37:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:55.257 17:37:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:55.257 17:37:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:55.257 17:37:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:55.257 "name": "raid_bdev1", 00:15:55.257 "uuid": "084a4405-2241-475a-b267-f790b58fd56c", 00:15:55.257 "strip_size_kb": 64, 00:15:55.257 "state": "online", 00:15:55.257 "raid_level": "raid5f", 00:15:55.257 "superblock": true, 00:15:55.257 "num_base_bdevs": 4, 00:15:55.257 "num_base_bdevs_discovered": 4, 00:15:55.257 "num_base_bdevs_operational": 4, 00:15:55.257 "process": { 00:15:55.257 "type": "rebuild", 00:15:55.257 "target": "spare", 00:15:55.257 "progress": { 00:15:55.257 "blocks": 19200, 00:15:55.257 "percent": 10 00:15:55.257 } 00:15:55.257 }, 00:15:55.257 "base_bdevs_list": [ 00:15:55.257 { 00:15:55.257 "name": "spare", 00:15:55.257 "uuid": "b97b13ff-990b-5307-ba00-202b7d15c61b", 00:15:55.257 "is_configured": true, 00:15:55.257 "data_offset": 2048, 00:15:55.257 "data_size": 63488 00:15:55.257 }, 00:15:55.257 { 00:15:55.257 "name": "BaseBdev2", 00:15:55.257 "uuid": "4d1cb354-a1cd-51f2-af06-696a7e2e1e92", 00:15:55.257 "is_configured": true, 00:15:55.257 "data_offset": 2048, 00:15:55.257 "data_size": 63488 00:15:55.257 }, 00:15:55.257 { 00:15:55.257 "name": "BaseBdev3", 00:15:55.257 "uuid": "caa820a2-49f8-5195-9dbe-f1fe056af663", 00:15:55.257 "is_configured": true, 00:15:55.257 "data_offset": 2048, 00:15:55.257 "data_size": 63488 00:15:55.257 }, 00:15:55.257 { 00:15:55.257 "name": "BaseBdev4", 00:15:55.257 "uuid": "887e80d8-ac51-55e0-a496-c3d1d2171574", 00:15:55.257 "is_configured": true, 00:15:55.257 "data_offset": 2048, 00:15:55.257 "data_size": 63488 00:15:55.257 } 00:15:55.257 ] 00:15:55.257 }' 00:15:55.257 17:37:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:55.257 17:37:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:15:55.257 17:37:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:55.257 17:37:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:15:55.257 17:37:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:15:55.257 17:37:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:55.257 17:37:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:55.257 [2024-11-27 17:37:26.348279] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:55.257 [2024-11-27 17:37:26.416766] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:15:55.257 [2024-11-27 17:37:26.416837] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:15:55.257 [2024-11-27 17:37:26.416853] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:15:55.257 [2024-11-27 17:37:26.416864] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:15:55.257 17:37:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:55.257 17:37:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:15:55.257 17:37:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:55.257 17:37:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:55.257 17:37:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:55.257 17:37:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:55.257 17:37:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:15:55.257 17:37:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:55.257 17:37:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:55.258 17:37:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:55.258 17:37:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:55.258 17:37:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:55.258 17:37:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:55.258 17:37:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:55.258 17:37:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:55.517 17:37:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:55.517 17:37:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:55.517 "name": "raid_bdev1", 00:15:55.517 "uuid": "084a4405-2241-475a-b267-f790b58fd56c", 00:15:55.517 "strip_size_kb": 64, 00:15:55.517 "state": "online", 00:15:55.517 "raid_level": "raid5f", 00:15:55.517 "superblock": true, 00:15:55.517 "num_base_bdevs": 4, 00:15:55.517 "num_base_bdevs_discovered": 3, 00:15:55.517 "num_base_bdevs_operational": 3, 00:15:55.517 "base_bdevs_list": [ 00:15:55.517 { 00:15:55.517 "name": null, 00:15:55.517 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:55.517 "is_configured": false, 00:15:55.517 "data_offset": 0, 00:15:55.517 "data_size": 63488 00:15:55.517 }, 00:15:55.517 { 00:15:55.517 "name": "BaseBdev2", 00:15:55.517 "uuid": "4d1cb354-a1cd-51f2-af06-696a7e2e1e92", 00:15:55.517 "is_configured": true, 00:15:55.517 "data_offset": 2048, 00:15:55.517 "data_size": 63488 00:15:55.517 }, 00:15:55.517 { 00:15:55.517 "name": "BaseBdev3", 00:15:55.517 "uuid": "caa820a2-49f8-5195-9dbe-f1fe056af663", 00:15:55.517 "is_configured": true, 00:15:55.517 "data_offset": 2048, 00:15:55.517 "data_size": 63488 00:15:55.517 }, 00:15:55.517 { 00:15:55.517 "name": "BaseBdev4", 00:15:55.517 "uuid": "887e80d8-ac51-55e0-a496-c3d1d2171574", 00:15:55.517 "is_configured": true, 00:15:55.517 "data_offset": 2048, 00:15:55.517 "data_size": 63488 00:15:55.517 } 00:15:55.517 ] 00:15:55.517 }' 00:15:55.517 17:37:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:55.517 17:37:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:55.778 17:37:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:55.778 17:37:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:55.778 17:37:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:55.778 17:37:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:55.778 17:37:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:55.778 17:37:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:55.778 17:37:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:55.778 17:37:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:55.778 17:37:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:55.778 17:37:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:55.778 17:37:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:55.778 "name": "raid_bdev1", 00:15:55.778 "uuid": "084a4405-2241-475a-b267-f790b58fd56c", 00:15:55.778 "strip_size_kb": 64, 00:15:55.778 "state": "online", 00:15:55.778 "raid_level": "raid5f", 00:15:55.778 "superblock": true, 00:15:55.778 "num_base_bdevs": 4, 00:15:55.778 "num_base_bdevs_discovered": 3, 00:15:55.778 "num_base_bdevs_operational": 3, 00:15:55.778 "base_bdevs_list": [ 00:15:55.778 { 00:15:55.778 "name": null, 00:15:55.778 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:55.778 "is_configured": false, 00:15:55.778 "data_offset": 0, 00:15:55.778 "data_size": 63488 00:15:55.778 }, 00:15:55.778 { 00:15:55.778 "name": "BaseBdev2", 00:15:55.778 "uuid": "4d1cb354-a1cd-51f2-af06-696a7e2e1e92", 00:15:55.778 "is_configured": true, 00:15:55.778 "data_offset": 2048, 00:15:55.778 "data_size": 63488 00:15:55.778 }, 00:15:55.778 { 00:15:55.778 "name": "BaseBdev3", 00:15:55.778 "uuid": "caa820a2-49f8-5195-9dbe-f1fe056af663", 00:15:55.778 "is_configured": true, 00:15:55.778 "data_offset": 2048, 00:15:55.778 "data_size": 63488 00:15:55.778 }, 00:15:55.778 { 00:15:55.778 "name": "BaseBdev4", 00:15:55.778 "uuid": "887e80d8-ac51-55e0-a496-c3d1d2171574", 00:15:55.778 "is_configured": true, 00:15:55.778 "data_offset": 2048, 00:15:55.778 "data_size": 63488 00:15:55.778 } 00:15:55.778 ] 00:15:55.778 }' 00:15:55.778 17:37:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:55.778 17:37:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:55.778 17:37:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:56.038 17:37:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:15:56.038 17:37:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:15:56.038 17:37:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:56.038 17:37:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:56.038 17:37:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:56.038 17:37:26 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:15:56.038 17:37:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:56.038 17:37:26 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:56.038 [2024-11-27 17:37:27.004562] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:15:56.038 [2024-11-27 17:37:27.004611] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:15:56.038 [2024-11-27 17:37:27.004630] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000bd80 00:15:56.038 [2024-11-27 17:37:27.004641] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:15:56.038 [2024-11-27 17:37:27.005029] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:15:56.038 [2024-11-27 17:37:27.005056] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:15:56.038 [2024-11-27 17:37:27.005120] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:15:56.038 [2024-11-27 17:37:27.005165] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:15:56.038 [2024-11-27 17:37:27.005179] bdev_raid.c:3709:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:15:56.038 [2024-11-27 17:37:27.005191] bdev_raid.c:3884:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:15:56.038 BaseBdev1 00:15:56.038 17:37:27 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:56.038 17:37:27 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@775 -- # sleep 1 00:15:56.978 17:37:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:15:56.978 17:37:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:56.978 17:37:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:56.978 17:37:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:56.978 17:37:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:56.978 17:37:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:15:56.978 17:37:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:56.978 17:37:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:56.978 17:37:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:56.978 17:37:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:56.978 17:37:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:56.978 17:37:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:56.978 17:37:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:56.978 17:37:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:56.978 17:37:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:56.978 17:37:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:56.978 "name": "raid_bdev1", 00:15:56.978 "uuid": "084a4405-2241-475a-b267-f790b58fd56c", 00:15:56.978 "strip_size_kb": 64, 00:15:56.978 "state": "online", 00:15:56.978 "raid_level": "raid5f", 00:15:56.978 "superblock": true, 00:15:56.978 "num_base_bdevs": 4, 00:15:56.978 "num_base_bdevs_discovered": 3, 00:15:56.978 "num_base_bdevs_operational": 3, 00:15:56.978 "base_bdevs_list": [ 00:15:56.978 { 00:15:56.978 "name": null, 00:15:56.978 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:56.978 "is_configured": false, 00:15:56.978 "data_offset": 0, 00:15:56.978 "data_size": 63488 00:15:56.978 }, 00:15:56.978 { 00:15:56.978 "name": "BaseBdev2", 00:15:56.978 "uuid": "4d1cb354-a1cd-51f2-af06-696a7e2e1e92", 00:15:56.978 "is_configured": true, 00:15:56.978 "data_offset": 2048, 00:15:56.978 "data_size": 63488 00:15:56.978 }, 00:15:56.978 { 00:15:56.978 "name": "BaseBdev3", 00:15:56.978 "uuid": "caa820a2-49f8-5195-9dbe-f1fe056af663", 00:15:56.978 "is_configured": true, 00:15:56.978 "data_offset": 2048, 00:15:56.978 "data_size": 63488 00:15:56.978 }, 00:15:56.978 { 00:15:56.978 "name": "BaseBdev4", 00:15:56.978 "uuid": "887e80d8-ac51-55e0-a496-c3d1d2171574", 00:15:56.978 "is_configured": true, 00:15:56.978 "data_offset": 2048, 00:15:56.978 "data_size": 63488 00:15:56.978 } 00:15:56.978 ] 00:15:56.978 }' 00:15:56.978 17:37:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:56.978 17:37:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:57.547 17:37:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:57.547 17:37:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:57.547 17:37:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:57.547 17:37:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:57.548 17:37:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:57.548 17:37:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:57.548 17:37:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:57.548 17:37:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:57.548 17:37:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:57.548 17:37:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:57.548 17:37:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:57.548 "name": "raid_bdev1", 00:15:57.548 "uuid": "084a4405-2241-475a-b267-f790b58fd56c", 00:15:57.548 "strip_size_kb": 64, 00:15:57.548 "state": "online", 00:15:57.548 "raid_level": "raid5f", 00:15:57.548 "superblock": true, 00:15:57.548 "num_base_bdevs": 4, 00:15:57.548 "num_base_bdevs_discovered": 3, 00:15:57.548 "num_base_bdevs_operational": 3, 00:15:57.548 "base_bdevs_list": [ 00:15:57.548 { 00:15:57.548 "name": null, 00:15:57.548 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:57.548 "is_configured": false, 00:15:57.548 "data_offset": 0, 00:15:57.548 "data_size": 63488 00:15:57.548 }, 00:15:57.548 { 00:15:57.548 "name": "BaseBdev2", 00:15:57.548 "uuid": "4d1cb354-a1cd-51f2-af06-696a7e2e1e92", 00:15:57.548 "is_configured": true, 00:15:57.548 "data_offset": 2048, 00:15:57.548 "data_size": 63488 00:15:57.548 }, 00:15:57.548 { 00:15:57.548 "name": "BaseBdev3", 00:15:57.548 "uuid": "caa820a2-49f8-5195-9dbe-f1fe056af663", 00:15:57.548 "is_configured": true, 00:15:57.548 "data_offset": 2048, 00:15:57.548 "data_size": 63488 00:15:57.548 }, 00:15:57.548 { 00:15:57.548 "name": "BaseBdev4", 00:15:57.548 "uuid": "887e80d8-ac51-55e0-a496-c3d1d2171574", 00:15:57.548 "is_configured": true, 00:15:57.548 "data_offset": 2048, 00:15:57.548 "data_size": 63488 00:15:57.548 } 00:15:57.548 ] 00:15:57.548 }' 00:15:57.548 17:37:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:57.548 17:37:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:57.548 17:37:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:57.548 17:37:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:15:57.548 17:37:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:15:57.548 17:37:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@650 -- # local es=0 00:15:57.548 17:37:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:15:57.548 17:37:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:15:57.548 17:37:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:15:57.548 17:37:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:15:57.548 17:37:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:15:57.548 17:37:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:15:57.548 17:37:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:57.548 17:37:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:57.548 [2024-11-27 17:37:28.614128] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:15:57.548 [2024-11-27 17:37:28.614264] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:15:57.548 [2024-11-27 17:37:28.614282] bdev_raid.c:3709:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:15:57.548 request: 00:15:57.548 { 00:15:57.548 "base_bdev": "BaseBdev1", 00:15:57.548 "raid_bdev": "raid_bdev1", 00:15:57.548 "method": "bdev_raid_add_base_bdev", 00:15:57.548 "req_id": 1 00:15:57.548 } 00:15:57.548 Got JSON-RPC error response 00:15:57.548 response: 00:15:57.548 { 00:15:57.548 "code": -22, 00:15:57.548 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:15:57.548 } 00:15:57.548 17:37:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:15:57.548 17:37:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@653 -- # es=1 00:15:57.548 17:37:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:15:57.548 17:37:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:15:57.548 17:37:28 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:15:57.548 17:37:28 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@779 -- # sleep 1 00:15:58.486 17:37:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid5f 64 3 00:15:58.486 17:37:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:15:58.486 17:37:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:15:58.486 17:37:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@105 -- # local raid_level=raid5f 00:15:58.486 17:37:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@106 -- # local strip_size=64 00:15:58.486 17:37:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=3 00:15:58.486 17:37:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:15:58.486 17:37:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:15:58.486 17:37:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:15:58.486 17:37:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@111 -- # local tmp 00:15:58.486 17:37:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:58.486 17:37:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:58.486 17:37:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:58.486 17:37:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:58.486 17:37:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:58.746 17:37:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:15:58.746 "name": "raid_bdev1", 00:15:58.746 "uuid": "084a4405-2241-475a-b267-f790b58fd56c", 00:15:58.746 "strip_size_kb": 64, 00:15:58.746 "state": "online", 00:15:58.746 "raid_level": "raid5f", 00:15:58.746 "superblock": true, 00:15:58.746 "num_base_bdevs": 4, 00:15:58.746 "num_base_bdevs_discovered": 3, 00:15:58.746 "num_base_bdevs_operational": 3, 00:15:58.746 "base_bdevs_list": [ 00:15:58.746 { 00:15:58.746 "name": null, 00:15:58.746 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:58.746 "is_configured": false, 00:15:58.746 "data_offset": 0, 00:15:58.746 "data_size": 63488 00:15:58.746 }, 00:15:58.746 { 00:15:58.746 "name": "BaseBdev2", 00:15:58.746 "uuid": "4d1cb354-a1cd-51f2-af06-696a7e2e1e92", 00:15:58.746 "is_configured": true, 00:15:58.746 "data_offset": 2048, 00:15:58.746 "data_size": 63488 00:15:58.746 }, 00:15:58.746 { 00:15:58.746 "name": "BaseBdev3", 00:15:58.746 "uuid": "caa820a2-49f8-5195-9dbe-f1fe056af663", 00:15:58.746 "is_configured": true, 00:15:58.746 "data_offset": 2048, 00:15:58.746 "data_size": 63488 00:15:58.746 }, 00:15:58.746 { 00:15:58.746 "name": "BaseBdev4", 00:15:58.746 "uuid": "887e80d8-ac51-55e0-a496-c3d1d2171574", 00:15:58.746 "is_configured": true, 00:15:58.746 "data_offset": 2048, 00:15:58.746 "data_size": 63488 00:15:58.746 } 00:15:58.746 ] 00:15:58.746 }' 00:15:58.746 17:37:29 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:15:58.746 17:37:29 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:59.006 17:37:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:15:59.006 17:37:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:15:59.006 17:37:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:15:59.006 17:37:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@171 -- # local target=none 00:15:59.006 17:37:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:15:59.006 17:37:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:15:59.006 17:37:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:15:59.006 17:37:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:59.006 17:37:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:59.006 17:37:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:59.006 17:37:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:15:59.006 "name": "raid_bdev1", 00:15:59.006 "uuid": "084a4405-2241-475a-b267-f790b58fd56c", 00:15:59.006 "strip_size_kb": 64, 00:15:59.006 "state": "online", 00:15:59.006 "raid_level": "raid5f", 00:15:59.006 "superblock": true, 00:15:59.006 "num_base_bdevs": 4, 00:15:59.006 "num_base_bdevs_discovered": 3, 00:15:59.006 "num_base_bdevs_operational": 3, 00:15:59.006 "base_bdevs_list": [ 00:15:59.006 { 00:15:59.006 "name": null, 00:15:59.006 "uuid": "00000000-0000-0000-0000-000000000000", 00:15:59.006 "is_configured": false, 00:15:59.006 "data_offset": 0, 00:15:59.006 "data_size": 63488 00:15:59.006 }, 00:15:59.006 { 00:15:59.006 "name": "BaseBdev2", 00:15:59.006 "uuid": "4d1cb354-a1cd-51f2-af06-696a7e2e1e92", 00:15:59.006 "is_configured": true, 00:15:59.006 "data_offset": 2048, 00:15:59.006 "data_size": 63488 00:15:59.006 }, 00:15:59.006 { 00:15:59.006 "name": "BaseBdev3", 00:15:59.006 "uuid": "caa820a2-49f8-5195-9dbe-f1fe056af663", 00:15:59.006 "is_configured": true, 00:15:59.006 "data_offset": 2048, 00:15:59.006 "data_size": 63488 00:15:59.006 }, 00:15:59.006 { 00:15:59.006 "name": "BaseBdev4", 00:15:59.006 "uuid": "887e80d8-ac51-55e0-a496-c3d1d2171574", 00:15:59.006 "is_configured": true, 00:15:59.006 "data_offset": 2048, 00:15:59.006 "data_size": 63488 00:15:59.006 } 00:15:59.006 ] 00:15:59.006 }' 00:15:59.006 17:37:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:15:59.006 17:37:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:15:59.006 17:37:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:15:59.006 17:37:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:15:59.006 17:37:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@784 -- # killprocess 95364 00:15:59.006 17:37:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@950 -- # '[' -z 95364 ']' 00:15:59.006 17:37:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@954 -- # kill -0 95364 00:15:59.006 17:37:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@955 -- # uname 00:15:59.006 17:37:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:15:59.006 17:37:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 95364 00:15:59.006 17:37:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:15:59.006 17:37:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:15:59.006 killing process with pid 95364 00:15:59.006 17:37:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@968 -- # echo 'killing process with pid 95364' 00:15:59.006 17:37:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@969 -- # kill 95364 00:15:59.006 Received shutdown signal, test time was about 60.000000 seconds 00:15:59.006 00:15:59.006 Latency(us) 00:15:59.006 [2024-11-27T17:37:30.198Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:15:59.006 [2024-11-27T17:37:30.198Z] =================================================================================================================== 00:15:59.006 [2024-11-27T17:37:30.198Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:15:59.006 [2024-11-27 17:37:30.194010] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:15:59.006 [2024-11-27 17:37:30.194147] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:15:59.006 17:37:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@974 -- # wait 95364 00:15:59.006 [2024-11-27 17:37:30.194226] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:15:59.006 [2024-11-27 17:37:30.194235] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state offline 00:15:59.266 [2024-11-27 17:37:30.243626] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:15:59.526 17:37:30 bdev_raid.raid5f_rebuild_test_sb -- bdev/bdev_raid.sh@786 -- # return 0 00:15:59.526 00:15:59.526 real 0m25.328s 00:15:59.526 user 0m32.060s 00:15:59.526 sys 0m3.201s 00:15:59.526 17:37:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@1126 -- # xtrace_disable 00:15:59.526 17:37:30 bdev_raid.raid5f_rebuild_test_sb -- common/autotest_common.sh@10 -- # set +x 00:15:59.526 ************************************ 00:15:59.526 END TEST raid5f_rebuild_test_sb 00:15:59.526 ************************************ 00:15:59.526 17:37:30 bdev_raid -- bdev/bdev_raid.sh@995 -- # base_blocklen=4096 00:15:59.526 17:37:30 bdev_raid -- bdev/bdev_raid.sh@997 -- # run_test raid_state_function_test_sb_4k raid_state_function_test raid1 2 true 00:15:59.526 17:37:30 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:15:59.526 17:37:30 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:15:59.526 17:37:30 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:15:59.526 ************************************ 00:15:59.526 START TEST raid_state_function_test_sb_4k 00:15:59.526 ************************************ 00:15:59.526 17:37:30 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@1125 -- # raid_state_function_test raid1 2 true 00:15:59.526 17:37:30 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:15:59.526 17:37:30 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:15:59.526 17:37:30 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:15:59.526 17:37:30 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:15:59.526 17:37:30 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:15:59.526 17:37:30 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:15:59.526 17:37:30 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:15:59.526 17:37:30 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:15:59.526 17:37:30 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:15:59.526 17:37:30 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:15:59.526 17:37:30 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:15:59.526 17:37:30 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:15:59.526 17:37:30 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:15:59.526 17:37:30 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:15:59.526 17:37:30 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:15:59.526 17:37:30 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@211 -- # local strip_size 00:15:59.526 17:37:30 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:15:59.526 17:37:30 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:15:59.526 17:37:30 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:15:59.526 17:37:30 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:15:59.526 17:37:30 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:15:59.526 17:37:30 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:15:59.526 17:37:30 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@229 -- # raid_pid=96170 00:15:59.526 17:37:30 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:15:59.526 Process raid pid: 96170 00:15:59.526 17:37:30 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 96170' 00:15:59.526 17:37:30 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@231 -- # waitforlisten 96170 00:15:59.526 17:37:30 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@831 -- # '[' -z 96170 ']' 00:15:59.526 17:37:30 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:15:59.526 17:37:30 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@836 -- # local max_retries=100 00:15:59.526 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:15:59.526 17:37:30 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:15:59.526 17:37:30 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@840 -- # xtrace_disable 00:15:59.526 17:37:30 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:15:59.526 [2024-11-27 17:37:30.636241] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:15:59.526 [2024-11-27 17:37:30.636396] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:15:59.786 [2024-11-27 17:37:30.761545] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:15:59.786 [2024-11-27 17:37:30.804862] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:15:59.786 [2024-11-27 17:37:30.846497] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:15:59.786 [2024-11-27 17:37:30.846536] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:16:00.505 17:37:31 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:16:00.505 17:37:31 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@864 -- # return 0 00:16:00.505 17:37:31 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:16:00.505 17:37:31 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:00.505 17:37:31 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:00.505 [2024-11-27 17:37:31.475419] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:16:00.505 [2024-11-27 17:37:31.475465] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:16:00.505 [2024-11-27 17:37:31.475489] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:16:00.505 [2024-11-27 17:37:31.475501] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:16:00.505 17:37:31 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:00.505 17:37:31 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:16:00.505 17:37:31 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:16:00.505 17:37:31 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:16:00.505 17:37:31 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:00.505 17:37:31 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:00.505 17:37:31 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:00.505 17:37:31 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:00.505 17:37:31 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:00.505 17:37:31 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:00.505 17:37:31 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:00.505 17:37:31 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:00.505 17:37:31 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:16:00.505 17:37:31 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:00.505 17:37:31 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:00.505 17:37:31 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:00.505 17:37:31 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:00.505 "name": "Existed_Raid", 00:16:00.505 "uuid": "8a0f1e2e-0e3b-4a0f-a5b0-d279b8c35059", 00:16:00.505 "strip_size_kb": 0, 00:16:00.505 "state": "configuring", 00:16:00.505 "raid_level": "raid1", 00:16:00.505 "superblock": true, 00:16:00.505 "num_base_bdevs": 2, 00:16:00.505 "num_base_bdevs_discovered": 0, 00:16:00.505 "num_base_bdevs_operational": 2, 00:16:00.505 "base_bdevs_list": [ 00:16:00.505 { 00:16:00.505 "name": "BaseBdev1", 00:16:00.505 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:00.505 "is_configured": false, 00:16:00.505 "data_offset": 0, 00:16:00.505 "data_size": 0 00:16:00.505 }, 00:16:00.505 { 00:16:00.505 "name": "BaseBdev2", 00:16:00.505 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:00.505 "is_configured": false, 00:16:00.505 "data_offset": 0, 00:16:00.505 "data_size": 0 00:16:00.505 } 00:16:00.505 ] 00:16:00.505 }' 00:16:00.505 17:37:31 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:00.505 17:37:31 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:00.765 17:37:31 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:16:00.765 17:37:31 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:00.765 17:37:31 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:00.765 [2024-11-27 17:37:31.938519] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:16:00.765 [2024-11-27 17:37:31.938558] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:16:00.765 17:37:31 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:00.765 17:37:31 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:16:00.765 17:37:31 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:00.765 17:37:31 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:00.765 [2024-11-27 17:37:31.950512] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:16:00.765 [2024-11-27 17:37:31.950549] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:16:00.765 [2024-11-27 17:37:31.950564] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:16:00.765 [2024-11-27 17:37:31.950574] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:16:01.026 17:37:31 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:01.026 17:37:31 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 4096 -b BaseBdev1 00:16:01.026 17:37:31 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:01.026 17:37:31 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:01.026 [2024-11-27 17:37:31.971176] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:16:01.026 BaseBdev1 00:16:01.026 17:37:31 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:01.026 17:37:31 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:16:01.026 17:37:31 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:16:01.026 17:37:31 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:16:01.026 17:37:31 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@901 -- # local i 00:16:01.026 17:37:31 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:16:01.026 17:37:31 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:16:01.026 17:37:31 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:16:01.026 17:37:31 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:01.026 17:37:31 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:01.026 17:37:31 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:01.026 17:37:31 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:16:01.026 17:37:31 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:01.026 17:37:31 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:01.026 [ 00:16:01.026 { 00:16:01.026 "name": "BaseBdev1", 00:16:01.026 "aliases": [ 00:16:01.026 "5df0df72-5584-4bc6-be79-7b5d4068aab7" 00:16:01.026 ], 00:16:01.026 "product_name": "Malloc disk", 00:16:01.026 "block_size": 4096, 00:16:01.026 "num_blocks": 8192, 00:16:01.026 "uuid": "5df0df72-5584-4bc6-be79-7b5d4068aab7", 00:16:01.026 "assigned_rate_limits": { 00:16:01.026 "rw_ios_per_sec": 0, 00:16:01.026 "rw_mbytes_per_sec": 0, 00:16:01.026 "r_mbytes_per_sec": 0, 00:16:01.026 "w_mbytes_per_sec": 0 00:16:01.026 }, 00:16:01.026 "claimed": true, 00:16:01.026 "claim_type": "exclusive_write", 00:16:01.026 "zoned": false, 00:16:01.026 "supported_io_types": { 00:16:01.026 "read": true, 00:16:01.026 "write": true, 00:16:01.026 "unmap": true, 00:16:01.026 "flush": true, 00:16:01.026 "reset": true, 00:16:01.026 "nvme_admin": false, 00:16:01.026 "nvme_io": false, 00:16:01.026 "nvme_io_md": false, 00:16:01.026 "write_zeroes": true, 00:16:01.026 "zcopy": true, 00:16:01.026 "get_zone_info": false, 00:16:01.026 "zone_management": false, 00:16:01.026 "zone_append": false, 00:16:01.026 "compare": false, 00:16:01.026 "compare_and_write": false, 00:16:01.026 "abort": true, 00:16:01.026 "seek_hole": false, 00:16:01.026 "seek_data": false, 00:16:01.026 "copy": true, 00:16:01.026 "nvme_iov_md": false 00:16:01.026 }, 00:16:01.026 "memory_domains": [ 00:16:01.026 { 00:16:01.026 "dma_device_id": "system", 00:16:01.026 "dma_device_type": 1 00:16:01.026 }, 00:16:01.026 { 00:16:01.026 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:01.026 "dma_device_type": 2 00:16:01.026 } 00:16:01.026 ], 00:16:01.026 "driver_specific": {} 00:16:01.026 } 00:16:01.026 ] 00:16:01.026 17:37:32 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:01.026 17:37:32 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@907 -- # return 0 00:16:01.026 17:37:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:16:01.026 17:37:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:16:01.026 17:37:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:16:01.026 17:37:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:01.026 17:37:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:01.026 17:37:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:01.026 17:37:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:01.026 17:37:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:01.026 17:37:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:01.026 17:37:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:01.026 17:37:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:01.026 17:37:32 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:01.026 17:37:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:16:01.026 17:37:32 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:01.026 17:37:32 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:01.026 17:37:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:01.026 "name": "Existed_Raid", 00:16:01.026 "uuid": "c5e04e38-c5e8-4640-a7aa-087985a90be1", 00:16:01.026 "strip_size_kb": 0, 00:16:01.026 "state": "configuring", 00:16:01.026 "raid_level": "raid1", 00:16:01.026 "superblock": true, 00:16:01.026 "num_base_bdevs": 2, 00:16:01.026 "num_base_bdevs_discovered": 1, 00:16:01.026 "num_base_bdevs_operational": 2, 00:16:01.026 "base_bdevs_list": [ 00:16:01.026 { 00:16:01.026 "name": "BaseBdev1", 00:16:01.026 "uuid": "5df0df72-5584-4bc6-be79-7b5d4068aab7", 00:16:01.026 "is_configured": true, 00:16:01.026 "data_offset": 256, 00:16:01.027 "data_size": 7936 00:16:01.027 }, 00:16:01.027 { 00:16:01.027 "name": "BaseBdev2", 00:16:01.027 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:01.027 "is_configured": false, 00:16:01.027 "data_offset": 0, 00:16:01.027 "data_size": 0 00:16:01.027 } 00:16:01.027 ] 00:16:01.027 }' 00:16:01.027 17:37:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:01.027 17:37:32 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:01.287 17:37:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:16:01.287 17:37:32 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:01.287 17:37:32 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:01.287 [2024-11-27 17:37:32.410456] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:16:01.287 [2024-11-27 17:37:32.410502] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:16:01.287 17:37:32 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:01.287 17:37:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:16:01.287 17:37:32 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:01.287 17:37:32 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:01.287 [2024-11-27 17:37:32.426484] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:16:01.287 [2024-11-27 17:37:32.428278] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:16:01.287 [2024-11-27 17:37:32.428318] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:16:01.287 17:37:32 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:01.287 17:37:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:16:01.287 17:37:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:16:01.287 17:37:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:16:01.287 17:37:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:16:01.287 17:37:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:16:01.287 17:37:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:01.287 17:37:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:01.287 17:37:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:01.287 17:37:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:01.287 17:37:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:01.287 17:37:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:01.287 17:37:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:01.287 17:37:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:01.287 17:37:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:16:01.287 17:37:32 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:01.287 17:37:32 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:01.287 17:37:32 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:01.547 17:37:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:01.547 "name": "Existed_Raid", 00:16:01.547 "uuid": "e097aeb4-caf6-4035-931b-9cb6f4b57525", 00:16:01.547 "strip_size_kb": 0, 00:16:01.547 "state": "configuring", 00:16:01.547 "raid_level": "raid1", 00:16:01.547 "superblock": true, 00:16:01.547 "num_base_bdevs": 2, 00:16:01.547 "num_base_bdevs_discovered": 1, 00:16:01.547 "num_base_bdevs_operational": 2, 00:16:01.547 "base_bdevs_list": [ 00:16:01.547 { 00:16:01.547 "name": "BaseBdev1", 00:16:01.547 "uuid": "5df0df72-5584-4bc6-be79-7b5d4068aab7", 00:16:01.547 "is_configured": true, 00:16:01.547 "data_offset": 256, 00:16:01.547 "data_size": 7936 00:16:01.547 }, 00:16:01.547 { 00:16:01.547 "name": "BaseBdev2", 00:16:01.547 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:01.547 "is_configured": false, 00:16:01.547 "data_offset": 0, 00:16:01.547 "data_size": 0 00:16:01.547 } 00:16:01.547 ] 00:16:01.547 }' 00:16:01.547 17:37:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:01.547 17:37:32 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:01.807 17:37:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 4096 -b BaseBdev2 00:16:01.807 17:37:32 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:01.807 17:37:32 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:01.807 [2024-11-27 17:37:32.875289] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:16:01.807 [2024-11-27 17:37:32.875836] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:16:01.807 [2024-11-27 17:37:32.875903] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:16:01.807 BaseBdev2 00:16:01.807 [2024-11-27 17:37:32.876936] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:16:01.807 17:37:32 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:01.807 [2024-11-27 17:37:32.877441] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:16:01.807 [2024-11-27 17:37:32.877547] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:16:01.807 17:37:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:16:01.807 17:37:32 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:16:01.807 [2024-11-27 17:37:32.878036] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:01.807 17:37:32 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:16:01.807 17:37:32 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@901 -- # local i 00:16:01.807 17:37:32 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:16:01.807 17:37:32 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:16:01.807 17:37:32 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:16:01.807 17:37:32 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:01.807 17:37:32 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:01.807 17:37:32 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:01.807 17:37:32 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:16:01.807 17:37:32 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:01.807 17:37:32 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:01.807 [ 00:16:01.807 { 00:16:01.807 "name": "BaseBdev2", 00:16:01.807 "aliases": [ 00:16:01.807 "e21bf88b-5c06-462e-80a3-ce8df83b2896" 00:16:01.807 ], 00:16:01.807 "product_name": "Malloc disk", 00:16:01.807 "block_size": 4096, 00:16:01.807 "num_blocks": 8192, 00:16:01.807 "uuid": "e21bf88b-5c06-462e-80a3-ce8df83b2896", 00:16:01.807 "assigned_rate_limits": { 00:16:01.807 "rw_ios_per_sec": 0, 00:16:01.807 "rw_mbytes_per_sec": 0, 00:16:01.807 "r_mbytes_per_sec": 0, 00:16:01.807 "w_mbytes_per_sec": 0 00:16:01.807 }, 00:16:01.807 "claimed": true, 00:16:01.807 "claim_type": "exclusive_write", 00:16:01.807 "zoned": false, 00:16:01.807 "supported_io_types": { 00:16:01.807 "read": true, 00:16:01.807 "write": true, 00:16:01.807 "unmap": true, 00:16:01.807 "flush": true, 00:16:01.807 "reset": true, 00:16:01.807 "nvme_admin": false, 00:16:01.808 "nvme_io": false, 00:16:01.808 "nvme_io_md": false, 00:16:01.808 "write_zeroes": true, 00:16:01.808 "zcopy": true, 00:16:01.808 "get_zone_info": false, 00:16:01.808 "zone_management": false, 00:16:01.808 "zone_append": false, 00:16:01.808 "compare": false, 00:16:01.808 "compare_and_write": false, 00:16:01.808 "abort": true, 00:16:01.808 "seek_hole": false, 00:16:01.808 "seek_data": false, 00:16:01.808 "copy": true, 00:16:01.808 "nvme_iov_md": false 00:16:01.808 }, 00:16:01.808 "memory_domains": [ 00:16:01.808 { 00:16:01.808 "dma_device_id": "system", 00:16:01.808 "dma_device_type": 1 00:16:01.808 }, 00:16:01.808 { 00:16:01.808 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:01.808 "dma_device_type": 2 00:16:01.808 } 00:16:01.808 ], 00:16:01.808 "driver_specific": {} 00:16:01.808 } 00:16:01.808 ] 00:16:01.808 17:37:32 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:01.808 17:37:32 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@907 -- # return 0 00:16:01.808 17:37:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:16:01.808 17:37:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:16:01.808 17:37:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 2 00:16:01.808 17:37:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:16:01.808 17:37:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:01.808 17:37:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:01.808 17:37:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:01.808 17:37:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:01.808 17:37:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:01.808 17:37:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:01.808 17:37:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:01.808 17:37:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:01.808 17:37:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:01.808 17:37:32 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:01.808 17:37:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:16:01.808 17:37:32 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:01.808 17:37:32 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:01.808 17:37:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:01.808 "name": "Existed_Raid", 00:16:01.808 "uuid": "e097aeb4-caf6-4035-931b-9cb6f4b57525", 00:16:01.808 "strip_size_kb": 0, 00:16:01.808 "state": "online", 00:16:01.808 "raid_level": "raid1", 00:16:01.808 "superblock": true, 00:16:01.808 "num_base_bdevs": 2, 00:16:01.808 "num_base_bdevs_discovered": 2, 00:16:01.808 "num_base_bdevs_operational": 2, 00:16:01.808 "base_bdevs_list": [ 00:16:01.808 { 00:16:01.808 "name": "BaseBdev1", 00:16:01.808 "uuid": "5df0df72-5584-4bc6-be79-7b5d4068aab7", 00:16:01.808 "is_configured": true, 00:16:01.808 "data_offset": 256, 00:16:01.808 "data_size": 7936 00:16:01.808 }, 00:16:01.808 { 00:16:01.808 "name": "BaseBdev2", 00:16:01.808 "uuid": "e21bf88b-5c06-462e-80a3-ce8df83b2896", 00:16:01.808 "is_configured": true, 00:16:01.808 "data_offset": 256, 00:16:01.808 "data_size": 7936 00:16:01.808 } 00:16:01.808 ] 00:16:01.808 }' 00:16:01.808 17:37:32 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:01.808 17:37:32 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:02.378 17:37:33 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:16:02.378 17:37:33 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:16:02.378 17:37:33 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:16:02.378 17:37:33 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:16:02.378 17:37:33 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@184 -- # local name 00:16:02.378 17:37:33 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:16:02.378 17:37:33 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:16:02.378 17:37:33 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:16:02.378 17:37:33 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:02.378 17:37:33 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:02.378 [2024-11-27 17:37:33.334702] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:02.378 17:37:33 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:02.378 17:37:33 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:16:02.378 "name": "Existed_Raid", 00:16:02.378 "aliases": [ 00:16:02.378 "e097aeb4-caf6-4035-931b-9cb6f4b57525" 00:16:02.378 ], 00:16:02.378 "product_name": "Raid Volume", 00:16:02.378 "block_size": 4096, 00:16:02.378 "num_blocks": 7936, 00:16:02.378 "uuid": "e097aeb4-caf6-4035-931b-9cb6f4b57525", 00:16:02.378 "assigned_rate_limits": { 00:16:02.378 "rw_ios_per_sec": 0, 00:16:02.378 "rw_mbytes_per_sec": 0, 00:16:02.378 "r_mbytes_per_sec": 0, 00:16:02.378 "w_mbytes_per_sec": 0 00:16:02.378 }, 00:16:02.378 "claimed": false, 00:16:02.378 "zoned": false, 00:16:02.378 "supported_io_types": { 00:16:02.378 "read": true, 00:16:02.378 "write": true, 00:16:02.378 "unmap": false, 00:16:02.378 "flush": false, 00:16:02.378 "reset": true, 00:16:02.378 "nvme_admin": false, 00:16:02.378 "nvme_io": false, 00:16:02.378 "nvme_io_md": false, 00:16:02.378 "write_zeroes": true, 00:16:02.378 "zcopy": false, 00:16:02.378 "get_zone_info": false, 00:16:02.378 "zone_management": false, 00:16:02.378 "zone_append": false, 00:16:02.378 "compare": false, 00:16:02.378 "compare_and_write": false, 00:16:02.378 "abort": false, 00:16:02.378 "seek_hole": false, 00:16:02.378 "seek_data": false, 00:16:02.378 "copy": false, 00:16:02.378 "nvme_iov_md": false 00:16:02.378 }, 00:16:02.378 "memory_domains": [ 00:16:02.378 { 00:16:02.378 "dma_device_id": "system", 00:16:02.378 "dma_device_type": 1 00:16:02.378 }, 00:16:02.378 { 00:16:02.378 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:02.378 "dma_device_type": 2 00:16:02.378 }, 00:16:02.378 { 00:16:02.378 "dma_device_id": "system", 00:16:02.378 "dma_device_type": 1 00:16:02.378 }, 00:16:02.378 { 00:16:02.378 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:02.378 "dma_device_type": 2 00:16:02.378 } 00:16:02.378 ], 00:16:02.378 "driver_specific": { 00:16:02.378 "raid": { 00:16:02.378 "uuid": "e097aeb4-caf6-4035-931b-9cb6f4b57525", 00:16:02.378 "strip_size_kb": 0, 00:16:02.378 "state": "online", 00:16:02.378 "raid_level": "raid1", 00:16:02.378 "superblock": true, 00:16:02.378 "num_base_bdevs": 2, 00:16:02.378 "num_base_bdevs_discovered": 2, 00:16:02.378 "num_base_bdevs_operational": 2, 00:16:02.378 "base_bdevs_list": [ 00:16:02.378 { 00:16:02.378 "name": "BaseBdev1", 00:16:02.378 "uuid": "5df0df72-5584-4bc6-be79-7b5d4068aab7", 00:16:02.378 "is_configured": true, 00:16:02.378 "data_offset": 256, 00:16:02.378 "data_size": 7936 00:16:02.378 }, 00:16:02.378 { 00:16:02.378 "name": "BaseBdev2", 00:16:02.378 "uuid": "e21bf88b-5c06-462e-80a3-ce8df83b2896", 00:16:02.378 "is_configured": true, 00:16:02.378 "data_offset": 256, 00:16:02.378 "data_size": 7936 00:16:02.378 } 00:16:02.378 ] 00:16:02.378 } 00:16:02.378 } 00:16:02.378 }' 00:16:02.378 17:37:33 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:16:02.378 17:37:33 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:16:02.378 BaseBdev2' 00:16:02.378 17:37:33 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:02.378 17:37:33 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4096 ' 00:16:02.378 17:37:33 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:16:02.378 17:37:33 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:02.378 17:37:33 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:16:02.378 17:37:33 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:02.378 17:37:33 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:02.378 17:37:33 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:02.378 17:37:33 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 ' 00:16:02.378 17:37:33 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@193 -- # [[ 4096 == \4\0\9\6\ \ \ ]] 00:16:02.378 17:37:33 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:16:02.378 17:37:33 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:16:02.378 17:37:33 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:02.378 17:37:33 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:02.379 17:37:33 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:02.379 17:37:33 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:02.379 17:37:33 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 ' 00:16:02.379 17:37:33 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@193 -- # [[ 4096 == \4\0\9\6\ \ \ ]] 00:16:02.379 17:37:33 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:16:02.379 17:37:33 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:02.379 17:37:33 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:02.639 [2024-11-27 17:37:33.570109] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:16:02.639 17:37:33 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:02.639 17:37:33 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@260 -- # local expected_state 00:16:02.639 17:37:33 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:16:02.639 17:37:33 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@198 -- # case $1 in 00:16:02.639 17:37:33 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@199 -- # return 0 00:16:02.639 17:37:33 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:16:02.639 17:37:33 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 1 00:16:02.639 17:37:33 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:16:02.639 17:37:33 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:02.639 17:37:33 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:02.639 17:37:33 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:02.639 17:37:33 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:02.639 17:37:33 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:02.639 17:37:33 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:02.639 17:37:33 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:02.640 17:37:33 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:02.640 17:37:33 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:16:02.640 17:37:33 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:02.640 17:37:33 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:02.640 17:37:33 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:02.640 17:37:33 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:02.640 17:37:33 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:02.640 "name": "Existed_Raid", 00:16:02.640 "uuid": "e097aeb4-caf6-4035-931b-9cb6f4b57525", 00:16:02.640 "strip_size_kb": 0, 00:16:02.640 "state": "online", 00:16:02.640 "raid_level": "raid1", 00:16:02.640 "superblock": true, 00:16:02.640 "num_base_bdevs": 2, 00:16:02.640 "num_base_bdevs_discovered": 1, 00:16:02.640 "num_base_bdevs_operational": 1, 00:16:02.640 "base_bdevs_list": [ 00:16:02.640 { 00:16:02.640 "name": null, 00:16:02.640 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:02.640 "is_configured": false, 00:16:02.640 "data_offset": 0, 00:16:02.640 "data_size": 7936 00:16:02.640 }, 00:16:02.640 { 00:16:02.640 "name": "BaseBdev2", 00:16:02.640 "uuid": "e21bf88b-5c06-462e-80a3-ce8df83b2896", 00:16:02.640 "is_configured": true, 00:16:02.640 "data_offset": 256, 00:16:02.640 "data_size": 7936 00:16:02.640 } 00:16:02.640 ] 00:16:02.640 }' 00:16:02.640 17:37:33 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:02.640 17:37:33 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:02.900 17:37:34 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:16:02.900 17:37:34 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:16:02.900 17:37:34 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:02.900 17:37:34 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:02.900 17:37:34 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:02.900 17:37:34 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:16:02.900 17:37:34 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:02.900 17:37:34 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:16:02.900 17:37:34 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:16:02.900 17:37:34 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:16:02.900 17:37:34 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:02.900 17:37:34 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:02.900 [2024-11-27 17:37:34.080322] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:16:02.900 [2024-11-27 17:37:34.080422] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:16:03.160 [2024-11-27 17:37:34.091918] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:03.160 [2024-11-27 17:37:34.091973] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:03.160 [2024-11-27 17:37:34.091991] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:16:03.160 17:37:34 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:03.160 17:37:34 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:16:03.160 17:37:34 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:16:03.160 17:37:34 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:03.160 17:37:34 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:03.160 17:37:34 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:03.160 17:37:34 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:16:03.160 17:37:34 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:03.160 17:37:34 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:16:03.160 17:37:34 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:16:03.160 17:37:34 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:16:03.160 17:37:34 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@326 -- # killprocess 96170 00:16:03.160 17:37:34 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@950 -- # '[' -z 96170 ']' 00:16:03.160 17:37:34 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@954 -- # kill -0 96170 00:16:03.160 17:37:34 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@955 -- # uname 00:16:03.160 17:37:34 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:16:03.160 17:37:34 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 96170 00:16:03.160 17:37:34 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:16:03.160 17:37:34 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:16:03.160 killing process with pid 96170 00:16:03.160 17:37:34 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@968 -- # echo 'killing process with pid 96170' 00:16:03.160 17:37:34 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@969 -- # kill 96170 00:16:03.160 [2024-11-27 17:37:34.183293] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:16:03.160 17:37:34 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@974 -- # wait 96170 00:16:03.160 [2024-11-27 17:37:34.184277] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:16:03.420 17:37:34 bdev_raid.raid_state_function_test_sb_4k -- bdev/bdev_raid.sh@328 -- # return 0 00:16:03.420 00:16:03.420 real 0m3.881s 00:16:03.420 user 0m6.090s 00:16:03.420 sys 0m0.798s 00:16:03.420 17:37:34 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@1126 -- # xtrace_disable 00:16:03.420 17:37:34 bdev_raid.raid_state_function_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:03.420 ************************************ 00:16:03.420 END TEST raid_state_function_test_sb_4k 00:16:03.420 ************************************ 00:16:03.420 17:37:34 bdev_raid -- bdev/bdev_raid.sh@998 -- # run_test raid_superblock_test_4k raid_superblock_test raid1 2 00:16:03.420 17:37:34 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:16:03.420 17:37:34 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:16:03.420 17:37:34 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:16:03.420 ************************************ 00:16:03.420 START TEST raid_superblock_test_4k 00:16:03.420 ************************************ 00:16:03.420 17:37:34 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@1125 -- # raid_superblock_test raid1 2 00:16:03.420 17:37:34 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@393 -- # local raid_level=raid1 00:16:03.420 17:37:34 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=2 00:16:03.420 17:37:34 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:16:03.420 17:37:34 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:16:03.420 17:37:34 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:16:03.420 17:37:34 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:16:03.420 17:37:34 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:16:03.420 17:37:34 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:16:03.420 17:37:34 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:16:03.420 17:37:34 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@399 -- # local strip_size 00:16:03.420 17:37:34 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:16:03.420 17:37:34 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:16:03.420 17:37:34 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:16:03.420 17:37:34 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@404 -- # '[' raid1 '!=' raid1 ']' 00:16:03.420 17:37:34 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@408 -- # strip_size=0 00:16:03.420 17:37:34 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@412 -- # raid_pid=96411 00:16:03.420 17:37:34 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:16:03.420 17:37:34 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@413 -- # waitforlisten 96411 00:16:03.420 17:37:34 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@831 -- # '[' -z 96411 ']' 00:16:03.420 17:37:34 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:16:03.420 17:37:34 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@836 -- # local max_retries=100 00:16:03.420 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:16:03.420 17:37:34 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:16:03.420 17:37:34 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@840 -- # xtrace_disable 00:16:03.420 17:37:34 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:03.420 [2024-11-27 17:37:34.580927] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:16:03.420 [2024-11-27 17:37:34.581065] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid96411 ] 00:16:03.680 [2024-11-27 17:37:34.726643] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:16:03.680 [2024-11-27 17:37:34.770859] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:16:03.680 [2024-11-27 17:37:34.812463] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:16:03.680 [2024-11-27 17:37:34.812507] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:16:04.250 17:37:35 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:16:04.250 17:37:35 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@864 -- # return 0 00:16:04.250 17:37:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:16:04.250 17:37:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:16:04.250 17:37:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:16:04.250 17:37:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:16:04.250 17:37:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:16:04.250 17:37:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:16:04.250 17:37:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:16:04.250 17:37:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:16:04.250 17:37:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 4096 -b malloc1 00:16:04.250 17:37:35 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:04.250 17:37:35 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:04.250 malloc1 00:16:04.250 17:37:35 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:04.250 17:37:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:16:04.250 17:37:35 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:04.250 17:37:35 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:04.250 [2024-11-27 17:37:35.422058] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:16:04.250 [2024-11-27 17:37:35.422123] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:04.250 [2024-11-27 17:37:35.422159] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:16:04.250 [2024-11-27 17:37:35.422180] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:04.250 [2024-11-27 17:37:35.424151] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:04.250 [2024-11-27 17:37:35.424190] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:16:04.250 pt1 00:16:04.250 17:37:35 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:04.250 17:37:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:16:04.250 17:37:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:16:04.250 17:37:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:16:04.250 17:37:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:16:04.250 17:37:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:16:04.250 17:37:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:16:04.250 17:37:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:16:04.250 17:37:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:16:04.250 17:37:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 4096 -b malloc2 00:16:04.250 17:37:35 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:04.250 17:37:35 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:04.511 malloc2 00:16:04.511 17:37:35 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:04.511 17:37:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:16:04.511 17:37:35 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:04.511 17:37:35 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:04.511 [2024-11-27 17:37:35.465324] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:16:04.511 [2024-11-27 17:37:35.465442] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:04.511 [2024-11-27 17:37:35.465483] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:16:04.511 [2024-11-27 17:37:35.465517] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:04.511 [2024-11-27 17:37:35.470716] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:04.511 [2024-11-27 17:37:35.470801] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:16:04.511 pt2 00:16:04.511 17:37:35 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:04.511 17:37:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:16:04.511 17:37:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:16:04.511 17:37:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''pt1 pt2'\''' -n raid_bdev1 -s 00:16:04.511 17:37:35 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:04.511 17:37:35 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:04.511 [2024-11-27 17:37:35.479108] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:16:04.511 [2024-11-27 17:37:35.482392] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:16:04.511 [2024-11-27 17:37:35.482630] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:16:04.511 [2024-11-27 17:37:35.482672] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:16:04.511 [2024-11-27 17:37:35.483109] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:16:04.511 [2024-11-27 17:37:35.483295] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:16:04.511 [2024-11-27 17:37:35.483315] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:16:04.511 [2024-11-27 17:37:35.483541] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:04.511 17:37:35 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:04.511 17:37:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:16:04.511 17:37:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:04.511 17:37:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:04.511 17:37:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:04.511 17:37:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:04.511 17:37:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:04.511 17:37:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:04.511 17:37:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:04.511 17:37:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:04.511 17:37:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:04.511 17:37:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:04.511 17:37:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:04.511 17:37:35 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:04.511 17:37:35 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:04.511 17:37:35 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:04.511 17:37:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:04.511 "name": "raid_bdev1", 00:16:04.511 "uuid": "bcbc5ee9-d373-4e2d-ac11-c480a3a0f00d", 00:16:04.511 "strip_size_kb": 0, 00:16:04.511 "state": "online", 00:16:04.511 "raid_level": "raid1", 00:16:04.511 "superblock": true, 00:16:04.511 "num_base_bdevs": 2, 00:16:04.511 "num_base_bdevs_discovered": 2, 00:16:04.511 "num_base_bdevs_operational": 2, 00:16:04.511 "base_bdevs_list": [ 00:16:04.511 { 00:16:04.511 "name": "pt1", 00:16:04.511 "uuid": "00000000-0000-0000-0000-000000000001", 00:16:04.511 "is_configured": true, 00:16:04.511 "data_offset": 256, 00:16:04.511 "data_size": 7936 00:16:04.511 }, 00:16:04.511 { 00:16:04.511 "name": "pt2", 00:16:04.511 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:04.511 "is_configured": true, 00:16:04.511 "data_offset": 256, 00:16:04.511 "data_size": 7936 00:16:04.511 } 00:16:04.511 ] 00:16:04.511 }' 00:16:04.511 17:37:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:04.511 17:37:35 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:04.771 17:37:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:16:04.771 17:37:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:16:04.771 17:37:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:16:04.771 17:37:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:16:04.771 17:37:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@184 -- # local name 00:16:04.771 17:37:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:16:04.771 17:37:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:16:04.771 17:37:35 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:04.772 17:37:35 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:04.772 17:37:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:16:04.772 [2024-11-27 17:37:35.911028] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:04.772 17:37:35 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:04.772 17:37:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:16:04.772 "name": "raid_bdev1", 00:16:04.772 "aliases": [ 00:16:04.772 "bcbc5ee9-d373-4e2d-ac11-c480a3a0f00d" 00:16:04.772 ], 00:16:04.772 "product_name": "Raid Volume", 00:16:04.772 "block_size": 4096, 00:16:04.772 "num_blocks": 7936, 00:16:04.772 "uuid": "bcbc5ee9-d373-4e2d-ac11-c480a3a0f00d", 00:16:04.772 "assigned_rate_limits": { 00:16:04.772 "rw_ios_per_sec": 0, 00:16:04.772 "rw_mbytes_per_sec": 0, 00:16:04.772 "r_mbytes_per_sec": 0, 00:16:04.772 "w_mbytes_per_sec": 0 00:16:04.772 }, 00:16:04.772 "claimed": false, 00:16:04.772 "zoned": false, 00:16:04.772 "supported_io_types": { 00:16:04.772 "read": true, 00:16:04.772 "write": true, 00:16:04.772 "unmap": false, 00:16:04.772 "flush": false, 00:16:04.772 "reset": true, 00:16:04.772 "nvme_admin": false, 00:16:04.772 "nvme_io": false, 00:16:04.772 "nvme_io_md": false, 00:16:04.772 "write_zeroes": true, 00:16:04.772 "zcopy": false, 00:16:04.772 "get_zone_info": false, 00:16:04.772 "zone_management": false, 00:16:04.772 "zone_append": false, 00:16:04.772 "compare": false, 00:16:04.772 "compare_and_write": false, 00:16:04.772 "abort": false, 00:16:04.772 "seek_hole": false, 00:16:04.772 "seek_data": false, 00:16:04.772 "copy": false, 00:16:04.772 "nvme_iov_md": false 00:16:04.772 }, 00:16:04.772 "memory_domains": [ 00:16:04.772 { 00:16:04.772 "dma_device_id": "system", 00:16:04.772 "dma_device_type": 1 00:16:04.772 }, 00:16:04.772 { 00:16:04.772 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:04.772 "dma_device_type": 2 00:16:04.772 }, 00:16:04.772 { 00:16:04.772 "dma_device_id": "system", 00:16:04.772 "dma_device_type": 1 00:16:04.772 }, 00:16:04.772 { 00:16:04.772 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:04.772 "dma_device_type": 2 00:16:04.772 } 00:16:04.772 ], 00:16:04.772 "driver_specific": { 00:16:04.772 "raid": { 00:16:04.772 "uuid": "bcbc5ee9-d373-4e2d-ac11-c480a3a0f00d", 00:16:04.772 "strip_size_kb": 0, 00:16:04.772 "state": "online", 00:16:04.772 "raid_level": "raid1", 00:16:04.772 "superblock": true, 00:16:04.772 "num_base_bdevs": 2, 00:16:04.772 "num_base_bdevs_discovered": 2, 00:16:04.772 "num_base_bdevs_operational": 2, 00:16:04.772 "base_bdevs_list": [ 00:16:04.772 { 00:16:04.772 "name": "pt1", 00:16:04.772 "uuid": "00000000-0000-0000-0000-000000000001", 00:16:04.772 "is_configured": true, 00:16:04.772 "data_offset": 256, 00:16:04.772 "data_size": 7936 00:16:04.772 }, 00:16:04.772 { 00:16:04.772 "name": "pt2", 00:16:04.772 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:04.772 "is_configured": true, 00:16:04.772 "data_offset": 256, 00:16:04.772 "data_size": 7936 00:16:04.772 } 00:16:04.772 ] 00:16:04.772 } 00:16:04.772 } 00:16:04.772 }' 00:16:04.772 17:37:35 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:16:05.033 17:37:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:16:05.033 pt2' 00:16:05.033 17:37:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:05.033 17:37:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4096 ' 00:16:05.033 17:37:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:16:05.033 17:37:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:05.033 17:37:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:16:05.033 17:37:36 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:05.033 17:37:36 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:05.033 17:37:36 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:05.033 17:37:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 ' 00:16:05.033 17:37:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@193 -- # [[ 4096 == \4\0\9\6\ \ \ ]] 00:16:05.033 17:37:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:16:05.033 17:37:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:16:05.033 17:37:36 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:05.033 17:37:36 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:05.033 17:37:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:05.033 17:37:36 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:05.033 17:37:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 ' 00:16:05.033 17:37:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@193 -- # [[ 4096 == \4\0\9\6\ \ \ ]] 00:16:05.033 17:37:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:16:05.033 17:37:36 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:05.033 17:37:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:16:05.033 17:37:36 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:05.033 [2024-11-27 17:37:36.138540] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:05.033 17:37:36 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:05.033 17:37:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=bcbc5ee9-d373-4e2d-ac11-c480a3a0f00d 00:16:05.033 17:37:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@436 -- # '[' -z bcbc5ee9-d373-4e2d-ac11-c480a3a0f00d ']' 00:16:05.033 17:37:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:16:05.033 17:37:36 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:05.033 17:37:36 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:05.033 [2024-11-27 17:37:36.182252] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:16:05.033 [2024-11-27 17:37:36.182278] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:16:05.033 [2024-11-27 17:37:36.182342] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:05.033 [2024-11-27 17:37:36.182396] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:05.033 [2024-11-27 17:37:36.182411] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:16:05.033 17:37:36 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:05.033 17:37:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:05.033 17:37:36 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:05.033 17:37:36 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:05.033 17:37:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:16:05.033 17:37:36 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:05.294 17:37:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:16:05.294 17:37:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:16:05.294 17:37:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:16:05.294 17:37:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:16:05.294 17:37:36 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:05.294 17:37:36 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:05.294 17:37:36 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:05.294 17:37:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:16:05.294 17:37:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:16:05.294 17:37:36 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:05.294 17:37:36 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:05.294 17:37:36 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:05.294 17:37:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:16:05.294 17:37:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:16:05.294 17:37:36 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:05.294 17:37:36 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:05.294 17:37:36 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:05.294 17:37:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:16:05.294 17:37:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:16:05.294 17:37:36 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@650 -- # local es=0 00:16:05.294 17:37:36 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:16:05.294 17:37:36 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:16:05.294 17:37:36 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:16:05.294 17:37:36 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:16:05.294 17:37:36 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:16:05.294 17:37:36 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:16:05.294 17:37:36 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:05.294 17:37:36 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:05.294 [2024-11-27 17:37:36.306054] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:16:05.294 [2024-11-27 17:37:36.307791] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:16:05.294 [2024-11-27 17:37:36.307851] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:16:05.294 [2024-11-27 17:37:36.307896] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:16:05.294 [2024-11-27 17:37:36.307913] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:16:05.294 [2024-11-27 17:37:36.307921] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state configuring 00:16:05.294 request: 00:16:05.294 { 00:16:05.294 "name": "raid_bdev1", 00:16:05.294 "raid_level": "raid1", 00:16:05.294 "base_bdevs": [ 00:16:05.294 "malloc1", 00:16:05.294 "malloc2" 00:16:05.294 ], 00:16:05.294 "superblock": false, 00:16:05.294 "method": "bdev_raid_create", 00:16:05.294 "req_id": 1 00:16:05.294 } 00:16:05.294 Got JSON-RPC error response 00:16:05.294 response: 00:16:05.294 { 00:16:05.294 "code": -17, 00:16:05.294 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:16:05.294 } 00:16:05.294 17:37:36 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:16:05.294 17:37:36 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@653 -- # es=1 00:16:05.294 17:37:36 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:16:05.294 17:37:36 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:16:05.294 17:37:36 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:16:05.294 17:37:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:05.294 17:37:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:16:05.294 17:37:36 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:05.294 17:37:36 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:05.294 17:37:36 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:05.294 17:37:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:16:05.294 17:37:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:16:05.294 17:37:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:16:05.294 17:37:36 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:05.294 17:37:36 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:05.294 [2024-11-27 17:37:36.373915] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:16:05.294 [2024-11-27 17:37:36.373966] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:05.294 [2024-11-27 17:37:36.373987] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:16:05.294 [2024-11-27 17:37:36.373995] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:05.294 [2024-11-27 17:37:36.375929] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:05.294 [2024-11-27 17:37:36.375968] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:16:05.294 [2024-11-27 17:37:36.376023] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:16:05.294 [2024-11-27 17:37:36.376049] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:16:05.294 pt1 00:16:05.294 17:37:36 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:05.294 17:37:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 2 00:16:05.294 17:37:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:05.294 17:37:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:16:05.294 17:37:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:05.294 17:37:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:05.294 17:37:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:05.294 17:37:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:05.294 17:37:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:05.294 17:37:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:05.294 17:37:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:05.294 17:37:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:05.294 17:37:36 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:05.294 17:37:36 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:05.294 17:37:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:05.294 17:37:36 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:05.294 17:37:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:05.294 "name": "raid_bdev1", 00:16:05.294 "uuid": "bcbc5ee9-d373-4e2d-ac11-c480a3a0f00d", 00:16:05.294 "strip_size_kb": 0, 00:16:05.294 "state": "configuring", 00:16:05.294 "raid_level": "raid1", 00:16:05.294 "superblock": true, 00:16:05.294 "num_base_bdevs": 2, 00:16:05.294 "num_base_bdevs_discovered": 1, 00:16:05.294 "num_base_bdevs_operational": 2, 00:16:05.294 "base_bdevs_list": [ 00:16:05.294 { 00:16:05.294 "name": "pt1", 00:16:05.294 "uuid": "00000000-0000-0000-0000-000000000001", 00:16:05.294 "is_configured": true, 00:16:05.294 "data_offset": 256, 00:16:05.294 "data_size": 7936 00:16:05.294 }, 00:16:05.294 { 00:16:05.294 "name": null, 00:16:05.294 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:05.294 "is_configured": false, 00:16:05.294 "data_offset": 256, 00:16:05.294 "data_size": 7936 00:16:05.294 } 00:16:05.294 ] 00:16:05.294 }' 00:16:05.294 17:37:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:05.294 17:37:36 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:05.864 17:37:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@470 -- # '[' 2 -gt 2 ']' 00:16:05.864 17:37:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:16:05.864 17:37:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:16:05.864 17:37:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:16:05.864 17:37:36 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:05.864 17:37:36 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:05.864 [2024-11-27 17:37:36.833124] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:16:05.864 [2024-11-27 17:37:36.833177] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:05.864 [2024-11-27 17:37:36.833195] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:16:05.864 [2024-11-27 17:37:36.833203] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:05.864 [2024-11-27 17:37:36.833547] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:05.864 [2024-11-27 17:37:36.833574] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:16:05.864 [2024-11-27 17:37:36.833632] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:16:05.864 [2024-11-27 17:37:36.833654] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:16:05.864 [2024-11-27 17:37:36.833734] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:16:05.864 [2024-11-27 17:37:36.833747] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:16:05.864 [2024-11-27 17:37:36.833968] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:16:05.864 [2024-11-27 17:37:36.834078] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:16:05.865 [2024-11-27 17:37:36.834095] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:16:05.865 [2024-11-27 17:37:36.834211] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:05.865 pt2 00:16:05.865 17:37:36 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:05.865 17:37:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:16:05.865 17:37:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:16:05.865 17:37:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:16:05.865 17:37:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:05.865 17:37:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:05.865 17:37:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:05.865 17:37:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:05.865 17:37:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:05.865 17:37:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:05.865 17:37:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:05.865 17:37:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:05.865 17:37:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:05.865 17:37:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:05.865 17:37:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:05.865 17:37:36 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:05.865 17:37:36 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:05.865 17:37:36 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:05.865 17:37:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:05.865 "name": "raid_bdev1", 00:16:05.865 "uuid": "bcbc5ee9-d373-4e2d-ac11-c480a3a0f00d", 00:16:05.865 "strip_size_kb": 0, 00:16:05.865 "state": "online", 00:16:05.865 "raid_level": "raid1", 00:16:05.865 "superblock": true, 00:16:05.865 "num_base_bdevs": 2, 00:16:05.865 "num_base_bdevs_discovered": 2, 00:16:05.865 "num_base_bdevs_operational": 2, 00:16:05.865 "base_bdevs_list": [ 00:16:05.865 { 00:16:05.865 "name": "pt1", 00:16:05.865 "uuid": "00000000-0000-0000-0000-000000000001", 00:16:05.865 "is_configured": true, 00:16:05.865 "data_offset": 256, 00:16:05.865 "data_size": 7936 00:16:05.865 }, 00:16:05.865 { 00:16:05.865 "name": "pt2", 00:16:05.865 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:05.865 "is_configured": true, 00:16:05.865 "data_offset": 256, 00:16:05.865 "data_size": 7936 00:16:05.865 } 00:16:05.865 ] 00:16:05.865 }' 00:16:05.865 17:37:36 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:05.865 17:37:36 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:06.125 17:37:37 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:16:06.125 17:37:37 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:16:06.125 17:37:37 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:16:06.125 17:37:37 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:16:06.125 17:37:37 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@184 -- # local name 00:16:06.125 17:37:37 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:16:06.125 17:37:37 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:16:06.125 17:37:37 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:16:06.125 17:37:37 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:06.125 17:37:37 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:06.125 [2024-11-27 17:37:37.256633] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:06.125 17:37:37 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:06.125 17:37:37 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:16:06.125 "name": "raid_bdev1", 00:16:06.125 "aliases": [ 00:16:06.125 "bcbc5ee9-d373-4e2d-ac11-c480a3a0f00d" 00:16:06.125 ], 00:16:06.125 "product_name": "Raid Volume", 00:16:06.125 "block_size": 4096, 00:16:06.125 "num_blocks": 7936, 00:16:06.125 "uuid": "bcbc5ee9-d373-4e2d-ac11-c480a3a0f00d", 00:16:06.125 "assigned_rate_limits": { 00:16:06.125 "rw_ios_per_sec": 0, 00:16:06.125 "rw_mbytes_per_sec": 0, 00:16:06.125 "r_mbytes_per_sec": 0, 00:16:06.125 "w_mbytes_per_sec": 0 00:16:06.125 }, 00:16:06.125 "claimed": false, 00:16:06.125 "zoned": false, 00:16:06.125 "supported_io_types": { 00:16:06.125 "read": true, 00:16:06.125 "write": true, 00:16:06.125 "unmap": false, 00:16:06.125 "flush": false, 00:16:06.125 "reset": true, 00:16:06.125 "nvme_admin": false, 00:16:06.125 "nvme_io": false, 00:16:06.125 "nvme_io_md": false, 00:16:06.125 "write_zeroes": true, 00:16:06.125 "zcopy": false, 00:16:06.125 "get_zone_info": false, 00:16:06.125 "zone_management": false, 00:16:06.125 "zone_append": false, 00:16:06.125 "compare": false, 00:16:06.125 "compare_and_write": false, 00:16:06.125 "abort": false, 00:16:06.125 "seek_hole": false, 00:16:06.125 "seek_data": false, 00:16:06.125 "copy": false, 00:16:06.125 "nvme_iov_md": false 00:16:06.125 }, 00:16:06.125 "memory_domains": [ 00:16:06.125 { 00:16:06.125 "dma_device_id": "system", 00:16:06.125 "dma_device_type": 1 00:16:06.125 }, 00:16:06.125 { 00:16:06.125 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:06.125 "dma_device_type": 2 00:16:06.125 }, 00:16:06.125 { 00:16:06.125 "dma_device_id": "system", 00:16:06.125 "dma_device_type": 1 00:16:06.125 }, 00:16:06.125 { 00:16:06.125 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:06.125 "dma_device_type": 2 00:16:06.125 } 00:16:06.125 ], 00:16:06.125 "driver_specific": { 00:16:06.125 "raid": { 00:16:06.125 "uuid": "bcbc5ee9-d373-4e2d-ac11-c480a3a0f00d", 00:16:06.125 "strip_size_kb": 0, 00:16:06.125 "state": "online", 00:16:06.125 "raid_level": "raid1", 00:16:06.125 "superblock": true, 00:16:06.125 "num_base_bdevs": 2, 00:16:06.125 "num_base_bdevs_discovered": 2, 00:16:06.125 "num_base_bdevs_operational": 2, 00:16:06.125 "base_bdevs_list": [ 00:16:06.125 { 00:16:06.125 "name": "pt1", 00:16:06.125 "uuid": "00000000-0000-0000-0000-000000000001", 00:16:06.125 "is_configured": true, 00:16:06.125 "data_offset": 256, 00:16:06.125 "data_size": 7936 00:16:06.125 }, 00:16:06.125 { 00:16:06.125 "name": "pt2", 00:16:06.125 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:06.125 "is_configured": true, 00:16:06.125 "data_offset": 256, 00:16:06.125 "data_size": 7936 00:16:06.125 } 00:16:06.125 ] 00:16:06.125 } 00:16:06.125 } 00:16:06.125 }' 00:16:06.125 17:37:37 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:16:06.386 17:37:37 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:16:06.386 pt2' 00:16:06.386 17:37:37 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:06.386 17:37:37 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4096 ' 00:16:06.386 17:37:37 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:16:06.386 17:37:37 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:06.386 17:37:37 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:16:06.386 17:37:37 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:06.386 17:37:37 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:06.386 17:37:37 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:06.386 17:37:37 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 ' 00:16:06.386 17:37:37 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@193 -- # [[ 4096 == \4\0\9\6\ \ \ ]] 00:16:06.386 17:37:37 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:16:06.386 17:37:37 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:16:06.386 17:37:37 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:06.386 17:37:37 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:06.386 17:37:37 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:06.386 17:37:37 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:06.386 17:37:37 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 ' 00:16:06.386 17:37:37 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@193 -- # [[ 4096 == \4\0\9\6\ \ \ ]] 00:16:06.386 17:37:37 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:16:06.386 17:37:37 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:06.386 17:37:37 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:06.386 17:37:37 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:16:06.386 [2024-11-27 17:37:37.480261] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:06.386 17:37:37 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:06.386 17:37:37 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@487 -- # '[' bcbc5ee9-d373-4e2d-ac11-c480a3a0f00d '!=' bcbc5ee9-d373-4e2d-ac11-c480a3a0f00d ']' 00:16:06.386 17:37:37 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@491 -- # has_redundancy raid1 00:16:06.386 17:37:37 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@198 -- # case $1 in 00:16:06.386 17:37:37 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@199 -- # return 0 00:16:06.386 17:37:37 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@493 -- # rpc_cmd bdev_passthru_delete pt1 00:16:06.386 17:37:37 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:06.386 17:37:37 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:06.386 [2024-11-27 17:37:37.527992] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: pt1 00:16:06.386 17:37:37 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:06.386 17:37:37 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@496 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:06.386 17:37:37 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:06.386 17:37:37 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:06.386 17:37:37 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:06.386 17:37:37 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:06.386 17:37:37 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:06.386 17:37:37 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:06.386 17:37:37 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:06.386 17:37:37 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:06.386 17:37:37 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:06.386 17:37:37 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:06.386 17:37:37 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:06.386 17:37:37 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:06.386 17:37:37 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:06.386 17:37:37 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:06.386 17:37:37 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:06.386 "name": "raid_bdev1", 00:16:06.386 "uuid": "bcbc5ee9-d373-4e2d-ac11-c480a3a0f00d", 00:16:06.386 "strip_size_kb": 0, 00:16:06.386 "state": "online", 00:16:06.386 "raid_level": "raid1", 00:16:06.386 "superblock": true, 00:16:06.386 "num_base_bdevs": 2, 00:16:06.386 "num_base_bdevs_discovered": 1, 00:16:06.386 "num_base_bdevs_operational": 1, 00:16:06.386 "base_bdevs_list": [ 00:16:06.386 { 00:16:06.386 "name": null, 00:16:06.386 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:06.386 "is_configured": false, 00:16:06.386 "data_offset": 0, 00:16:06.386 "data_size": 7936 00:16:06.386 }, 00:16:06.386 { 00:16:06.386 "name": "pt2", 00:16:06.386 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:06.386 "is_configured": true, 00:16:06.386 "data_offset": 256, 00:16:06.386 "data_size": 7936 00:16:06.386 } 00:16:06.386 ] 00:16:06.386 }' 00:16:06.386 17:37:37 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:06.386 17:37:37 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:06.957 17:37:37 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@499 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:16:06.957 17:37:37 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:06.957 17:37:37 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:06.957 [2024-11-27 17:37:37.963201] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:16:06.957 [2024-11-27 17:37:37.963228] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:16:06.958 [2024-11-27 17:37:37.963286] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:06.958 [2024-11-27 17:37:37.963325] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:06.958 [2024-11-27 17:37:37.963333] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:16:06.958 17:37:37 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:06.958 17:37:37 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@500 -- # jq -r '.[]' 00:16:06.958 17:37:37 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@500 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:06.958 17:37:37 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:06.958 17:37:37 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:06.958 17:37:37 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:06.958 17:37:37 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@500 -- # raid_bdev= 00:16:06.958 17:37:37 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@501 -- # '[' -n '' ']' 00:16:06.958 17:37:37 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@506 -- # (( i = 1 )) 00:16:06.958 17:37:37 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:16:06.958 17:37:37 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt2 00:16:06.958 17:37:37 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:06.958 17:37:37 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:06.958 17:37:38 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:06.958 17:37:38 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:16:06.958 17:37:38 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:16:06.958 17:37:38 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@511 -- # (( i = 1 )) 00:16:06.958 17:37:38 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:16:06.958 17:37:38 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@519 -- # i=1 00:16:06.958 17:37:38 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@520 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:16:06.958 17:37:38 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:06.958 17:37:38 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:06.958 [2024-11-27 17:37:38.015132] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:16:06.958 [2024-11-27 17:37:38.015189] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:06.958 [2024-11-27 17:37:38.015207] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008780 00:16:06.958 [2024-11-27 17:37:38.015216] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:06.958 [2024-11-27 17:37:38.017156] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:06.958 [2024-11-27 17:37:38.017199] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:16:06.958 [2024-11-27 17:37:38.017277] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:16:06.958 [2024-11-27 17:37:38.017304] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:16:06.958 [2024-11-27 17:37:38.017369] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:16:06.958 [2024-11-27 17:37:38.017376] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:16:06.958 [2024-11-27 17:37:38.017576] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:16:06.958 [2024-11-27 17:37:38.017683] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:16:06.958 [2024-11-27 17:37:38.017697] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001c80 00:16:06.958 [2024-11-27 17:37:38.017786] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:06.958 pt2 00:16:06.958 17:37:38 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:06.958 17:37:38 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@523 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:06.958 17:37:38 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:06.958 17:37:38 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:06.958 17:37:38 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:06.958 17:37:38 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:06.958 17:37:38 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:06.958 17:37:38 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:06.958 17:37:38 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:06.958 17:37:38 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:06.958 17:37:38 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:06.958 17:37:38 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:06.958 17:37:38 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:06.958 17:37:38 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:06.958 17:37:38 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:06.958 17:37:38 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:06.958 17:37:38 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:06.958 "name": "raid_bdev1", 00:16:06.958 "uuid": "bcbc5ee9-d373-4e2d-ac11-c480a3a0f00d", 00:16:06.958 "strip_size_kb": 0, 00:16:06.958 "state": "online", 00:16:06.958 "raid_level": "raid1", 00:16:06.958 "superblock": true, 00:16:06.958 "num_base_bdevs": 2, 00:16:06.958 "num_base_bdevs_discovered": 1, 00:16:06.958 "num_base_bdevs_operational": 1, 00:16:06.958 "base_bdevs_list": [ 00:16:06.958 { 00:16:06.958 "name": null, 00:16:06.958 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:06.958 "is_configured": false, 00:16:06.958 "data_offset": 256, 00:16:06.958 "data_size": 7936 00:16:06.958 }, 00:16:06.958 { 00:16:06.958 "name": "pt2", 00:16:06.958 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:06.958 "is_configured": true, 00:16:06.958 "data_offset": 256, 00:16:06.958 "data_size": 7936 00:16:06.958 } 00:16:06.958 ] 00:16:06.958 }' 00:16:06.958 17:37:38 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:06.958 17:37:38 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:07.528 17:37:38 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@526 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:16:07.528 17:37:38 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:07.528 17:37:38 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:07.529 [2024-11-27 17:37:38.442378] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:16:07.529 [2024-11-27 17:37:38.442401] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:16:07.529 [2024-11-27 17:37:38.442450] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:07.529 [2024-11-27 17:37:38.442484] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:07.529 [2024-11-27 17:37:38.442493] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name raid_bdev1, state offline 00:16:07.529 17:37:38 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:07.529 17:37:38 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@527 -- # jq -r '.[]' 00:16:07.529 17:37:38 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@527 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:07.529 17:37:38 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:07.529 17:37:38 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:07.529 17:37:38 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:07.529 17:37:38 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@527 -- # raid_bdev= 00:16:07.529 17:37:38 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@528 -- # '[' -n '' ']' 00:16:07.529 17:37:38 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@532 -- # '[' 2 -gt 2 ']' 00:16:07.529 17:37:38 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@540 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:16:07.529 17:37:38 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:07.529 17:37:38 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:07.529 [2024-11-27 17:37:38.486331] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:16:07.529 [2024-11-27 17:37:38.486394] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:07.529 [2024-11-27 17:37:38.486407] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008d80 00:16:07.529 [2024-11-27 17:37:38.486419] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:07.529 [2024-11-27 17:37:38.488329] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:07.529 [2024-11-27 17:37:38.488366] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:16:07.529 [2024-11-27 17:37:38.488450] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:16:07.529 [2024-11-27 17:37:38.488492] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:16:07.529 [2024-11-27 17:37:38.488575] bdev_raid.c:3675:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev pt2 (4) greater than existing raid bdev raid_bdev1 (2) 00:16:07.529 [2024-11-27 17:37:38.488586] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:16:07.529 [2024-11-27 17:37:38.488600] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002000 name raid_bdev1, state configuring 00:16:07.529 [2024-11-27 17:37:38.488626] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:16:07.529 [2024-11-27 17:37:38.488692] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000002380 00:16:07.529 [2024-11-27 17:37:38.488705] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:16:07.529 [2024-11-27 17:37:38.488894] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:16:07.529 [2024-11-27 17:37:38.489010] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000002380 00:16:07.529 [2024-11-27 17:37:38.489024] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000002380 00:16:07.529 [2024-11-27 17:37:38.489123] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:07.529 pt1 00:16:07.529 17:37:38 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:07.529 17:37:38 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@542 -- # '[' 2 -gt 2 ']' 00:16:07.529 17:37:38 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@554 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:07.529 17:37:38 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:07.529 17:37:38 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:07.529 17:37:38 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:07.529 17:37:38 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:07.529 17:37:38 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:07.529 17:37:38 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:07.529 17:37:38 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:07.529 17:37:38 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:07.529 17:37:38 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:07.529 17:37:38 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:07.529 17:37:38 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:07.529 17:37:38 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:07.529 17:37:38 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:07.529 17:37:38 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:07.529 17:37:38 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:07.529 "name": "raid_bdev1", 00:16:07.529 "uuid": "bcbc5ee9-d373-4e2d-ac11-c480a3a0f00d", 00:16:07.529 "strip_size_kb": 0, 00:16:07.529 "state": "online", 00:16:07.529 "raid_level": "raid1", 00:16:07.529 "superblock": true, 00:16:07.529 "num_base_bdevs": 2, 00:16:07.529 "num_base_bdevs_discovered": 1, 00:16:07.529 "num_base_bdevs_operational": 1, 00:16:07.529 "base_bdevs_list": [ 00:16:07.529 { 00:16:07.529 "name": null, 00:16:07.529 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:07.529 "is_configured": false, 00:16:07.529 "data_offset": 256, 00:16:07.529 "data_size": 7936 00:16:07.529 }, 00:16:07.529 { 00:16:07.529 "name": "pt2", 00:16:07.529 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:07.529 "is_configured": true, 00:16:07.529 "data_offset": 256, 00:16:07.529 "data_size": 7936 00:16:07.529 } 00:16:07.529 ] 00:16:07.529 }' 00:16:07.529 17:37:38 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:07.529 17:37:38 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:07.790 17:37:38 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@555 -- # rpc_cmd bdev_raid_get_bdevs online 00:16:07.790 17:37:38 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:07.790 17:37:38 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@555 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:16:07.790 17:37:38 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:07.790 17:37:38 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:07.790 17:37:38 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@555 -- # [[ false == \f\a\l\s\e ]] 00:16:07.790 17:37:38 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@558 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:16:07.790 17:37:38 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@558 -- # jq -r '.[] | .uuid' 00:16:07.790 17:37:38 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:07.790 17:37:38 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:07.790 [2024-11-27 17:37:38.965716] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:08.050 17:37:38 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:08.050 17:37:39 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@558 -- # '[' bcbc5ee9-d373-4e2d-ac11-c480a3a0f00d '!=' bcbc5ee9-d373-4e2d-ac11-c480a3a0f00d ']' 00:16:08.050 17:37:39 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@563 -- # killprocess 96411 00:16:08.050 17:37:39 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@950 -- # '[' -z 96411 ']' 00:16:08.050 17:37:39 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@954 -- # kill -0 96411 00:16:08.050 17:37:39 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@955 -- # uname 00:16:08.050 17:37:39 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:16:08.050 17:37:39 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 96411 00:16:08.050 17:37:39 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:16:08.050 17:37:39 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:16:08.050 17:37:39 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@968 -- # echo 'killing process with pid 96411' 00:16:08.050 killing process with pid 96411 00:16:08.050 17:37:39 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@969 -- # kill 96411 00:16:08.050 [2024-11-27 17:37:39.048420] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:16:08.050 [2024-11-27 17:37:39.048488] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:08.050 [2024-11-27 17:37:39.048529] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:08.050 [2024-11-27 17:37:39.048539] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002380 name raid_bdev1, state offline 00:16:08.050 17:37:39 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@974 -- # wait 96411 00:16:08.050 [2024-11-27 17:37:39.071017] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:16:08.309 17:37:39 bdev_raid.raid_superblock_test_4k -- bdev/bdev_raid.sh@565 -- # return 0 00:16:08.309 00:16:08.310 real 0m4.811s 00:16:08.310 user 0m7.835s 00:16:08.310 sys 0m1.035s 00:16:08.310 17:37:39 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@1126 -- # xtrace_disable 00:16:08.310 17:37:39 bdev_raid.raid_superblock_test_4k -- common/autotest_common.sh@10 -- # set +x 00:16:08.310 ************************************ 00:16:08.310 END TEST raid_superblock_test_4k 00:16:08.310 ************************************ 00:16:08.310 17:37:39 bdev_raid -- bdev/bdev_raid.sh@999 -- # '[' true = true ']' 00:16:08.310 17:37:39 bdev_raid -- bdev/bdev_raid.sh@1000 -- # run_test raid_rebuild_test_sb_4k raid_rebuild_test raid1 2 true false true 00:16:08.310 17:37:39 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 7 -le 1 ']' 00:16:08.310 17:37:39 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:16:08.310 17:37:39 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:16:08.310 ************************************ 00:16:08.310 START TEST raid_rebuild_test_sb_4k 00:16:08.310 ************************************ 00:16:08.310 17:37:39 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@1125 -- # raid_rebuild_test raid1 2 true false true 00:16:08.310 17:37:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:16:08.310 17:37:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=2 00:16:08.310 17:37:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:16:08.310 17:37:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:16:08.310 17:37:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@573 -- # local verify=true 00:16:08.310 17:37:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:16:08.310 17:37:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:16:08.310 17:37:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:16:08.310 17:37:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:16:08.310 17:37:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:16:08.310 17:37:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:16:08.310 17:37:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:16:08.310 17:37:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:16:08.310 17:37:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:16:08.310 17:37:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:16:08.310 17:37:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:16:08.310 17:37:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@576 -- # local strip_size 00:16:08.310 17:37:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@577 -- # local create_arg 00:16:08.310 17:37:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:16:08.310 17:37:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@579 -- # local data_offset 00:16:08.310 17:37:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:16:08.310 17:37:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:16:08.310 17:37:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:16:08.310 17:37:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:16:08.310 17:37:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@597 -- # raid_pid=96717 00:16:08.310 17:37:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:16:08.310 17:37:39 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@598 -- # waitforlisten 96717 00:16:08.310 17:37:39 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@831 -- # '[' -z 96717 ']' 00:16:08.310 17:37:39 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:16:08.310 17:37:39 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@836 -- # local max_retries=100 00:16:08.310 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:16:08.310 17:37:39 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:16:08.310 17:37:39 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@840 -- # xtrace_disable 00:16:08.310 17:37:39 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:08.310 I/O size of 3145728 is greater than zero copy threshold (65536). 00:16:08.310 Zero copy mechanism will not be used. 00:16:08.310 [2024-11-27 17:37:39.476560] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:16:08.310 [2024-11-27 17:37:39.476687] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid96717 ] 00:16:08.569 [2024-11-27 17:37:39.611822] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:16:08.569 [2024-11-27 17:37:39.656186] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:16:08.569 [2024-11-27 17:37:39.697555] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:16:08.569 [2024-11-27 17:37:39.697591] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:16:09.139 17:37:40 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:16:09.139 17:37:40 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@864 -- # return 0 00:16:09.139 17:37:40 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:16:09.139 17:37:40 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 4096 -b BaseBdev1_malloc 00:16:09.139 17:37:40 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:09.139 17:37:40 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:09.139 BaseBdev1_malloc 00:16:09.139 17:37:40 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:09.139 17:37:40 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:16:09.139 17:37:40 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:09.139 17:37:40 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:09.139 [2024-11-27 17:37:40.310862] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:16:09.139 [2024-11-27 17:37:40.310930] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:09.139 [2024-11-27 17:37:40.310959] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:16:09.139 [2024-11-27 17:37:40.310975] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:09.139 [2024-11-27 17:37:40.312954] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:09.139 [2024-11-27 17:37:40.312988] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:16:09.139 BaseBdev1 00:16:09.139 17:37:40 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:09.139 17:37:40 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:16:09.139 17:37:40 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 4096 -b BaseBdev2_malloc 00:16:09.139 17:37:40 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:09.139 17:37:40 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:09.399 BaseBdev2_malloc 00:16:09.399 17:37:40 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:09.399 17:37:40 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:16:09.399 17:37:40 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:09.399 17:37:40 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:09.399 [2024-11-27 17:37:40.347441] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:16:09.399 [2024-11-27 17:37:40.347490] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:09.399 [2024-11-27 17:37:40.347510] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:16:09.399 [2024-11-27 17:37:40.347519] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:09.399 [2024-11-27 17:37:40.349518] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:09.399 [2024-11-27 17:37:40.349552] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:16:09.399 BaseBdev2 00:16:09.399 17:37:40 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:09.399 17:37:40 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 4096 -b spare_malloc 00:16:09.399 17:37:40 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:09.399 17:37:40 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:09.399 spare_malloc 00:16:09.399 17:37:40 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:09.399 17:37:40 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:16:09.399 17:37:40 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:09.399 17:37:40 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:09.399 spare_delay 00:16:09.399 17:37:40 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:09.399 17:37:40 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:16:09.399 17:37:40 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:09.399 17:37:40 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:09.399 [2024-11-27 17:37:40.387724] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:16:09.399 [2024-11-27 17:37:40.387777] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:09.399 [2024-11-27 17:37:40.387797] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:16:09.399 [2024-11-27 17:37:40.387806] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:09.399 [2024-11-27 17:37:40.389696] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:09.399 [2024-11-27 17:37:40.389729] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:16:09.399 spare 00:16:09.399 17:37:40 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:09.399 17:37:40 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 00:16:09.399 17:37:40 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:09.399 17:37:40 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:09.399 [2024-11-27 17:37:40.399757] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:16:09.399 [2024-11-27 17:37:40.401434] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:16:09.399 [2024-11-27 17:37:40.401597] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:16:09.399 [2024-11-27 17:37:40.401609] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:16:09.399 [2024-11-27 17:37:40.401863] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:16:09.399 [2024-11-27 17:37:40.401990] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:16:09.399 [2024-11-27 17:37:40.402009] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:16:09.400 [2024-11-27 17:37:40.402127] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:09.400 17:37:40 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:09.400 17:37:40 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:16:09.400 17:37:40 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:09.400 17:37:40 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:09.400 17:37:40 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:09.400 17:37:40 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:09.400 17:37:40 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:09.400 17:37:40 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:09.400 17:37:40 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:09.400 17:37:40 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:09.400 17:37:40 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:09.400 17:37:40 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:09.400 17:37:40 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:09.400 17:37:40 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:09.400 17:37:40 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:09.400 17:37:40 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:09.400 17:37:40 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:09.400 "name": "raid_bdev1", 00:16:09.400 "uuid": "ca18a2f3-1538-4867-ae49-d203a82ee759", 00:16:09.400 "strip_size_kb": 0, 00:16:09.400 "state": "online", 00:16:09.400 "raid_level": "raid1", 00:16:09.400 "superblock": true, 00:16:09.400 "num_base_bdevs": 2, 00:16:09.400 "num_base_bdevs_discovered": 2, 00:16:09.400 "num_base_bdevs_operational": 2, 00:16:09.400 "base_bdevs_list": [ 00:16:09.400 { 00:16:09.400 "name": "BaseBdev1", 00:16:09.400 "uuid": "abc5eb4a-6ac8-547b-8a79-e710a46c7197", 00:16:09.400 "is_configured": true, 00:16:09.400 "data_offset": 256, 00:16:09.400 "data_size": 7936 00:16:09.400 }, 00:16:09.400 { 00:16:09.400 "name": "BaseBdev2", 00:16:09.400 "uuid": "8da903e4-ba16-5f8d-8a8f-8ec78df3de55", 00:16:09.400 "is_configured": true, 00:16:09.400 "data_offset": 256, 00:16:09.400 "data_size": 7936 00:16:09.400 } 00:16:09.400 ] 00:16:09.400 }' 00:16:09.400 17:37:40 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:09.400 17:37:40 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:09.660 17:37:40 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:16:09.660 17:37:40 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:09.660 17:37:40 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:09.660 17:37:40 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:16:09.660 [2024-11-27 17:37:40.787417] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:09.660 17:37:40 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:09.660 17:37:40 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=7936 00:16:09.660 17:37:40 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:16:09.660 17:37:40 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:09.660 17:37:40 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:09.660 17:37:40 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:09.660 17:37:40 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:09.921 17:37:40 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@619 -- # data_offset=256 00:16:09.921 17:37:40 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:16:09.921 17:37:40 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:16:09.921 17:37:40 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:16:09.921 17:37:40 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:16:09.921 17:37:40 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:16:09.921 17:37:40 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:16:09.921 17:37:40 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@10 -- # local bdev_list 00:16:09.921 17:37:40 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:16:09.921 17:37:40 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@11 -- # local nbd_list 00:16:09.921 17:37:40 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@12 -- # local i 00:16:09.921 17:37:40 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:16:09.921 17:37:40 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:16:09.921 17:37:40 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:16:09.921 [2024-11-27 17:37:41.054745] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:16:09.921 /dev/nbd0 00:16:09.921 17:37:41 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:16:09.921 17:37:41 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:16:09.921 17:37:41 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:16:09.921 17:37:41 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@869 -- # local i 00:16:09.921 17:37:41 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:16:09.921 17:37:41 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:16:09.921 17:37:41 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:16:09.921 17:37:41 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@873 -- # break 00:16:09.921 17:37:41 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:16:09.921 17:37:41 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:16:09.921 17:37:41 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:16:09.921 1+0 records in 00:16:09.921 1+0 records out 00:16:09.921 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000398377 s, 10.3 MB/s 00:16:10.181 17:37:41 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:16:10.181 17:37:41 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@886 -- # size=4096 00:16:10.181 17:37:41 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:16:10.181 17:37:41 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:16:10.181 17:37:41 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@889 -- # return 0 00:16:10.181 17:37:41 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:16:10.181 17:37:41 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:16:10.181 17:37:41 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@629 -- # '[' raid1 = raid5f ']' 00:16:10.181 17:37:41 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@633 -- # write_unit_size=1 00:16:10.181 17:37:41 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=4096 count=7936 oflag=direct 00:16:10.750 7936+0 records in 00:16:10.750 7936+0 records out 00:16:10.750 32505856 bytes (33 MB, 31 MiB) copied, 0.585655 s, 55.5 MB/s 00:16:10.750 17:37:41 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:16:10.750 17:37:41 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:16:10.750 17:37:41 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:16:10.750 17:37:41 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@50 -- # local nbd_list 00:16:10.750 17:37:41 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@51 -- # local i 00:16:10.750 17:37:41 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:16:10.750 17:37:41 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:16:10.750 17:37:41 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:16:10.750 17:37:41 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:16:10.750 [2024-11-27 17:37:41.917122] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:10.750 17:37:41 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:16:10.750 17:37:41 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:16:10.750 17:37:41 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:16:10.750 17:37:41 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:16:10.750 17:37:41 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@41 -- # break 00:16:10.750 17:37:41 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@45 -- # return 0 00:16:10.750 17:37:41 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:16:10.750 17:37:41 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:10.750 17:37:41 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:10.750 [2024-11-27 17:37:41.933199] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:16:10.750 17:37:41 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:10.750 17:37:41 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:10.750 17:37:41 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:10.750 17:37:41 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:10.750 17:37:41 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:10.751 17:37:41 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:11.011 17:37:41 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:11.011 17:37:41 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:11.011 17:37:41 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:11.011 17:37:41 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:11.011 17:37:41 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:11.011 17:37:41 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:11.011 17:37:41 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:11.011 17:37:41 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:11.011 17:37:41 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:11.011 17:37:41 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:11.011 17:37:41 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:11.011 "name": "raid_bdev1", 00:16:11.011 "uuid": "ca18a2f3-1538-4867-ae49-d203a82ee759", 00:16:11.011 "strip_size_kb": 0, 00:16:11.011 "state": "online", 00:16:11.011 "raid_level": "raid1", 00:16:11.011 "superblock": true, 00:16:11.011 "num_base_bdevs": 2, 00:16:11.011 "num_base_bdevs_discovered": 1, 00:16:11.011 "num_base_bdevs_operational": 1, 00:16:11.011 "base_bdevs_list": [ 00:16:11.011 { 00:16:11.011 "name": null, 00:16:11.011 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:11.011 "is_configured": false, 00:16:11.011 "data_offset": 0, 00:16:11.011 "data_size": 7936 00:16:11.011 }, 00:16:11.011 { 00:16:11.011 "name": "BaseBdev2", 00:16:11.011 "uuid": "8da903e4-ba16-5f8d-8a8f-8ec78df3de55", 00:16:11.011 "is_configured": true, 00:16:11.011 "data_offset": 256, 00:16:11.011 "data_size": 7936 00:16:11.011 } 00:16:11.011 ] 00:16:11.011 }' 00:16:11.011 17:37:41 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:11.011 17:37:41 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:11.271 17:37:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:16:11.271 17:37:42 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:11.271 17:37:42 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:11.271 [2024-11-27 17:37:42.384388] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:16:11.271 [2024-11-27 17:37:42.388561] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d00019c960 00:16:11.271 17:37:42 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:11.271 17:37:42 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@647 -- # sleep 1 00:16:11.271 [2024-11-27 17:37:42.390433] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:16:12.209 17:37:43 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:12.209 17:37:43 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:12.209 17:37:43 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:12.209 17:37:43 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:12.209 17:37:43 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:12.470 17:37:43 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:12.470 17:37:43 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:12.470 17:37:43 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:12.470 17:37:43 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:12.470 17:37:43 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:12.470 17:37:43 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:12.470 "name": "raid_bdev1", 00:16:12.470 "uuid": "ca18a2f3-1538-4867-ae49-d203a82ee759", 00:16:12.470 "strip_size_kb": 0, 00:16:12.470 "state": "online", 00:16:12.470 "raid_level": "raid1", 00:16:12.470 "superblock": true, 00:16:12.470 "num_base_bdevs": 2, 00:16:12.470 "num_base_bdevs_discovered": 2, 00:16:12.470 "num_base_bdevs_operational": 2, 00:16:12.470 "process": { 00:16:12.470 "type": "rebuild", 00:16:12.470 "target": "spare", 00:16:12.470 "progress": { 00:16:12.470 "blocks": 2560, 00:16:12.470 "percent": 32 00:16:12.470 } 00:16:12.470 }, 00:16:12.470 "base_bdevs_list": [ 00:16:12.470 { 00:16:12.470 "name": "spare", 00:16:12.470 "uuid": "470cfe62-9599-5986-90ee-67afdd4359fa", 00:16:12.470 "is_configured": true, 00:16:12.470 "data_offset": 256, 00:16:12.470 "data_size": 7936 00:16:12.470 }, 00:16:12.470 { 00:16:12.470 "name": "BaseBdev2", 00:16:12.470 "uuid": "8da903e4-ba16-5f8d-8a8f-8ec78df3de55", 00:16:12.470 "is_configured": true, 00:16:12.470 "data_offset": 256, 00:16:12.470 "data_size": 7936 00:16:12.470 } 00:16:12.470 ] 00:16:12.470 }' 00:16:12.470 17:37:43 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:12.470 17:37:43 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:12.470 17:37:43 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:12.470 17:37:43 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:12.470 17:37:43 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:16:12.470 17:37:43 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:12.470 17:37:43 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:12.470 [2024-11-27 17:37:43.554935] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:12.470 [2024-11-27 17:37:43.594836] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:16:12.470 [2024-11-27 17:37:43.594888] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:12.470 [2024-11-27 17:37:43.594905] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:12.470 [2024-11-27 17:37:43.594912] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:16:12.470 17:37:43 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:12.470 17:37:43 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:12.470 17:37:43 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:12.470 17:37:43 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:12.470 17:37:43 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:12.470 17:37:43 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:12.470 17:37:43 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:12.470 17:37:43 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:12.470 17:37:43 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:12.470 17:37:43 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:12.470 17:37:43 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:12.470 17:37:43 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:12.470 17:37:43 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:12.470 17:37:43 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:12.470 17:37:43 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:12.470 17:37:43 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:12.470 17:37:43 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:12.470 "name": "raid_bdev1", 00:16:12.470 "uuid": "ca18a2f3-1538-4867-ae49-d203a82ee759", 00:16:12.470 "strip_size_kb": 0, 00:16:12.470 "state": "online", 00:16:12.470 "raid_level": "raid1", 00:16:12.470 "superblock": true, 00:16:12.470 "num_base_bdevs": 2, 00:16:12.470 "num_base_bdevs_discovered": 1, 00:16:12.470 "num_base_bdevs_operational": 1, 00:16:12.470 "base_bdevs_list": [ 00:16:12.470 { 00:16:12.470 "name": null, 00:16:12.470 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:12.470 "is_configured": false, 00:16:12.470 "data_offset": 0, 00:16:12.470 "data_size": 7936 00:16:12.470 }, 00:16:12.470 { 00:16:12.470 "name": "BaseBdev2", 00:16:12.470 "uuid": "8da903e4-ba16-5f8d-8a8f-8ec78df3de55", 00:16:12.470 "is_configured": true, 00:16:12.470 "data_offset": 256, 00:16:12.471 "data_size": 7936 00:16:12.471 } 00:16:12.471 ] 00:16:12.471 }' 00:16:12.471 17:37:43 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:12.471 17:37:43 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:13.041 17:37:44 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:16:13.041 17:37:44 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:13.041 17:37:44 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:16:13.041 17:37:44 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=none 00:16:13.041 17:37:44 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:13.041 17:37:44 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:13.041 17:37:44 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:13.041 17:37:44 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:13.041 17:37:44 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:13.041 17:37:44 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:13.041 17:37:44 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:13.041 "name": "raid_bdev1", 00:16:13.041 "uuid": "ca18a2f3-1538-4867-ae49-d203a82ee759", 00:16:13.041 "strip_size_kb": 0, 00:16:13.041 "state": "online", 00:16:13.041 "raid_level": "raid1", 00:16:13.041 "superblock": true, 00:16:13.041 "num_base_bdevs": 2, 00:16:13.041 "num_base_bdevs_discovered": 1, 00:16:13.041 "num_base_bdevs_operational": 1, 00:16:13.041 "base_bdevs_list": [ 00:16:13.041 { 00:16:13.041 "name": null, 00:16:13.041 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:13.041 "is_configured": false, 00:16:13.041 "data_offset": 0, 00:16:13.041 "data_size": 7936 00:16:13.041 }, 00:16:13.041 { 00:16:13.041 "name": "BaseBdev2", 00:16:13.041 "uuid": "8da903e4-ba16-5f8d-8a8f-8ec78df3de55", 00:16:13.041 "is_configured": true, 00:16:13.041 "data_offset": 256, 00:16:13.041 "data_size": 7936 00:16:13.041 } 00:16:13.041 ] 00:16:13.041 }' 00:16:13.041 17:37:44 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:13.041 17:37:44 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:16:13.041 17:37:44 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:13.041 17:37:44 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:16:13.041 17:37:44 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:16:13.041 17:37:44 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:13.041 17:37:44 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:13.041 [2024-11-27 17:37:44.174051] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:16:13.041 [2024-11-27 17:37:44.177585] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d00019ca30 00:16:13.041 [2024-11-27 17:37:44.179373] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:16:13.041 17:37:44 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:13.041 17:37:44 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@663 -- # sleep 1 00:16:14.010 17:37:45 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:14.011 17:37:45 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:14.011 17:37:45 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:14.011 17:37:45 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:14.011 17:37:45 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:14.011 17:37:45 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:14.011 17:37:45 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:14.011 17:37:45 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:14.011 17:37:45 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:14.271 17:37:45 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:14.271 17:37:45 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:14.271 "name": "raid_bdev1", 00:16:14.271 "uuid": "ca18a2f3-1538-4867-ae49-d203a82ee759", 00:16:14.271 "strip_size_kb": 0, 00:16:14.271 "state": "online", 00:16:14.271 "raid_level": "raid1", 00:16:14.271 "superblock": true, 00:16:14.271 "num_base_bdevs": 2, 00:16:14.271 "num_base_bdevs_discovered": 2, 00:16:14.271 "num_base_bdevs_operational": 2, 00:16:14.271 "process": { 00:16:14.271 "type": "rebuild", 00:16:14.271 "target": "spare", 00:16:14.271 "progress": { 00:16:14.271 "blocks": 2560, 00:16:14.271 "percent": 32 00:16:14.271 } 00:16:14.271 }, 00:16:14.271 "base_bdevs_list": [ 00:16:14.271 { 00:16:14.271 "name": "spare", 00:16:14.271 "uuid": "470cfe62-9599-5986-90ee-67afdd4359fa", 00:16:14.271 "is_configured": true, 00:16:14.271 "data_offset": 256, 00:16:14.271 "data_size": 7936 00:16:14.271 }, 00:16:14.271 { 00:16:14.271 "name": "BaseBdev2", 00:16:14.271 "uuid": "8da903e4-ba16-5f8d-8a8f-8ec78df3de55", 00:16:14.271 "is_configured": true, 00:16:14.271 "data_offset": 256, 00:16:14.271 "data_size": 7936 00:16:14.271 } 00:16:14.271 ] 00:16:14.271 }' 00:16:14.271 17:37:45 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:14.271 17:37:45 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:14.271 17:37:45 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:14.271 17:37:45 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:14.271 17:37:45 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:16:14.271 17:37:45 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:16:14.271 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:16:14.271 17:37:45 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=2 00:16:14.271 17:37:45 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:16:14.271 17:37:45 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@693 -- # '[' 2 -gt 2 ']' 00:16:14.271 17:37:45 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@706 -- # local timeout=573 00:16:14.271 17:37:45 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:16:14.271 17:37:45 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:14.271 17:37:45 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:14.271 17:37:45 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:14.271 17:37:45 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:14.271 17:37:45 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:14.271 17:37:45 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:14.271 17:37:45 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:14.271 17:37:45 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:14.271 17:37:45 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:14.271 17:37:45 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:14.271 17:37:45 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:14.271 "name": "raid_bdev1", 00:16:14.271 "uuid": "ca18a2f3-1538-4867-ae49-d203a82ee759", 00:16:14.271 "strip_size_kb": 0, 00:16:14.271 "state": "online", 00:16:14.271 "raid_level": "raid1", 00:16:14.271 "superblock": true, 00:16:14.271 "num_base_bdevs": 2, 00:16:14.271 "num_base_bdevs_discovered": 2, 00:16:14.271 "num_base_bdevs_operational": 2, 00:16:14.271 "process": { 00:16:14.271 "type": "rebuild", 00:16:14.271 "target": "spare", 00:16:14.271 "progress": { 00:16:14.271 "blocks": 2816, 00:16:14.271 "percent": 35 00:16:14.271 } 00:16:14.271 }, 00:16:14.271 "base_bdevs_list": [ 00:16:14.271 { 00:16:14.271 "name": "spare", 00:16:14.271 "uuid": "470cfe62-9599-5986-90ee-67afdd4359fa", 00:16:14.271 "is_configured": true, 00:16:14.271 "data_offset": 256, 00:16:14.271 "data_size": 7936 00:16:14.271 }, 00:16:14.271 { 00:16:14.271 "name": "BaseBdev2", 00:16:14.271 "uuid": "8da903e4-ba16-5f8d-8a8f-8ec78df3de55", 00:16:14.271 "is_configured": true, 00:16:14.271 "data_offset": 256, 00:16:14.271 "data_size": 7936 00:16:14.271 } 00:16:14.271 ] 00:16:14.271 }' 00:16:14.271 17:37:45 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:14.271 17:37:45 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:14.271 17:37:45 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:14.531 17:37:45 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:14.531 17:37:45 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@711 -- # sleep 1 00:16:15.471 17:37:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:16:15.471 17:37:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:15.471 17:37:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:15.471 17:37:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:15.471 17:37:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:15.471 17:37:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:15.471 17:37:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:15.471 17:37:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:15.471 17:37:46 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:15.471 17:37:46 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:15.471 17:37:46 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:15.471 17:37:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:15.471 "name": "raid_bdev1", 00:16:15.471 "uuid": "ca18a2f3-1538-4867-ae49-d203a82ee759", 00:16:15.471 "strip_size_kb": 0, 00:16:15.471 "state": "online", 00:16:15.471 "raid_level": "raid1", 00:16:15.471 "superblock": true, 00:16:15.471 "num_base_bdevs": 2, 00:16:15.471 "num_base_bdevs_discovered": 2, 00:16:15.471 "num_base_bdevs_operational": 2, 00:16:15.471 "process": { 00:16:15.471 "type": "rebuild", 00:16:15.471 "target": "spare", 00:16:15.471 "progress": { 00:16:15.471 "blocks": 5888, 00:16:15.471 "percent": 74 00:16:15.471 } 00:16:15.471 }, 00:16:15.471 "base_bdevs_list": [ 00:16:15.471 { 00:16:15.471 "name": "spare", 00:16:15.471 "uuid": "470cfe62-9599-5986-90ee-67afdd4359fa", 00:16:15.471 "is_configured": true, 00:16:15.471 "data_offset": 256, 00:16:15.471 "data_size": 7936 00:16:15.471 }, 00:16:15.471 { 00:16:15.472 "name": "BaseBdev2", 00:16:15.472 "uuid": "8da903e4-ba16-5f8d-8a8f-8ec78df3de55", 00:16:15.472 "is_configured": true, 00:16:15.472 "data_offset": 256, 00:16:15.472 "data_size": 7936 00:16:15.472 } 00:16:15.472 ] 00:16:15.472 }' 00:16:15.472 17:37:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:15.472 17:37:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:15.472 17:37:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:15.472 17:37:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:15.472 17:37:46 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@711 -- # sleep 1 00:16:16.412 [2024-11-27 17:37:47.289189] bdev_raid.c:2896:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:16:16.412 [2024-11-27 17:37:47.289254] bdev_raid.c:2558:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:16:16.412 [2024-11-27 17:37:47.289382] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:16.673 17:37:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:16:16.673 17:37:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:16.673 17:37:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:16.673 17:37:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:16.673 17:37:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:16.673 17:37:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:16.673 17:37:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:16.673 17:37:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:16.673 17:37:47 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:16.673 17:37:47 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:16.673 17:37:47 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:16.673 17:37:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:16.673 "name": "raid_bdev1", 00:16:16.673 "uuid": "ca18a2f3-1538-4867-ae49-d203a82ee759", 00:16:16.673 "strip_size_kb": 0, 00:16:16.673 "state": "online", 00:16:16.673 "raid_level": "raid1", 00:16:16.673 "superblock": true, 00:16:16.673 "num_base_bdevs": 2, 00:16:16.673 "num_base_bdevs_discovered": 2, 00:16:16.673 "num_base_bdevs_operational": 2, 00:16:16.673 "base_bdevs_list": [ 00:16:16.673 { 00:16:16.673 "name": "spare", 00:16:16.673 "uuid": "470cfe62-9599-5986-90ee-67afdd4359fa", 00:16:16.673 "is_configured": true, 00:16:16.673 "data_offset": 256, 00:16:16.673 "data_size": 7936 00:16:16.673 }, 00:16:16.673 { 00:16:16.673 "name": "BaseBdev2", 00:16:16.673 "uuid": "8da903e4-ba16-5f8d-8a8f-8ec78df3de55", 00:16:16.673 "is_configured": true, 00:16:16.673 "data_offset": 256, 00:16:16.673 "data_size": 7936 00:16:16.673 } 00:16:16.673 ] 00:16:16.673 }' 00:16:16.673 17:37:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:16.673 17:37:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:16:16.673 17:37:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:16.673 17:37:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:16:16.673 17:37:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@709 -- # break 00:16:16.673 17:37:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:16:16.673 17:37:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:16.673 17:37:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:16:16.673 17:37:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=none 00:16:16.673 17:37:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:16.673 17:37:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:16.673 17:37:47 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:16.673 17:37:47 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:16.673 17:37:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:16.673 17:37:47 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:16.673 17:37:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:16.673 "name": "raid_bdev1", 00:16:16.673 "uuid": "ca18a2f3-1538-4867-ae49-d203a82ee759", 00:16:16.673 "strip_size_kb": 0, 00:16:16.673 "state": "online", 00:16:16.673 "raid_level": "raid1", 00:16:16.673 "superblock": true, 00:16:16.673 "num_base_bdevs": 2, 00:16:16.673 "num_base_bdevs_discovered": 2, 00:16:16.673 "num_base_bdevs_operational": 2, 00:16:16.673 "base_bdevs_list": [ 00:16:16.674 { 00:16:16.674 "name": "spare", 00:16:16.674 "uuid": "470cfe62-9599-5986-90ee-67afdd4359fa", 00:16:16.674 "is_configured": true, 00:16:16.674 "data_offset": 256, 00:16:16.674 "data_size": 7936 00:16:16.674 }, 00:16:16.674 { 00:16:16.674 "name": "BaseBdev2", 00:16:16.674 "uuid": "8da903e4-ba16-5f8d-8a8f-8ec78df3de55", 00:16:16.674 "is_configured": true, 00:16:16.674 "data_offset": 256, 00:16:16.674 "data_size": 7936 00:16:16.674 } 00:16:16.674 ] 00:16:16.674 }' 00:16:16.674 17:37:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:16.674 17:37:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:16:16.674 17:37:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:16.934 17:37:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:16:16.934 17:37:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:16:16.934 17:37:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:16.934 17:37:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:16.934 17:37:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:16.934 17:37:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:16.934 17:37:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:16.934 17:37:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:16.934 17:37:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:16.934 17:37:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:16.934 17:37:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:16.934 17:37:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:16.934 17:37:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:16.934 17:37:47 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:16.934 17:37:47 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:16.934 17:37:47 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:16.934 17:37:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:16.934 "name": "raid_bdev1", 00:16:16.934 "uuid": "ca18a2f3-1538-4867-ae49-d203a82ee759", 00:16:16.934 "strip_size_kb": 0, 00:16:16.934 "state": "online", 00:16:16.934 "raid_level": "raid1", 00:16:16.934 "superblock": true, 00:16:16.934 "num_base_bdevs": 2, 00:16:16.934 "num_base_bdevs_discovered": 2, 00:16:16.934 "num_base_bdevs_operational": 2, 00:16:16.934 "base_bdevs_list": [ 00:16:16.934 { 00:16:16.934 "name": "spare", 00:16:16.934 "uuid": "470cfe62-9599-5986-90ee-67afdd4359fa", 00:16:16.934 "is_configured": true, 00:16:16.934 "data_offset": 256, 00:16:16.934 "data_size": 7936 00:16:16.934 }, 00:16:16.934 { 00:16:16.934 "name": "BaseBdev2", 00:16:16.934 "uuid": "8da903e4-ba16-5f8d-8a8f-8ec78df3de55", 00:16:16.934 "is_configured": true, 00:16:16.934 "data_offset": 256, 00:16:16.934 "data_size": 7936 00:16:16.934 } 00:16:16.934 ] 00:16:16.934 }' 00:16:16.934 17:37:47 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:16.934 17:37:47 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:17.195 17:37:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:16:17.195 17:37:48 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:17.195 17:37:48 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:17.195 [2024-11-27 17:37:48.359226] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:16:17.195 [2024-11-27 17:37:48.359294] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:16:17.195 [2024-11-27 17:37:48.359393] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:17.195 [2024-11-27 17:37:48.359492] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:17.195 [2024-11-27 17:37:48.359586] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:16:17.195 17:37:48 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:17.195 17:37:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:17.195 17:37:48 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:17.195 17:37:48 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:17.195 17:37:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@720 -- # jq length 00:16:17.195 17:37:48 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:17.455 17:37:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:16:17.455 17:37:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:16:17.455 17:37:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:16:17.455 17:37:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:16:17.455 17:37:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:16:17.455 17:37:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:16:17.455 17:37:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@10 -- # local bdev_list 00:16:17.455 17:37:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:16:17.455 17:37:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@11 -- # local nbd_list 00:16:17.455 17:37:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@12 -- # local i 00:16:17.455 17:37:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:16:17.455 17:37:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:16:17.455 17:37:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:16:17.455 /dev/nbd0 00:16:17.455 17:37:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:16:17.714 17:37:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:16:17.714 17:37:48 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:16:17.714 17:37:48 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@869 -- # local i 00:16:17.714 17:37:48 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:16:17.714 17:37:48 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:16:17.714 17:37:48 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:16:17.715 17:37:48 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@873 -- # break 00:16:17.715 17:37:48 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:16:17.715 17:37:48 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:16:17.715 17:37:48 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:16:17.715 1+0 records in 00:16:17.715 1+0 records out 00:16:17.715 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000582349 s, 7.0 MB/s 00:16:17.715 17:37:48 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:16:17.715 17:37:48 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@886 -- # size=4096 00:16:17.715 17:37:48 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:16:17.715 17:37:48 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:16:17.715 17:37:48 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@889 -- # return 0 00:16:17.715 17:37:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:16:17.715 17:37:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:16:17.715 17:37:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:16:17.715 /dev/nbd1 00:16:17.715 17:37:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:16:17.974 17:37:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:16:17.974 17:37:48 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@868 -- # local nbd_name=nbd1 00:16:17.974 17:37:48 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@869 -- # local i 00:16:17.974 17:37:48 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:16:17.974 17:37:48 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:16:17.974 17:37:48 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@872 -- # grep -q -w nbd1 /proc/partitions 00:16:17.974 17:37:48 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@873 -- # break 00:16:17.974 17:37:48 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:16:17.974 17:37:48 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:16:17.974 17:37:48 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@885 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:16:17.974 1+0 records in 00:16:17.974 1+0 records out 00:16:17.974 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000254572 s, 16.1 MB/s 00:16:17.974 17:37:48 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:16:17.974 17:37:48 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@886 -- # size=4096 00:16:17.974 17:37:48 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:16:17.974 17:37:48 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:16:17.974 17:37:48 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@889 -- # return 0 00:16:17.974 17:37:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:16:17.974 17:37:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:16:17.974 17:37:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@738 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:16:17.974 17:37:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:16:17.974 17:37:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:16:17.974 17:37:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:16:17.974 17:37:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@50 -- # local nbd_list 00:16:17.974 17:37:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@51 -- # local i 00:16:17.974 17:37:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:16:17.974 17:37:48 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:16:18.233 17:37:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:16:18.233 17:37:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:16:18.233 17:37:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:16:18.233 17:37:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:16:18.233 17:37:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:16:18.233 17:37:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:16:18.233 17:37:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@41 -- # break 00:16:18.233 17:37:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@45 -- # return 0 00:16:18.233 17:37:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:16:18.233 17:37:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:16:18.233 17:37:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:16:18.233 17:37:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:16:18.233 17:37:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:16:18.233 17:37:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:16:18.233 17:37:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:16:18.233 17:37:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:16:18.233 17:37:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@41 -- # break 00:16:18.233 17:37:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/nbd_common.sh@45 -- # return 0 00:16:18.233 17:37:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:16:18.233 17:37:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:16:18.233 17:37:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:18.233 17:37:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:18.492 17:37:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:18.492 17:37:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:16:18.492 17:37:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:18.492 17:37:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:18.492 [2024-11-27 17:37:49.430099] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:16:18.492 [2024-11-27 17:37:49.430176] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:18.492 [2024-11-27 17:37:49.430197] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:16:18.492 [2024-11-27 17:37:49.430209] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:18.492 [2024-11-27 17:37:49.432231] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:18.492 [2024-11-27 17:37:49.432272] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:16:18.492 [2024-11-27 17:37:49.432339] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:16:18.492 [2024-11-27 17:37:49.432383] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:16:18.492 [2024-11-27 17:37:49.432497] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:16:18.492 spare 00:16:18.492 17:37:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:18.492 17:37:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:16:18.492 17:37:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:18.492 17:37:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:18.492 [2024-11-27 17:37:49.532380] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001580 00:16:18.492 [2024-11-27 17:37:49.532442] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:16:18.492 [2024-11-27 17:37:49.532694] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0001bb1b0 00:16:18.492 [2024-11-27 17:37:49.532858] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001580 00:16:18.492 [2024-11-27 17:37:49.532903] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001580 00:16:18.492 [2024-11-27 17:37:49.533071] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:18.492 17:37:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:18.492 17:37:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:16:18.492 17:37:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:18.492 17:37:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:18.492 17:37:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:18.492 17:37:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:18.492 17:37:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:18.492 17:37:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:18.492 17:37:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:18.492 17:37:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:18.492 17:37:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:18.492 17:37:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:18.492 17:37:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:18.492 17:37:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:18.492 17:37:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:18.492 17:37:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:18.492 17:37:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:18.492 "name": "raid_bdev1", 00:16:18.493 "uuid": "ca18a2f3-1538-4867-ae49-d203a82ee759", 00:16:18.493 "strip_size_kb": 0, 00:16:18.493 "state": "online", 00:16:18.493 "raid_level": "raid1", 00:16:18.493 "superblock": true, 00:16:18.493 "num_base_bdevs": 2, 00:16:18.493 "num_base_bdevs_discovered": 2, 00:16:18.493 "num_base_bdevs_operational": 2, 00:16:18.493 "base_bdevs_list": [ 00:16:18.493 { 00:16:18.493 "name": "spare", 00:16:18.493 "uuid": "470cfe62-9599-5986-90ee-67afdd4359fa", 00:16:18.493 "is_configured": true, 00:16:18.493 "data_offset": 256, 00:16:18.493 "data_size": 7936 00:16:18.493 }, 00:16:18.493 { 00:16:18.493 "name": "BaseBdev2", 00:16:18.493 "uuid": "8da903e4-ba16-5f8d-8a8f-8ec78df3de55", 00:16:18.493 "is_configured": true, 00:16:18.493 "data_offset": 256, 00:16:18.493 "data_size": 7936 00:16:18.493 } 00:16:18.493 ] 00:16:18.493 }' 00:16:18.493 17:37:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:18.493 17:37:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:19.062 17:37:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:16:19.062 17:37:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:19.062 17:37:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:16:19.062 17:37:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=none 00:16:19.062 17:37:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:19.062 17:37:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:19.062 17:37:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:19.062 17:37:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:19.062 17:37:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:19.062 17:37:49 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:19.062 17:37:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:19.062 "name": "raid_bdev1", 00:16:19.062 "uuid": "ca18a2f3-1538-4867-ae49-d203a82ee759", 00:16:19.062 "strip_size_kb": 0, 00:16:19.062 "state": "online", 00:16:19.062 "raid_level": "raid1", 00:16:19.062 "superblock": true, 00:16:19.062 "num_base_bdevs": 2, 00:16:19.062 "num_base_bdevs_discovered": 2, 00:16:19.062 "num_base_bdevs_operational": 2, 00:16:19.062 "base_bdevs_list": [ 00:16:19.062 { 00:16:19.062 "name": "spare", 00:16:19.062 "uuid": "470cfe62-9599-5986-90ee-67afdd4359fa", 00:16:19.062 "is_configured": true, 00:16:19.062 "data_offset": 256, 00:16:19.062 "data_size": 7936 00:16:19.062 }, 00:16:19.062 { 00:16:19.062 "name": "BaseBdev2", 00:16:19.062 "uuid": "8da903e4-ba16-5f8d-8a8f-8ec78df3de55", 00:16:19.062 "is_configured": true, 00:16:19.062 "data_offset": 256, 00:16:19.062 "data_size": 7936 00:16:19.062 } 00:16:19.062 ] 00:16:19.062 }' 00:16:19.062 17:37:49 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:19.062 17:37:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:16:19.062 17:37:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:19.062 17:37:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:16:19.062 17:37:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:16:19.062 17:37:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:19.062 17:37:50 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:19.062 17:37:50 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:19.062 17:37:50 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:19.062 17:37:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:16:19.062 17:37:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:16:19.062 17:37:50 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:19.062 17:37:50 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:19.062 [2024-11-27 17:37:50.132920] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:19.062 17:37:50 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:19.062 17:37:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:19.062 17:37:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:19.062 17:37:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:19.062 17:37:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:19.062 17:37:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:19.062 17:37:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:19.062 17:37:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:19.062 17:37:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:19.062 17:37:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:19.062 17:37:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:19.062 17:37:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:19.062 17:37:50 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:19.062 17:37:50 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:19.062 17:37:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:19.062 17:37:50 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:19.062 17:37:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:19.062 "name": "raid_bdev1", 00:16:19.062 "uuid": "ca18a2f3-1538-4867-ae49-d203a82ee759", 00:16:19.062 "strip_size_kb": 0, 00:16:19.062 "state": "online", 00:16:19.062 "raid_level": "raid1", 00:16:19.062 "superblock": true, 00:16:19.062 "num_base_bdevs": 2, 00:16:19.062 "num_base_bdevs_discovered": 1, 00:16:19.062 "num_base_bdevs_operational": 1, 00:16:19.062 "base_bdevs_list": [ 00:16:19.062 { 00:16:19.062 "name": null, 00:16:19.062 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:19.062 "is_configured": false, 00:16:19.062 "data_offset": 0, 00:16:19.062 "data_size": 7936 00:16:19.062 }, 00:16:19.062 { 00:16:19.062 "name": "BaseBdev2", 00:16:19.062 "uuid": "8da903e4-ba16-5f8d-8a8f-8ec78df3de55", 00:16:19.062 "is_configured": true, 00:16:19.062 "data_offset": 256, 00:16:19.062 "data_size": 7936 00:16:19.062 } 00:16:19.062 ] 00:16:19.062 }' 00:16:19.062 17:37:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:19.062 17:37:50 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:19.633 17:37:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:16:19.633 17:37:50 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:19.633 17:37:50 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:19.633 [2024-11-27 17:37:50.588215] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:16:19.633 [2024-11-27 17:37:50.588402] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:16:19.633 [2024-11-27 17:37:50.588418] bdev_raid.c:3748:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:16:19.633 [2024-11-27 17:37:50.588460] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:16:19.633 [2024-11-27 17:37:50.592348] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0001bb280 00:16:19.633 17:37:50 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:19.633 17:37:50 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@757 -- # sleep 1 00:16:19.633 [2024-11-27 17:37:50.594094] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:16:20.576 17:37:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:20.576 17:37:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:20.576 17:37:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:20.576 17:37:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:20.576 17:37:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:20.576 17:37:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:20.576 17:37:51 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:20.576 17:37:51 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:20.576 17:37:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:20.576 17:37:51 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:20.576 17:37:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:20.576 "name": "raid_bdev1", 00:16:20.576 "uuid": "ca18a2f3-1538-4867-ae49-d203a82ee759", 00:16:20.576 "strip_size_kb": 0, 00:16:20.576 "state": "online", 00:16:20.576 "raid_level": "raid1", 00:16:20.576 "superblock": true, 00:16:20.576 "num_base_bdevs": 2, 00:16:20.576 "num_base_bdevs_discovered": 2, 00:16:20.576 "num_base_bdevs_operational": 2, 00:16:20.576 "process": { 00:16:20.576 "type": "rebuild", 00:16:20.576 "target": "spare", 00:16:20.576 "progress": { 00:16:20.576 "blocks": 2560, 00:16:20.576 "percent": 32 00:16:20.576 } 00:16:20.576 }, 00:16:20.576 "base_bdevs_list": [ 00:16:20.576 { 00:16:20.576 "name": "spare", 00:16:20.576 "uuid": "470cfe62-9599-5986-90ee-67afdd4359fa", 00:16:20.576 "is_configured": true, 00:16:20.576 "data_offset": 256, 00:16:20.576 "data_size": 7936 00:16:20.576 }, 00:16:20.576 { 00:16:20.576 "name": "BaseBdev2", 00:16:20.576 "uuid": "8da903e4-ba16-5f8d-8a8f-8ec78df3de55", 00:16:20.576 "is_configured": true, 00:16:20.576 "data_offset": 256, 00:16:20.576 "data_size": 7936 00:16:20.576 } 00:16:20.576 ] 00:16:20.576 }' 00:16:20.576 17:37:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:20.576 17:37:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:20.576 17:37:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:20.576 17:37:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:20.576 17:37:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:16:20.576 17:37:51 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:20.576 17:37:51 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:20.576 [2024-11-27 17:37:51.759276] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:20.836 [2024-11-27 17:37:51.797958] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:16:20.836 [2024-11-27 17:37:51.798059] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:20.836 [2024-11-27 17:37:51.798126] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:20.836 [2024-11-27 17:37:51.798175] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:16:20.836 17:37:51 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:20.836 17:37:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:20.836 17:37:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:20.836 17:37:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:20.836 17:37:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:20.836 17:37:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:20.836 17:37:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:20.836 17:37:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:20.836 17:37:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:20.836 17:37:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:20.836 17:37:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:20.836 17:37:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:20.836 17:37:51 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:20.836 17:37:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:20.836 17:37:51 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:20.836 17:37:51 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:20.836 17:37:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:20.836 "name": "raid_bdev1", 00:16:20.836 "uuid": "ca18a2f3-1538-4867-ae49-d203a82ee759", 00:16:20.836 "strip_size_kb": 0, 00:16:20.836 "state": "online", 00:16:20.837 "raid_level": "raid1", 00:16:20.837 "superblock": true, 00:16:20.837 "num_base_bdevs": 2, 00:16:20.837 "num_base_bdevs_discovered": 1, 00:16:20.837 "num_base_bdevs_operational": 1, 00:16:20.837 "base_bdevs_list": [ 00:16:20.837 { 00:16:20.837 "name": null, 00:16:20.837 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:20.837 "is_configured": false, 00:16:20.837 "data_offset": 0, 00:16:20.837 "data_size": 7936 00:16:20.837 }, 00:16:20.837 { 00:16:20.837 "name": "BaseBdev2", 00:16:20.837 "uuid": "8da903e4-ba16-5f8d-8a8f-8ec78df3de55", 00:16:20.837 "is_configured": true, 00:16:20.837 "data_offset": 256, 00:16:20.837 "data_size": 7936 00:16:20.837 } 00:16:20.837 ] 00:16:20.837 }' 00:16:20.837 17:37:51 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:20.837 17:37:51 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:21.097 17:37:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:16:21.097 17:37:52 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:21.097 17:37:52 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:21.097 [2024-11-27 17:37:52.265308] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:16:21.097 [2024-11-27 17:37:52.265444] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:21.097 [2024-11-27 17:37:52.265495] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009f80 00:16:21.097 [2024-11-27 17:37:52.265526] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:21.097 [2024-11-27 17:37:52.265974] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:21.097 [2024-11-27 17:37:52.266032] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:16:21.097 [2024-11-27 17:37:52.266158] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:16:21.097 [2024-11-27 17:37:52.266200] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:16:21.097 [2024-11-27 17:37:52.266244] bdev_raid.c:3748:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:16:21.097 [2024-11-27 17:37:52.266313] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:16:21.097 spare 00:16:21.097 [2024-11-27 17:37:52.270240] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0001bb350 00:16:21.097 17:37:52 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:21.097 17:37:52 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@764 -- # sleep 1 00:16:21.097 [2024-11-27 17:37:52.272041] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:16:22.478 17:37:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:22.479 17:37:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:22.479 17:37:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:22.479 17:37:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:22.479 17:37:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:22.479 17:37:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:22.479 17:37:53 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:22.479 17:37:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:22.479 17:37:53 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:22.479 17:37:53 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:22.479 17:37:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:22.479 "name": "raid_bdev1", 00:16:22.479 "uuid": "ca18a2f3-1538-4867-ae49-d203a82ee759", 00:16:22.479 "strip_size_kb": 0, 00:16:22.479 "state": "online", 00:16:22.479 "raid_level": "raid1", 00:16:22.479 "superblock": true, 00:16:22.479 "num_base_bdevs": 2, 00:16:22.479 "num_base_bdevs_discovered": 2, 00:16:22.479 "num_base_bdevs_operational": 2, 00:16:22.479 "process": { 00:16:22.479 "type": "rebuild", 00:16:22.479 "target": "spare", 00:16:22.479 "progress": { 00:16:22.479 "blocks": 2560, 00:16:22.479 "percent": 32 00:16:22.479 } 00:16:22.479 }, 00:16:22.479 "base_bdevs_list": [ 00:16:22.479 { 00:16:22.479 "name": "spare", 00:16:22.479 "uuid": "470cfe62-9599-5986-90ee-67afdd4359fa", 00:16:22.479 "is_configured": true, 00:16:22.479 "data_offset": 256, 00:16:22.479 "data_size": 7936 00:16:22.479 }, 00:16:22.479 { 00:16:22.479 "name": "BaseBdev2", 00:16:22.479 "uuid": "8da903e4-ba16-5f8d-8a8f-8ec78df3de55", 00:16:22.479 "is_configured": true, 00:16:22.479 "data_offset": 256, 00:16:22.479 "data_size": 7936 00:16:22.479 } 00:16:22.479 ] 00:16:22.479 }' 00:16:22.479 17:37:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:22.479 17:37:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:22.479 17:37:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:22.479 17:37:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:22.479 17:37:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:16:22.479 17:37:53 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:22.479 17:37:53 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:22.479 [2024-11-27 17:37:53.408426] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:22.479 [2024-11-27 17:37:53.475941] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:16:22.479 [2024-11-27 17:37:53.476054] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:22.479 [2024-11-27 17:37:53.476090] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:22.479 [2024-11-27 17:37:53.476114] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:16:22.479 17:37:53 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:22.479 17:37:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:22.479 17:37:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:22.479 17:37:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:22.479 17:37:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:22.479 17:37:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:22.479 17:37:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:22.479 17:37:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:22.479 17:37:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:22.479 17:37:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:22.479 17:37:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:22.479 17:37:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:22.479 17:37:53 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:22.479 17:37:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:22.479 17:37:53 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:22.479 17:37:53 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:22.479 17:37:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:22.479 "name": "raid_bdev1", 00:16:22.479 "uuid": "ca18a2f3-1538-4867-ae49-d203a82ee759", 00:16:22.479 "strip_size_kb": 0, 00:16:22.479 "state": "online", 00:16:22.479 "raid_level": "raid1", 00:16:22.479 "superblock": true, 00:16:22.479 "num_base_bdevs": 2, 00:16:22.479 "num_base_bdevs_discovered": 1, 00:16:22.479 "num_base_bdevs_operational": 1, 00:16:22.479 "base_bdevs_list": [ 00:16:22.479 { 00:16:22.479 "name": null, 00:16:22.479 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:22.479 "is_configured": false, 00:16:22.479 "data_offset": 0, 00:16:22.479 "data_size": 7936 00:16:22.479 }, 00:16:22.479 { 00:16:22.479 "name": "BaseBdev2", 00:16:22.479 "uuid": "8da903e4-ba16-5f8d-8a8f-8ec78df3de55", 00:16:22.479 "is_configured": true, 00:16:22.479 "data_offset": 256, 00:16:22.479 "data_size": 7936 00:16:22.479 } 00:16:22.479 ] 00:16:22.479 }' 00:16:22.479 17:37:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:22.479 17:37:53 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:23.050 17:37:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:16:23.050 17:37:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:23.050 17:37:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:16:23.050 17:37:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=none 00:16:23.050 17:37:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:23.050 17:37:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:23.050 17:37:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:23.050 17:37:53 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:23.050 17:37:53 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:23.050 17:37:53 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:23.050 17:37:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:23.050 "name": "raid_bdev1", 00:16:23.050 "uuid": "ca18a2f3-1538-4867-ae49-d203a82ee759", 00:16:23.050 "strip_size_kb": 0, 00:16:23.050 "state": "online", 00:16:23.050 "raid_level": "raid1", 00:16:23.050 "superblock": true, 00:16:23.050 "num_base_bdevs": 2, 00:16:23.050 "num_base_bdevs_discovered": 1, 00:16:23.050 "num_base_bdevs_operational": 1, 00:16:23.050 "base_bdevs_list": [ 00:16:23.050 { 00:16:23.050 "name": null, 00:16:23.050 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:23.050 "is_configured": false, 00:16:23.050 "data_offset": 0, 00:16:23.050 "data_size": 7936 00:16:23.050 }, 00:16:23.050 { 00:16:23.050 "name": "BaseBdev2", 00:16:23.050 "uuid": "8da903e4-ba16-5f8d-8a8f-8ec78df3de55", 00:16:23.050 "is_configured": true, 00:16:23.050 "data_offset": 256, 00:16:23.050 "data_size": 7936 00:16:23.050 } 00:16:23.050 ] 00:16:23.050 }' 00:16:23.050 17:37:53 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:23.050 17:37:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:16:23.050 17:37:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:23.050 17:37:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:16:23.050 17:37:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:16:23.050 17:37:54 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:23.050 17:37:54 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:23.050 17:37:54 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:23.050 17:37:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:16:23.050 17:37:54 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:23.050 17:37:54 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:23.050 [2024-11-27 17:37:54.095133] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:16:23.050 [2024-11-27 17:37:54.095235] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:23.050 [2024-11-27 17:37:54.095272] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a580 00:16:23.050 [2024-11-27 17:37:54.095284] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:23.050 [2024-11-27 17:37:54.095672] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:23.050 [2024-11-27 17:37:54.095701] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:16:23.050 [2024-11-27 17:37:54.095767] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:16:23.050 [2024-11-27 17:37:54.095784] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:16:23.050 [2024-11-27 17:37:54.095791] bdev_raid.c:3709:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:16:23.050 [2024-11-27 17:37:54.095801] bdev_raid.c:3884:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:16:23.050 BaseBdev1 00:16:23.050 17:37:54 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:23.050 17:37:54 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@775 -- # sleep 1 00:16:23.990 17:37:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:23.990 17:37:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:23.990 17:37:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:23.990 17:37:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:23.990 17:37:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:23.990 17:37:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:23.990 17:37:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:23.990 17:37:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:23.990 17:37:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:23.990 17:37:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:23.990 17:37:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:23.990 17:37:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:23.990 17:37:55 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:23.990 17:37:55 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:23.990 17:37:55 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:23.990 17:37:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:23.990 "name": "raid_bdev1", 00:16:23.990 "uuid": "ca18a2f3-1538-4867-ae49-d203a82ee759", 00:16:23.990 "strip_size_kb": 0, 00:16:23.990 "state": "online", 00:16:23.990 "raid_level": "raid1", 00:16:23.990 "superblock": true, 00:16:23.990 "num_base_bdevs": 2, 00:16:23.990 "num_base_bdevs_discovered": 1, 00:16:23.990 "num_base_bdevs_operational": 1, 00:16:23.991 "base_bdevs_list": [ 00:16:23.991 { 00:16:23.991 "name": null, 00:16:23.991 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:23.991 "is_configured": false, 00:16:23.991 "data_offset": 0, 00:16:23.991 "data_size": 7936 00:16:23.991 }, 00:16:23.991 { 00:16:23.991 "name": "BaseBdev2", 00:16:23.991 "uuid": "8da903e4-ba16-5f8d-8a8f-8ec78df3de55", 00:16:23.991 "is_configured": true, 00:16:23.991 "data_offset": 256, 00:16:23.991 "data_size": 7936 00:16:23.991 } 00:16:23.991 ] 00:16:23.991 }' 00:16:23.991 17:37:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:23.991 17:37:55 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:24.561 17:37:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:16:24.561 17:37:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:24.561 17:37:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:16:24.561 17:37:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=none 00:16:24.561 17:37:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:24.561 17:37:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:24.561 17:37:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:24.561 17:37:55 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:24.561 17:37:55 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:24.561 17:37:55 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:24.561 17:37:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:24.561 "name": "raid_bdev1", 00:16:24.561 "uuid": "ca18a2f3-1538-4867-ae49-d203a82ee759", 00:16:24.561 "strip_size_kb": 0, 00:16:24.561 "state": "online", 00:16:24.561 "raid_level": "raid1", 00:16:24.561 "superblock": true, 00:16:24.561 "num_base_bdevs": 2, 00:16:24.561 "num_base_bdevs_discovered": 1, 00:16:24.561 "num_base_bdevs_operational": 1, 00:16:24.561 "base_bdevs_list": [ 00:16:24.561 { 00:16:24.561 "name": null, 00:16:24.561 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:24.561 "is_configured": false, 00:16:24.561 "data_offset": 0, 00:16:24.561 "data_size": 7936 00:16:24.561 }, 00:16:24.561 { 00:16:24.561 "name": "BaseBdev2", 00:16:24.561 "uuid": "8da903e4-ba16-5f8d-8a8f-8ec78df3de55", 00:16:24.561 "is_configured": true, 00:16:24.561 "data_offset": 256, 00:16:24.561 "data_size": 7936 00:16:24.561 } 00:16:24.561 ] 00:16:24.561 }' 00:16:24.561 17:37:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:24.562 17:37:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:16:24.562 17:37:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:24.562 17:37:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:16:24.562 17:37:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:16:24.562 17:37:55 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@650 -- # local es=0 00:16:24.562 17:37:55 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:16:24.562 17:37:55 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:16:24.562 17:37:55 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:16:24.562 17:37:55 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:16:24.562 17:37:55 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:16:24.562 17:37:55 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:16:24.562 17:37:55 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:24.562 17:37:55 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:24.562 [2024-11-27 17:37:55.704354] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:16:24.562 [2024-11-27 17:37:55.704523] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:16:24.562 [2024-11-27 17:37:55.704540] bdev_raid.c:3709:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:16:24.562 request: 00:16:24.562 { 00:16:24.562 "base_bdev": "BaseBdev1", 00:16:24.562 "raid_bdev": "raid_bdev1", 00:16:24.562 "method": "bdev_raid_add_base_bdev", 00:16:24.562 "req_id": 1 00:16:24.562 } 00:16:24.562 Got JSON-RPC error response 00:16:24.562 response: 00:16:24.562 { 00:16:24.562 "code": -22, 00:16:24.562 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:16:24.562 } 00:16:24.562 17:37:55 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:16:24.562 17:37:55 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@653 -- # es=1 00:16:24.562 17:37:55 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:16:24.562 17:37:55 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:16:24.562 17:37:55 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:16:24.562 17:37:55 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@779 -- # sleep 1 00:16:25.945 17:37:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:25.945 17:37:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:25.945 17:37:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:25.945 17:37:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:25.945 17:37:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:25.945 17:37:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:25.945 17:37:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:25.945 17:37:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:25.945 17:37:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:25.945 17:37:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:25.945 17:37:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:25.945 17:37:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:25.945 17:37:56 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:25.945 17:37:56 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:25.945 17:37:56 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:25.945 17:37:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:25.945 "name": "raid_bdev1", 00:16:25.945 "uuid": "ca18a2f3-1538-4867-ae49-d203a82ee759", 00:16:25.945 "strip_size_kb": 0, 00:16:25.945 "state": "online", 00:16:25.945 "raid_level": "raid1", 00:16:25.945 "superblock": true, 00:16:25.945 "num_base_bdevs": 2, 00:16:25.945 "num_base_bdevs_discovered": 1, 00:16:25.945 "num_base_bdevs_operational": 1, 00:16:25.945 "base_bdevs_list": [ 00:16:25.945 { 00:16:25.945 "name": null, 00:16:25.945 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:25.945 "is_configured": false, 00:16:25.945 "data_offset": 0, 00:16:25.945 "data_size": 7936 00:16:25.945 }, 00:16:25.945 { 00:16:25.945 "name": "BaseBdev2", 00:16:25.946 "uuid": "8da903e4-ba16-5f8d-8a8f-8ec78df3de55", 00:16:25.946 "is_configured": true, 00:16:25.946 "data_offset": 256, 00:16:25.946 "data_size": 7936 00:16:25.946 } 00:16:25.946 ] 00:16:25.946 }' 00:16:25.946 17:37:56 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:25.946 17:37:56 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:26.206 17:37:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:16:26.206 17:37:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:26.207 17:37:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:16:26.207 17:37:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@171 -- # local target=none 00:16:26.207 17:37:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:26.207 17:37:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:26.207 17:37:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:26.207 17:37:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:26.207 17:37:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:26.207 17:37:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:26.207 17:37:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:26.207 "name": "raid_bdev1", 00:16:26.207 "uuid": "ca18a2f3-1538-4867-ae49-d203a82ee759", 00:16:26.207 "strip_size_kb": 0, 00:16:26.207 "state": "online", 00:16:26.207 "raid_level": "raid1", 00:16:26.207 "superblock": true, 00:16:26.207 "num_base_bdevs": 2, 00:16:26.207 "num_base_bdevs_discovered": 1, 00:16:26.207 "num_base_bdevs_operational": 1, 00:16:26.207 "base_bdevs_list": [ 00:16:26.207 { 00:16:26.207 "name": null, 00:16:26.207 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:26.207 "is_configured": false, 00:16:26.207 "data_offset": 0, 00:16:26.207 "data_size": 7936 00:16:26.207 }, 00:16:26.207 { 00:16:26.207 "name": "BaseBdev2", 00:16:26.207 "uuid": "8da903e4-ba16-5f8d-8a8f-8ec78df3de55", 00:16:26.207 "is_configured": true, 00:16:26.207 "data_offset": 256, 00:16:26.207 "data_size": 7936 00:16:26.207 } 00:16:26.207 ] 00:16:26.207 }' 00:16:26.207 17:37:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:26.207 17:37:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:16:26.207 17:37:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:26.207 17:37:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:16:26.207 17:37:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@784 -- # killprocess 96717 00:16:26.207 17:37:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@950 -- # '[' -z 96717 ']' 00:16:26.207 17:37:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@954 -- # kill -0 96717 00:16:26.207 17:37:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@955 -- # uname 00:16:26.207 17:37:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:16:26.207 17:37:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 96717 00:16:26.207 17:37:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:16:26.207 killing process with pid 96717 00:16:26.207 Received shutdown signal, test time was about 60.000000 seconds 00:16:26.207 00:16:26.207 Latency(us) 00:16:26.207 [2024-11-27T17:37:57.399Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:16:26.207 [2024-11-27T17:37:57.399Z] =================================================================================================================== 00:16:26.207 [2024-11-27T17:37:57.399Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:16:26.207 17:37:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:16:26.207 17:37:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@968 -- # echo 'killing process with pid 96717' 00:16:26.207 17:37:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@969 -- # kill 96717 00:16:26.207 [2024-11-27 17:37:57.328836] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:16:26.207 [2024-11-27 17:37:57.328937] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:26.207 [2024-11-27 17:37:57.328981] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:26.207 [2024-11-27 17:37:57.328990] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state offline 00:16:26.207 17:37:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@974 -- # wait 96717 00:16:26.207 [2024-11-27 17:37:57.360370] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:16:26.468 17:37:57 bdev_raid.raid_rebuild_test_sb_4k -- bdev/bdev_raid.sh@786 -- # return 0 00:16:26.468 00:16:26.468 real 0m18.203s 00:16:26.468 user 0m24.177s 00:16:26.468 sys 0m2.557s 00:16:26.468 ************************************ 00:16:26.468 END TEST raid_rebuild_test_sb_4k 00:16:26.468 ************************************ 00:16:26.468 17:37:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@1126 -- # xtrace_disable 00:16:26.468 17:37:57 bdev_raid.raid_rebuild_test_sb_4k -- common/autotest_common.sh@10 -- # set +x 00:16:26.468 17:37:57 bdev_raid -- bdev/bdev_raid.sh@1003 -- # base_malloc_params='-m 32' 00:16:26.468 17:37:57 bdev_raid -- bdev/bdev_raid.sh@1004 -- # run_test raid_state_function_test_sb_md_separate raid_state_function_test raid1 2 true 00:16:26.468 17:37:57 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:16:26.468 17:37:57 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:16:26.468 17:37:57 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:16:26.728 ************************************ 00:16:26.728 START TEST raid_state_function_test_sb_md_separate 00:16:26.728 ************************************ 00:16:26.728 17:37:57 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@1125 -- # raid_state_function_test raid1 2 true 00:16:26.728 17:37:57 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:16:26.728 17:37:57 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:16:26.728 17:37:57 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:16:26.728 17:37:57 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:16:26.728 17:37:57 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:16:26.728 17:37:57 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:16:26.728 17:37:57 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:16:26.728 17:37:57 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:16:26.728 17:37:57 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:16:26.728 17:37:57 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:16:26.729 17:37:57 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:16:26.729 17:37:57 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:16:26.729 17:37:57 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:16:26.729 17:37:57 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:16:26.729 17:37:57 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:16:26.729 17:37:57 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@211 -- # local strip_size 00:16:26.729 17:37:57 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:16:26.729 17:37:57 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:16:26.729 17:37:57 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:16:26.729 17:37:57 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:16:26.729 17:37:57 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:16:26.729 17:37:57 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:16:26.729 17:37:57 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@229 -- # raid_pid=97402 00:16:26.729 Process raid pid: 97402 00:16:26.729 17:37:57 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:16:26.729 17:37:57 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 97402' 00:16:26.729 17:37:57 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@231 -- # waitforlisten 97402 00:16:26.729 17:37:57 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@831 -- # '[' -z 97402 ']' 00:16:26.729 17:37:57 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:16:26.729 17:37:57 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@836 -- # local max_retries=100 00:16:26.729 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:16:26.729 17:37:57 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:16:26.729 17:37:57 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@840 -- # xtrace_disable 00:16:26.729 17:37:57 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:26.729 [2024-11-27 17:37:57.772153] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:16:26.729 [2024-11-27 17:37:57.772294] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:16:26.988 [2024-11-27 17:37:57.919617] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:16:26.989 [2024-11-27 17:37:57.964590] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:16:26.989 [2024-11-27 17:37:58.008713] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:16:26.989 [2024-11-27 17:37:58.008756] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:16:27.559 17:37:58 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:16:27.559 17:37:58 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@864 -- # return 0 00:16:27.559 17:37:58 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:16:27.559 17:37:58 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:27.559 17:37:58 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:27.559 [2024-11-27 17:37:58.606982] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:16:27.559 [2024-11-27 17:37:58.607045] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:16:27.559 [2024-11-27 17:37:58.607059] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:16:27.559 [2024-11-27 17:37:58.607072] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:16:27.559 17:37:58 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:27.559 17:37:58 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:16:27.559 17:37:58 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:16:27.559 17:37:58 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:16:27.559 17:37:58 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:27.559 17:37:58 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:27.559 17:37:58 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:27.559 17:37:58 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:27.559 17:37:58 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:27.559 17:37:58 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:27.559 17:37:58 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:27.559 17:37:58 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:27.559 17:37:58 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:16:27.559 17:37:58 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:27.559 17:37:58 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:27.559 17:37:58 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:27.559 17:37:58 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:27.559 "name": "Existed_Raid", 00:16:27.559 "uuid": "e22f6b34-2857-408a-beb6-b8adf28768c7", 00:16:27.559 "strip_size_kb": 0, 00:16:27.559 "state": "configuring", 00:16:27.559 "raid_level": "raid1", 00:16:27.559 "superblock": true, 00:16:27.559 "num_base_bdevs": 2, 00:16:27.559 "num_base_bdevs_discovered": 0, 00:16:27.559 "num_base_bdevs_operational": 2, 00:16:27.559 "base_bdevs_list": [ 00:16:27.559 { 00:16:27.559 "name": "BaseBdev1", 00:16:27.559 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:27.559 "is_configured": false, 00:16:27.559 "data_offset": 0, 00:16:27.559 "data_size": 0 00:16:27.559 }, 00:16:27.559 { 00:16:27.559 "name": "BaseBdev2", 00:16:27.559 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:27.559 "is_configured": false, 00:16:27.559 "data_offset": 0, 00:16:27.559 "data_size": 0 00:16:27.559 } 00:16:27.559 ] 00:16:27.559 }' 00:16:27.559 17:37:58 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:27.559 17:37:58 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:28.129 17:37:59 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:16:28.130 17:37:59 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:28.130 17:37:59 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:28.130 [2024-11-27 17:37:59.082050] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:16:28.130 [2024-11-27 17:37:59.082091] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:16:28.130 17:37:59 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:28.130 17:37:59 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:16:28.130 17:37:59 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:28.130 17:37:59 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:28.130 [2024-11-27 17:37:59.094042] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:16:28.130 [2024-11-27 17:37:59.094132] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:16:28.130 [2024-11-27 17:37:59.094195] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:16:28.130 [2024-11-27 17:37:59.094225] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:16:28.130 17:37:59 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:28.130 17:37:59 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -b BaseBdev1 00:16:28.130 17:37:59 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:28.130 17:37:59 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:28.130 [2024-11-27 17:37:59.115738] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:16:28.130 BaseBdev1 00:16:28.130 17:37:59 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:28.130 17:37:59 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:16:28.130 17:37:59 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:16:28.130 17:37:59 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:16:28.130 17:37:59 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@901 -- # local i 00:16:28.130 17:37:59 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:16:28.130 17:37:59 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:16:28.130 17:37:59 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:16:28.130 17:37:59 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:28.130 17:37:59 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:28.130 17:37:59 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:28.130 17:37:59 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:16:28.130 17:37:59 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:28.130 17:37:59 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:28.130 [ 00:16:28.130 { 00:16:28.130 "name": "BaseBdev1", 00:16:28.130 "aliases": [ 00:16:28.130 "0fbf12f2-8544-4eb0-9efe-de25477935b7" 00:16:28.130 ], 00:16:28.130 "product_name": "Malloc disk", 00:16:28.130 "block_size": 4096, 00:16:28.130 "num_blocks": 8192, 00:16:28.130 "uuid": "0fbf12f2-8544-4eb0-9efe-de25477935b7", 00:16:28.130 "md_size": 32, 00:16:28.130 "md_interleave": false, 00:16:28.130 "dif_type": 0, 00:16:28.130 "assigned_rate_limits": { 00:16:28.130 "rw_ios_per_sec": 0, 00:16:28.130 "rw_mbytes_per_sec": 0, 00:16:28.130 "r_mbytes_per_sec": 0, 00:16:28.130 "w_mbytes_per_sec": 0 00:16:28.130 }, 00:16:28.130 "claimed": true, 00:16:28.130 "claim_type": "exclusive_write", 00:16:28.130 "zoned": false, 00:16:28.130 "supported_io_types": { 00:16:28.130 "read": true, 00:16:28.130 "write": true, 00:16:28.130 "unmap": true, 00:16:28.130 "flush": true, 00:16:28.130 "reset": true, 00:16:28.130 "nvme_admin": false, 00:16:28.130 "nvme_io": false, 00:16:28.130 "nvme_io_md": false, 00:16:28.130 "write_zeroes": true, 00:16:28.130 "zcopy": true, 00:16:28.130 "get_zone_info": false, 00:16:28.130 "zone_management": false, 00:16:28.130 "zone_append": false, 00:16:28.130 "compare": false, 00:16:28.130 "compare_and_write": false, 00:16:28.130 "abort": true, 00:16:28.130 "seek_hole": false, 00:16:28.130 "seek_data": false, 00:16:28.130 "copy": true, 00:16:28.130 "nvme_iov_md": false 00:16:28.130 }, 00:16:28.130 "memory_domains": [ 00:16:28.130 { 00:16:28.130 "dma_device_id": "system", 00:16:28.130 "dma_device_type": 1 00:16:28.130 }, 00:16:28.130 { 00:16:28.130 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:28.130 "dma_device_type": 2 00:16:28.130 } 00:16:28.130 ], 00:16:28.130 "driver_specific": {} 00:16:28.130 } 00:16:28.130 ] 00:16:28.130 17:37:59 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:28.130 17:37:59 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@907 -- # return 0 00:16:28.130 17:37:59 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:16:28.130 17:37:59 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:16:28.130 17:37:59 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:16:28.130 17:37:59 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:28.130 17:37:59 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:28.130 17:37:59 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:28.130 17:37:59 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:28.130 17:37:59 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:28.130 17:37:59 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:28.130 17:37:59 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:28.130 17:37:59 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:28.130 17:37:59 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:28.130 17:37:59 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:16:28.130 17:37:59 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:28.130 17:37:59 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:28.130 17:37:59 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:28.130 "name": "Existed_Raid", 00:16:28.130 "uuid": "99c6ebba-645b-495e-93c3-f67b69ab0f4e", 00:16:28.130 "strip_size_kb": 0, 00:16:28.130 "state": "configuring", 00:16:28.130 "raid_level": "raid1", 00:16:28.130 "superblock": true, 00:16:28.130 "num_base_bdevs": 2, 00:16:28.130 "num_base_bdevs_discovered": 1, 00:16:28.130 "num_base_bdevs_operational": 2, 00:16:28.130 "base_bdevs_list": [ 00:16:28.130 { 00:16:28.130 "name": "BaseBdev1", 00:16:28.130 "uuid": "0fbf12f2-8544-4eb0-9efe-de25477935b7", 00:16:28.130 "is_configured": true, 00:16:28.130 "data_offset": 256, 00:16:28.130 "data_size": 7936 00:16:28.130 }, 00:16:28.130 { 00:16:28.130 "name": "BaseBdev2", 00:16:28.130 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:28.130 "is_configured": false, 00:16:28.130 "data_offset": 0, 00:16:28.130 "data_size": 0 00:16:28.130 } 00:16:28.130 ] 00:16:28.130 }' 00:16:28.130 17:37:59 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:28.130 17:37:59 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:28.699 17:37:59 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:16:28.699 17:37:59 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:28.699 17:37:59 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:28.699 [2024-11-27 17:37:59.643204] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:16:28.699 [2024-11-27 17:37:59.643250] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:16:28.699 17:37:59 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:28.699 17:37:59 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:16:28.699 17:37:59 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:28.699 17:37:59 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:28.699 [2024-11-27 17:37:59.655274] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:16:28.699 [2024-11-27 17:37:59.657106] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:16:28.699 [2024-11-27 17:37:59.657175] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:16:28.699 17:37:59 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:28.699 17:37:59 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:16:28.699 17:37:59 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:16:28.699 17:37:59 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:16:28.699 17:37:59 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:16:28.699 17:37:59 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:16:28.699 17:37:59 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:28.699 17:37:59 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:28.699 17:37:59 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:28.699 17:37:59 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:28.699 17:37:59 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:28.699 17:37:59 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:28.699 17:37:59 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:28.699 17:37:59 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:28.699 17:37:59 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:28.699 17:37:59 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:28.700 17:37:59 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:16:28.700 17:37:59 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:28.700 17:37:59 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:28.700 "name": "Existed_Raid", 00:16:28.700 "uuid": "24337217-af66-4238-b7fa-cd5ca487e9fe", 00:16:28.700 "strip_size_kb": 0, 00:16:28.700 "state": "configuring", 00:16:28.700 "raid_level": "raid1", 00:16:28.700 "superblock": true, 00:16:28.700 "num_base_bdevs": 2, 00:16:28.700 "num_base_bdevs_discovered": 1, 00:16:28.700 "num_base_bdevs_operational": 2, 00:16:28.700 "base_bdevs_list": [ 00:16:28.700 { 00:16:28.700 "name": "BaseBdev1", 00:16:28.700 "uuid": "0fbf12f2-8544-4eb0-9efe-de25477935b7", 00:16:28.700 "is_configured": true, 00:16:28.700 "data_offset": 256, 00:16:28.700 "data_size": 7936 00:16:28.700 }, 00:16:28.700 { 00:16:28.700 "name": "BaseBdev2", 00:16:28.700 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:28.700 "is_configured": false, 00:16:28.700 "data_offset": 0, 00:16:28.700 "data_size": 0 00:16:28.700 } 00:16:28.700 ] 00:16:28.700 }' 00:16:28.700 17:37:59 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:28.700 17:37:59 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:28.960 17:38:00 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -b BaseBdev2 00:16:28.960 17:38:00 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:28.960 17:38:00 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:28.960 [2024-11-27 17:38:00.125654] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:16:28.960 [2024-11-27 17:38:00.126264] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:16:28.960 [2024-11-27 17:38:00.126334] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:16:28.960 BaseBdev2 00:16:28.960 [2024-11-27 17:38:00.126684] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:16:28.960 [2024-11-27 17:38:00.127075] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:16:28.960 [2024-11-27 17:38:00.127215] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:16:28.960 [2024-11-27 17:38:00.127486] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:28.960 17:38:00 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:28.960 17:38:00 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:16:28.960 17:38:00 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:16:28.960 17:38:00 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:16:28.960 17:38:00 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@901 -- # local i 00:16:28.960 17:38:00 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:16:28.960 17:38:00 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:16:28.960 17:38:00 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:16:28.960 17:38:00 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:28.960 17:38:00 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:28.960 17:38:00 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:28.960 17:38:00 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:16:28.960 17:38:00 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:28.960 17:38:00 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:29.220 [ 00:16:29.220 { 00:16:29.220 "name": "BaseBdev2", 00:16:29.220 "aliases": [ 00:16:29.220 "203a2ba1-633b-4eab-8c63-e5711af26b30" 00:16:29.220 ], 00:16:29.220 "product_name": "Malloc disk", 00:16:29.220 "block_size": 4096, 00:16:29.220 "num_blocks": 8192, 00:16:29.220 "uuid": "203a2ba1-633b-4eab-8c63-e5711af26b30", 00:16:29.220 "md_size": 32, 00:16:29.220 "md_interleave": false, 00:16:29.220 "dif_type": 0, 00:16:29.220 "assigned_rate_limits": { 00:16:29.220 "rw_ios_per_sec": 0, 00:16:29.220 "rw_mbytes_per_sec": 0, 00:16:29.220 "r_mbytes_per_sec": 0, 00:16:29.220 "w_mbytes_per_sec": 0 00:16:29.220 }, 00:16:29.220 "claimed": true, 00:16:29.220 "claim_type": "exclusive_write", 00:16:29.220 "zoned": false, 00:16:29.220 "supported_io_types": { 00:16:29.220 "read": true, 00:16:29.220 "write": true, 00:16:29.220 "unmap": true, 00:16:29.220 "flush": true, 00:16:29.220 "reset": true, 00:16:29.220 "nvme_admin": false, 00:16:29.220 "nvme_io": false, 00:16:29.220 "nvme_io_md": false, 00:16:29.220 "write_zeroes": true, 00:16:29.220 "zcopy": true, 00:16:29.220 "get_zone_info": false, 00:16:29.220 "zone_management": false, 00:16:29.220 "zone_append": false, 00:16:29.220 "compare": false, 00:16:29.220 "compare_and_write": false, 00:16:29.220 "abort": true, 00:16:29.220 "seek_hole": false, 00:16:29.220 "seek_data": false, 00:16:29.220 "copy": true, 00:16:29.220 "nvme_iov_md": false 00:16:29.220 }, 00:16:29.220 "memory_domains": [ 00:16:29.220 { 00:16:29.220 "dma_device_id": "system", 00:16:29.220 "dma_device_type": 1 00:16:29.220 }, 00:16:29.220 { 00:16:29.220 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:29.220 "dma_device_type": 2 00:16:29.220 } 00:16:29.220 ], 00:16:29.220 "driver_specific": {} 00:16:29.220 } 00:16:29.220 ] 00:16:29.220 17:38:00 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:29.220 17:38:00 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@907 -- # return 0 00:16:29.220 17:38:00 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:16:29.220 17:38:00 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:16:29.220 17:38:00 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 2 00:16:29.220 17:38:00 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:16:29.220 17:38:00 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:29.220 17:38:00 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:29.220 17:38:00 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:29.220 17:38:00 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:29.220 17:38:00 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:29.220 17:38:00 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:29.220 17:38:00 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:29.221 17:38:00 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:29.221 17:38:00 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:29.221 17:38:00 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:29.221 17:38:00 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:29.221 17:38:00 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:16:29.221 17:38:00 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:29.221 17:38:00 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:29.221 "name": "Existed_Raid", 00:16:29.221 "uuid": "24337217-af66-4238-b7fa-cd5ca487e9fe", 00:16:29.221 "strip_size_kb": 0, 00:16:29.221 "state": "online", 00:16:29.221 "raid_level": "raid1", 00:16:29.221 "superblock": true, 00:16:29.221 "num_base_bdevs": 2, 00:16:29.221 "num_base_bdevs_discovered": 2, 00:16:29.221 "num_base_bdevs_operational": 2, 00:16:29.221 "base_bdevs_list": [ 00:16:29.221 { 00:16:29.221 "name": "BaseBdev1", 00:16:29.221 "uuid": "0fbf12f2-8544-4eb0-9efe-de25477935b7", 00:16:29.221 "is_configured": true, 00:16:29.221 "data_offset": 256, 00:16:29.221 "data_size": 7936 00:16:29.221 }, 00:16:29.221 { 00:16:29.221 "name": "BaseBdev2", 00:16:29.221 "uuid": "203a2ba1-633b-4eab-8c63-e5711af26b30", 00:16:29.221 "is_configured": true, 00:16:29.221 "data_offset": 256, 00:16:29.221 "data_size": 7936 00:16:29.221 } 00:16:29.221 ] 00:16:29.221 }' 00:16:29.221 17:38:00 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:29.221 17:38:00 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:29.481 17:38:00 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:16:29.481 17:38:00 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:16:29.481 17:38:00 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:16:29.481 17:38:00 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:16:29.481 17:38:00 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@184 -- # local name 00:16:29.481 17:38:00 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:16:29.481 17:38:00 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:16:29.481 17:38:00 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:16:29.481 17:38:00 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:29.481 17:38:00 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:29.481 [2024-11-27 17:38:00.589112] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:29.481 17:38:00 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:29.481 17:38:00 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:16:29.481 "name": "Existed_Raid", 00:16:29.481 "aliases": [ 00:16:29.481 "24337217-af66-4238-b7fa-cd5ca487e9fe" 00:16:29.481 ], 00:16:29.481 "product_name": "Raid Volume", 00:16:29.481 "block_size": 4096, 00:16:29.481 "num_blocks": 7936, 00:16:29.481 "uuid": "24337217-af66-4238-b7fa-cd5ca487e9fe", 00:16:29.481 "md_size": 32, 00:16:29.481 "md_interleave": false, 00:16:29.481 "dif_type": 0, 00:16:29.481 "assigned_rate_limits": { 00:16:29.481 "rw_ios_per_sec": 0, 00:16:29.481 "rw_mbytes_per_sec": 0, 00:16:29.481 "r_mbytes_per_sec": 0, 00:16:29.481 "w_mbytes_per_sec": 0 00:16:29.481 }, 00:16:29.481 "claimed": false, 00:16:29.481 "zoned": false, 00:16:29.481 "supported_io_types": { 00:16:29.481 "read": true, 00:16:29.481 "write": true, 00:16:29.481 "unmap": false, 00:16:29.481 "flush": false, 00:16:29.481 "reset": true, 00:16:29.481 "nvme_admin": false, 00:16:29.481 "nvme_io": false, 00:16:29.481 "nvme_io_md": false, 00:16:29.481 "write_zeroes": true, 00:16:29.481 "zcopy": false, 00:16:29.481 "get_zone_info": false, 00:16:29.481 "zone_management": false, 00:16:29.481 "zone_append": false, 00:16:29.481 "compare": false, 00:16:29.481 "compare_and_write": false, 00:16:29.481 "abort": false, 00:16:29.481 "seek_hole": false, 00:16:29.481 "seek_data": false, 00:16:29.481 "copy": false, 00:16:29.481 "nvme_iov_md": false 00:16:29.481 }, 00:16:29.481 "memory_domains": [ 00:16:29.481 { 00:16:29.481 "dma_device_id": "system", 00:16:29.481 "dma_device_type": 1 00:16:29.481 }, 00:16:29.481 { 00:16:29.481 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:29.481 "dma_device_type": 2 00:16:29.481 }, 00:16:29.481 { 00:16:29.481 "dma_device_id": "system", 00:16:29.481 "dma_device_type": 1 00:16:29.481 }, 00:16:29.481 { 00:16:29.481 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:29.481 "dma_device_type": 2 00:16:29.481 } 00:16:29.481 ], 00:16:29.481 "driver_specific": { 00:16:29.481 "raid": { 00:16:29.481 "uuid": "24337217-af66-4238-b7fa-cd5ca487e9fe", 00:16:29.481 "strip_size_kb": 0, 00:16:29.481 "state": "online", 00:16:29.481 "raid_level": "raid1", 00:16:29.481 "superblock": true, 00:16:29.481 "num_base_bdevs": 2, 00:16:29.481 "num_base_bdevs_discovered": 2, 00:16:29.481 "num_base_bdevs_operational": 2, 00:16:29.481 "base_bdevs_list": [ 00:16:29.481 { 00:16:29.481 "name": "BaseBdev1", 00:16:29.481 "uuid": "0fbf12f2-8544-4eb0-9efe-de25477935b7", 00:16:29.481 "is_configured": true, 00:16:29.481 "data_offset": 256, 00:16:29.481 "data_size": 7936 00:16:29.481 }, 00:16:29.481 { 00:16:29.481 "name": "BaseBdev2", 00:16:29.481 "uuid": "203a2ba1-633b-4eab-8c63-e5711af26b30", 00:16:29.481 "is_configured": true, 00:16:29.481 "data_offset": 256, 00:16:29.481 "data_size": 7936 00:16:29.481 } 00:16:29.481 ] 00:16:29.481 } 00:16:29.481 } 00:16:29.481 }' 00:16:29.481 17:38:00 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:16:29.481 17:38:00 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:16:29.481 BaseBdev2' 00:16:29.482 17:38:00 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:29.742 17:38:00 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4096 32 false 0' 00:16:29.742 17:38:00 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:16:29.742 17:38:00 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:29.742 17:38:00 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:16:29.742 17:38:00 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:29.742 17:38:00 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:29.742 17:38:00 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:29.742 17:38:00 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 32 false 0' 00:16:29.742 17:38:00 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@193 -- # [[ 4096 32 false 0 == \4\0\9\6\ \3\2\ \f\a\l\s\e\ \0 ]] 00:16:29.742 17:38:00 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:16:29.742 17:38:00 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:16:29.742 17:38:00 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:29.742 17:38:00 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:29.742 17:38:00 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:29.742 17:38:00 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:29.742 17:38:00 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 32 false 0' 00:16:29.742 17:38:00 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@193 -- # [[ 4096 32 false 0 == \4\0\9\6\ \3\2\ \f\a\l\s\e\ \0 ]] 00:16:29.742 17:38:00 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:16:29.742 17:38:00 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:29.742 17:38:00 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:29.742 [2024-11-27 17:38:00.772594] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:16:29.742 17:38:00 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:29.742 17:38:00 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@260 -- # local expected_state 00:16:29.742 17:38:00 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:16:29.742 17:38:00 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@198 -- # case $1 in 00:16:29.742 17:38:00 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@199 -- # return 0 00:16:29.742 17:38:00 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:16:29.742 17:38:00 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 1 00:16:29.742 17:38:00 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:16:29.742 17:38:00 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:29.742 17:38:00 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:29.742 17:38:00 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:29.742 17:38:00 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:29.742 17:38:00 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:29.742 17:38:00 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:29.742 17:38:00 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:29.742 17:38:00 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:29.742 17:38:00 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:29.742 17:38:00 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:16:29.742 17:38:00 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:29.742 17:38:00 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:29.742 17:38:00 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:29.742 17:38:00 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:29.742 "name": "Existed_Raid", 00:16:29.742 "uuid": "24337217-af66-4238-b7fa-cd5ca487e9fe", 00:16:29.742 "strip_size_kb": 0, 00:16:29.742 "state": "online", 00:16:29.742 "raid_level": "raid1", 00:16:29.742 "superblock": true, 00:16:29.742 "num_base_bdevs": 2, 00:16:29.742 "num_base_bdevs_discovered": 1, 00:16:29.742 "num_base_bdevs_operational": 1, 00:16:29.742 "base_bdevs_list": [ 00:16:29.742 { 00:16:29.742 "name": null, 00:16:29.742 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:29.742 "is_configured": false, 00:16:29.742 "data_offset": 0, 00:16:29.742 "data_size": 7936 00:16:29.742 }, 00:16:29.742 { 00:16:29.742 "name": "BaseBdev2", 00:16:29.742 "uuid": "203a2ba1-633b-4eab-8c63-e5711af26b30", 00:16:29.742 "is_configured": true, 00:16:29.742 "data_offset": 256, 00:16:29.742 "data_size": 7936 00:16:29.742 } 00:16:29.742 ] 00:16:29.742 }' 00:16:29.742 17:38:00 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:29.742 17:38:00 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:30.313 17:38:01 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:16:30.313 17:38:01 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:16:30.313 17:38:01 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:30.313 17:38:01 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:16:30.313 17:38:01 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:30.313 17:38:01 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:30.313 17:38:01 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:30.313 17:38:01 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:16:30.313 17:38:01 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:16:30.313 17:38:01 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:16:30.313 17:38:01 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:30.313 17:38:01 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:30.313 [2024-11-27 17:38:01.324021] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:16:30.313 [2024-11-27 17:38:01.324133] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:16:30.313 [2024-11-27 17:38:01.336569] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:30.313 [2024-11-27 17:38:01.336638] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:30.313 [2024-11-27 17:38:01.336660] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:16:30.313 17:38:01 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:30.313 17:38:01 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:16:30.313 17:38:01 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:16:30.313 17:38:01 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:16:30.313 17:38:01 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:30.313 17:38:01 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:30.313 17:38:01 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:30.313 17:38:01 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:30.313 17:38:01 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:16:30.313 17:38:01 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:16:30.313 17:38:01 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:16:30.313 17:38:01 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@326 -- # killprocess 97402 00:16:30.313 17:38:01 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@950 -- # '[' -z 97402 ']' 00:16:30.313 17:38:01 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@954 -- # kill -0 97402 00:16:30.313 17:38:01 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@955 -- # uname 00:16:30.313 17:38:01 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:16:30.313 17:38:01 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 97402 00:16:30.313 killing process with pid 97402 00:16:30.313 17:38:01 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:16:30.313 17:38:01 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:16:30.313 17:38:01 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@968 -- # echo 'killing process with pid 97402' 00:16:30.313 17:38:01 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@969 -- # kill 97402 00:16:30.313 [2024-11-27 17:38:01.415615] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:16:30.313 17:38:01 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@974 -- # wait 97402 00:16:30.313 [2024-11-27 17:38:01.416615] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:16:30.573 17:38:01 bdev_raid.raid_state_function_test_sb_md_separate -- bdev/bdev_raid.sh@328 -- # return 0 00:16:30.573 00:16:30.573 real 0m3.998s 00:16:30.573 user 0m6.235s 00:16:30.573 sys 0m0.883s 00:16:30.573 17:38:01 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@1126 -- # xtrace_disable 00:16:30.573 17:38:01 bdev_raid.raid_state_function_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:30.573 ************************************ 00:16:30.573 END TEST raid_state_function_test_sb_md_separate 00:16:30.573 ************************************ 00:16:30.573 17:38:01 bdev_raid -- bdev/bdev_raid.sh@1005 -- # run_test raid_superblock_test_md_separate raid_superblock_test raid1 2 00:16:30.573 17:38:01 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:16:30.573 17:38:01 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:16:30.573 17:38:01 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:16:30.573 ************************************ 00:16:30.573 START TEST raid_superblock_test_md_separate 00:16:30.573 ************************************ 00:16:30.573 17:38:01 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@1125 -- # raid_superblock_test raid1 2 00:16:30.573 17:38:01 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@393 -- # local raid_level=raid1 00:16:30.573 17:38:01 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=2 00:16:30.573 17:38:01 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:16:30.573 17:38:01 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:16:30.573 17:38:01 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:16:30.573 17:38:01 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:16:30.573 17:38:01 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:16:30.573 17:38:01 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:16:30.573 17:38:01 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:16:30.573 17:38:01 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@399 -- # local strip_size 00:16:30.573 17:38:01 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:16:30.573 17:38:01 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:16:30.573 17:38:01 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:16:30.573 17:38:01 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@404 -- # '[' raid1 '!=' raid1 ']' 00:16:30.573 17:38:01 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@408 -- # strip_size=0 00:16:30.573 17:38:01 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@412 -- # raid_pid=97643 00:16:30.573 17:38:01 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:16:30.574 17:38:01 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@413 -- # waitforlisten 97643 00:16:30.574 17:38:01 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@831 -- # '[' -z 97643 ']' 00:16:30.574 17:38:01 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:16:30.574 17:38:01 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@836 -- # local max_retries=100 00:16:30.574 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:16:30.574 17:38:01 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:16:30.574 17:38:01 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@840 -- # xtrace_disable 00:16:30.574 17:38:01 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:30.833 [2024-11-27 17:38:01.831277] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:16:30.834 [2024-11-27 17:38:01.831420] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid97643 ] 00:16:30.834 [2024-11-27 17:38:01.977954] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:16:31.093 [2024-11-27 17:38:02.023795] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:16:31.093 [2024-11-27 17:38:02.067447] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:16:31.093 [2024-11-27 17:38:02.067493] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:16:31.664 17:38:02 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:16:31.664 17:38:02 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@864 -- # return 0 00:16:31.664 17:38:02 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:16:31.664 17:38:02 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:16:31.664 17:38:02 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:16:31.664 17:38:02 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:16:31.664 17:38:02 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:16:31.664 17:38:02 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:16:31.664 17:38:02 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:16:31.664 17:38:02 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:16:31.664 17:38:02 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -b malloc1 00:16:31.664 17:38:02 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:31.664 17:38:02 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:31.664 malloc1 00:16:31.664 17:38:02 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:31.664 17:38:02 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:16:31.664 17:38:02 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:31.664 17:38:02 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:31.664 [2024-11-27 17:38:02.671259] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:16:31.664 [2024-11-27 17:38:02.671317] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:31.664 [2024-11-27 17:38:02.671338] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:16:31.664 [2024-11-27 17:38:02.671351] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:31.664 [2024-11-27 17:38:02.673211] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:31.664 [2024-11-27 17:38:02.673261] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:16:31.664 pt1 00:16:31.664 17:38:02 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:31.664 17:38:02 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:16:31.664 17:38:02 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:16:31.664 17:38:02 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:16:31.664 17:38:02 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:16:31.664 17:38:02 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:16:31.664 17:38:02 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:16:31.664 17:38:02 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:16:31.664 17:38:02 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:16:31.664 17:38:02 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -b malloc2 00:16:31.664 17:38:02 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:31.664 17:38:02 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:31.664 malloc2 00:16:31.664 17:38:02 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:31.664 17:38:02 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:16:31.664 17:38:02 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:31.664 17:38:02 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:31.664 [2024-11-27 17:38:02.717167] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:16:31.664 [2024-11-27 17:38:02.717246] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:31.664 [2024-11-27 17:38:02.717273] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:16:31.664 [2024-11-27 17:38:02.717292] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:31.664 [2024-11-27 17:38:02.719861] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:31.664 [2024-11-27 17:38:02.719918] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:16:31.664 pt2 00:16:31.664 17:38:02 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:31.664 17:38:02 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:16:31.664 17:38:02 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:16:31.664 17:38:02 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''pt1 pt2'\''' -n raid_bdev1 -s 00:16:31.664 17:38:02 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:31.664 17:38:02 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:31.664 [2024-11-27 17:38:02.729135] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:16:31.664 [2024-11-27 17:38:02.730922] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:16:31.664 [2024-11-27 17:38:02.731104] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:16:31.664 [2024-11-27 17:38:02.731154] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:16:31.664 [2024-11-27 17:38:02.731242] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:16:31.664 [2024-11-27 17:38:02.731348] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:16:31.664 [2024-11-27 17:38:02.731367] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:16:31.664 [2024-11-27 17:38:02.731457] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:31.664 17:38:02 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:31.664 17:38:02 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:16:31.665 17:38:02 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:31.665 17:38:02 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:31.665 17:38:02 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:31.665 17:38:02 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:31.665 17:38:02 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:31.665 17:38:02 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:31.665 17:38:02 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:31.665 17:38:02 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:31.665 17:38:02 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:31.665 17:38:02 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:31.665 17:38:02 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:31.665 17:38:02 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:31.665 17:38:02 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:31.665 17:38:02 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:31.665 17:38:02 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:31.665 "name": "raid_bdev1", 00:16:31.665 "uuid": "d847b762-0e36-4d7f-9fd7-e7f3360b778d", 00:16:31.665 "strip_size_kb": 0, 00:16:31.665 "state": "online", 00:16:31.665 "raid_level": "raid1", 00:16:31.665 "superblock": true, 00:16:31.665 "num_base_bdevs": 2, 00:16:31.665 "num_base_bdevs_discovered": 2, 00:16:31.665 "num_base_bdevs_operational": 2, 00:16:31.665 "base_bdevs_list": [ 00:16:31.665 { 00:16:31.665 "name": "pt1", 00:16:31.665 "uuid": "00000000-0000-0000-0000-000000000001", 00:16:31.665 "is_configured": true, 00:16:31.665 "data_offset": 256, 00:16:31.665 "data_size": 7936 00:16:31.665 }, 00:16:31.665 { 00:16:31.665 "name": "pt2", 00:16:31.665 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:31.665 "is_configured": true, 00:16:31.665 "data_offset": 256, 00:16:31.665 "data_size": 7936 00:16:31.665 } 00:16:31.665 ] 00:16:31.665 }' 00:16:31.665 17:38:02 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:31.665 17:38:02 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:32.265 17:38:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:16:32.265 17:38:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:16:32.265 17:38:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:16:32.265 17:38:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:16:32.265 17:38:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@184 -- # local name 00:16:32.265 17:38:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:16:32.265 17:38:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:16:32.265 17:38:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:16:32.265 17:38:03 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:32.265 17:38:03 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:32.265 [2024-11-27 17:38:03.192556] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:32.265 17:38:03 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:32.265 17:38:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:16:32.265 "name": "raid_bdev1", 00:16:32.265 "aliases": [ 00:16:32.265 "d847b762-0e36-4d7f-9fd7-e7f3360b778d" 00:16:32.265 ], 00:16:32.265 "product_name": "Raid Volume", 00:16:32.265 "block_size": 4096, 00:16:32.265 "num_blocks": 7936, 00:16:32.265 "uuid": "d847b762-0e36-4d7f-9fd7-e7f3360b778d", 00:16:32.265 "md_size": 32, 00:16:32.265 "md_interleave": false, 00:16:32.265 "dif_type": 0, 00:16:32.265 "assigned_rate_limits": { 00:16:32.265 "rw_ios_per_sec": 0, 00:16:32.265 "rw_mbytes_per_sec": 0, 00:16:32.265 "r_mbytes_per_sec": 0, 00:16:32.265 "w_mbytes_per_sec": 0 00:16:32.265 }, 00:16:32.265 "claimed": false, 00:16:32.265 "zoned": false, 00:16:32.265 "supported_io_types": { 00:16:32.265 "read": true, 00:16:32.265 "write": true, 00:16:32.265 "unmap": false, 00:16:32.265 "flush": false, 00:16:32.265 "reset": true, 00:16:32.265 "nvme_admin": false, 00:16:32.265 "nvme_io": false, 00:16:32.265 "nvme_io_md": false, 00:16:32.265 "write_zeroes": true, 00:16:32.265 "zcopy": false, 00:16:32.265 "get_zone_info": false, 00:16:32.265 "zone_management": false, 00:16:32.265 "zone_append": false, 00:16:32.265 "compare": false, 00:16:32.265 "compare_and_write": false, 00:16:32.265 "abort": false, 00:16:32.265 "seek_hole": false, 00:16:32.265 "seek_data": false, 00:16:32.265 "copy": false, 00:16:32.265 "nvme_iov_md": false 00:16:32.265 }, 00:16:32.265 "memory_domains": [ 00:16:32.265 { 00:16:32.265 "dma_device_id": "system", 00:16:32.265 "dma_device_type": 1 00:16:32.265 }, 00:16:32.265 { 00:16:32.265 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:32.265 "dma_device_type": 2 00:16:32.265 }, 00:16:32.265 { 00:16:32.265 "dma_device_id": "system", 00:16:32.265 "dma_device_type": 1 00:16:32.265 }, 00:16:32.265 { 00:16:32.265 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:32.265 "dma_device_type": 2 00:16:32.265 } 00:16:32.265 ], 00:16:32.265 "driver_specific": { 00:16:32.265 "raid": { 00:16:32.265 "uuid": "d847b762-0e36-4d7f-9fd7-e7f3360b778d", 00:16:32.265 "strip_size_kb": 0, 00:16:32.265 "state": "online", 00:16:32.265 "raid_level": "raid1", 00:16:32.265 "superblock": true, 00:16:32.265 "num_base_bdevs": 2, 00:16:32.265 "num_base_bdevs_discovered": 2, 00:16:32.265 "num_base_bdevs_operational": 2, 00:16:32.265 "base_bdevs_list": [ 00:16:32.265 { 00:16:32.265 "name": "pt1", 00:16:32.265 "uuid": "00000000-0000-0000-0000-000000000001", 00:16:32.265 "is_configured": true, 00:16:32.265 "data_offset": 256, 00:16:32.265 "data_size": 7936 00:16:32.265 }, 00:16:32.265 { 00:16:32.265 "name": "pt2", 00:16:32.265 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:32.265 "is_configured": true, 00:16:32.265 "data_offset": 256, 00:16:32.265 "data_size": 7936 00:16:32.265 } 00:16:32.265 ] 00:16:32.265 } 00:16:32.265 } 00:16:32.265 }' 00:16:32.265 17:38:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:16:32.265 17:38:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:16:32.265 pt2' 00:16:32.265 17:38:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:32.265 17:38:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4096 32 false 0' 00:16:32.265 17:38:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:16:32.265 17:38:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:32.265 17:38:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:16:32.265 17:38:03 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:32.265 17:38:03 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:32.265 17:38:03 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:32.265 17:38:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 32 false 0' 00:16:32.265 17:38:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@193 -- # [[ 4096 32 false 0 == \4\0\9\6\ \3\2\ \f\a\l\s\e\ \0 ]] 00:16:32.265 17:38:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:16:32.265 17:38:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:16:32.265 17:38:03 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:32.265 17:38:03 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:32.265 17:38:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:32.265 17:38:03 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:32.265 17:38:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 32 false 0' 00:16:32.265 17:38:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@193 -- # [[ 4096 32 false 0 == \4\0\9\6\ \3\2\ \f\a\l\s\e\ \0 ]] 00:16:32.265 17:38:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:16:32.265 17:38:03 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:32.265 17:38:03 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:32.265 17:38:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:16:32.265 [2024-11-27 17:38:03.396173] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:32.265 17:38:03 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:32.265 17:38:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=d847b762-0e36-4d7f-9fd7-e7f3360b778d 00:16:32.265 17:38:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@436 -- # '[' -z d847b762-0e36-4d7f-9fd7-e7f3360b778d ']' 00:16:32.265 17:38:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:16:32.265 17:38:03 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:32.265 17:38:03 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:32.265 [2024-11-27 17:38:03.435866] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:16:32.265 [2024-11-27 17:38:03.435893] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:16:32.266 [2024-11-27 17:38:03.435973] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:32.266 [2024-11-27 17:38:03.436045] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:32.266 [2024-11-27 17:38:03.436060] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:16:32.266 17:38:03 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:32.266 17:38:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:32.266 17:38:03 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:32.266 17:38:03 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:32.266 17:38:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:16:32.266 17:38:03 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:32.525 17:38:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:16:32.525 17:38:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:16:32.525 17:38:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:16:32.525 17:38:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:16:32.525 17:38:03 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:32.525 17:38:03 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:32.525 17:38:03 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:32.525 17:38:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:16:32.525 17:38:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:16:32.525 17:38:03 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:32.525 17:38:03 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:32.525 17:38:03 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:32.525 17:38:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:16:32.525 17:38:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:16:32.525 17:38:03 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:32.525 17:38:03 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:32.525 17:38:03 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:32.525 17:38:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:16:32.525 17:38:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:16:32.525 17:38:03 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@650 -- # local es=0 00:16:32.525 17:38:03 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:16:32.525 17:38:03 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:16:32.525 17:38:03 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:16:32.525 17:38:03 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:16:32.525 17:38:03 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:16:32.525 17:38:03 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:16:32.525 17:38:03 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:32.525 17:38:03 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:32.525 [2024-11-27 17:38:03.559679] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:16:32.525 [2024-11-27 17:38:03.561522] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:16:32.525 [2024-11-27 17:38:03.561596] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:16:32.525 [2024-11-27 17:38:03.561641] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:16:32.525 [2024-11-27 17:38:03.561659] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:16:32.525 [2024-11-27 17:38:03.561680] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state configuring 00:16:32.525 request: 00:16:32.525 { 00:16:32.525 "name": "raid_bdev1", 00:16:32.526 "raid_level": "raid1", 00:16:32.526 "base_bdevs": [ 00:16:32.526 "malloc1", 00:16:32.526 "malloc2" 00:16:32.526 ], 00:16:32.526 "superblock": false, 00:16:32.526 "method": "bdev_raid_create", 00:16:32.526 "req_id": 1 00:16:32.526 } 00:16:32.526 Got JSON-RPC error response 00:16:32.526 response: 00:16:32.526 { 00:16:32.526 "code": -17, 00:16:32.526 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:16:32.526 } 00:16:32.526 17:38:03 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:16:32.526 17:38:03 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@653 -- # es=1 00:16:32.526 17:38:03 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:16:32.526 17:38:03 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:16:32.526 17:38:03 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:16:32.526 17:38:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:32.526 17:38:03 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:32.526 17:38:03 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:32.526 17:38:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:16:32.526 17:38:03 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:32.526 17:38:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:16:32.526 17:38:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:16:32.526 17:38:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:16:32.526 17:38:03 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:32.526 17:38:03 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:32.526 [2024-11-27 17:38:03.627521] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:16:32.526 [2024-11-27 17:38:03.627597] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:32.526 [2024-11-27 17:38:03.627620] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:16:32.526 [2024-11-27 17:38:03.627630] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:32.526 [2024-11-27 17:38:03.629430] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:32.526 [2024-11-27 17:38:03.629482] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:16:32.526 [2024-11-27 17:38:03.629550] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:16:32.526 [2024-11-27 17:38:03.629592] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:16:32.526 pt1 00:16:32.526 17:38:03 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:32.526 17:38:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 2 00:16:32.526 17:38:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:32.526 17:38:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:16:32.526 17:38:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:32.526 17:38:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:32.526 17:38:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:32.526 17:38:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:32.526 17:38:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:32.526 17:38:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:32.526 17:38:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:32.526 17:38:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:32.526 17:38:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:32.526 17:38:03 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:32.526 17:38:03 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:32.526 17:38:03 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:32.526 17:38:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:32.526 "name": "raid_bdev1", 00:16:32.526 "uuid": "d847b762-0e36-4d7f-9fd7-e7f3360b778d", 00:16:32.526 "strip_size_kb": 0, 00:16:32.526 "state": "configuring", 00:16:32.526 "raid_level": "raid1", 00:16:32.526 "superblock": true, 00:16:32.526 "num_base_bdevs": 2, 00:16:32.526 "num_base_bdevs_discovered": 1, 00:16:32.526 "num_base_bdevs_operational": 2, 00:16:32.526 "base_bdevs_list": [ 00:16:32.526 { 00:16:32.526 "name": "pt1", 00:16:32.526 "uuid": "00000000-0000-0000-0000-000000000001", 00:16:32.526 "is_configured": true, 00:16:32.526 "data_offset": 256, 00:16:32.526 "data_size": 7936 00:16:32.526 }, 00:16:32.526 { 00:16:32.526 "name": null, 00:16:32.526 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:32.526 "is_configured": false, 00:16:32.526 "data_offset": 256, 00:16:32.526 "data_size": 7936 00:16:32.526 } 00:16:32.526 ] 00:16:32.526 }' 00:16:32.526 17:38:03 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:32.526 17:38:03 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:33.096 17:38:04 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@470 -- # '[' 2 -gt 2 ']' 00:16:33.096 17:38:04 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:16:33.096 17:38:04 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:16:33.096 17:38:04 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:16:33.096 17:38:04 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:33.096 17:38:04 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:33.096 [2024-11-27 17:38:04.071246] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:16:33.096 [2024-11-27 17:38:04.071298] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:33.096 [2024-11-27 17:38:04.071321] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:16:33.096 [2024-11-27 17:38:04.071331] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:33.096 [2024-11-27 17:38:04.071497] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:33.096 [2024-11-27 17:38:04.071517] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:16:33.096 [2024-11-27 17:38:04.071562] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:16:33.096 [2024-11-27 17:38:04.071594] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:16:33.096 [2024-11-27 17:38:04.071682] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:16:33.096 [2024-11-27 17:38:04.071700] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:16:33.096 [2024-11-27 17:38:04.071780] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:16:33.096 [2024-11-27 17:38:04.071866] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:16:33.096 [2024-11-27 17:38:04.071884] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:16:33.096 [2024-11-27 17:38:04.071949] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:33.096 pt2 00:16:33.096 17:38:04 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:33.096 17:38:04 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:16:33.096 17:38:04 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:16:33.096 17:38:04 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:16:33.096 17:38:04 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:33.096 17:38:04 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:33.096 17:38:04 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:33.096 17:38:04 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:33.096 17:38:04 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:33.096 17:38:04 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:33.096 17:38:04 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:33.096 17:38:04 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:33.096 17:38:04 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:33.096 17:38:04 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:33.096 17:38:04 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:33.096 17:38:04 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:33.096 17:38:04 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:33.096 17:38:04 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:33.096 17:38:04 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:33.096 "name": "raid_bdev1", 00:16:33.096 "uuid": "d847b762-0e36-4d7f-9fd7-e7f3360b778d", 00:16:33.096 "strip_size_kb": 0, 00:16:33.096 "state": "online", 00:16:33.096 "raid_level": "raid1", 00:16:33.096 "superblock": true, 00:16:33.096 "num_base_bdevs": 2, 00:16:33.096 "num_base_bdevs_discovered": 2, 00:16:33.096 "num_base_bdevs_operational": 2, 00:16:33.096 "base_bdevs_list": [ 00:16:33.096 { 00:16:33.096 "name": "pt1", 00:16:33.096 "uuid": "00000000-0000-0000-0000-000000000001", 00:16:33.096 "is_configured": true, 00:16:33.096 "data_offset": 256, 00:16:33.096 "data_size": 7936 00:16:33.096 }, 00:16:33.096 { 00:16:33.096 "name": "pt2", 00:16:33.096 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:33.096 "is_configured": true, 00:16:33.096 "data_offset": 256, 00:16:33.096 "data_size": 7936 00:16:33.096 } 00:16:33.096 ] 00:16:33.096 }' 00:16:33.096 17:38:04 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:33.096 17:38:04 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:33.356 17:38:04 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:16:33.356 17:38:04 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:16:33.356 17:38:04 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:16:33.356 17:38:04 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:16:33.356 17:38:04 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@184 -- # local name 00:16:33.356 17:38:04 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:16:33.356 17:38:04 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:16:33.356 17:38:04 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:33.356 17:38:04 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:33.356 17:38:04 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:16:33.356 [2024-11-27 17:38:04.503527] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:33.356 17:38:04 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:33.356 17:38:04 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:16:33.356 "name": "raid_bdev1", 00:16:33.356 "aliases": [ 00:16:33.356 "d847b762-0e36-4d7f-9fd7-e7f3360b778d" 00:16:33.356 ], 00:16:33.356 "product_name": "Raid Volume", 00:16:33.356 "block_size": 4096, 00:16:33.356 "num_blocks": 7936, 00:16:33.356 "uuid": "d847b762-0e36-4d7f-9fd7-e7f3360b778d", 00:16:33.356 "md_size": 32, 00:16:33.356 "md_interleave": false, 00:16:33.356 "dif_type": 0, 00:16:33.356 "assigned_rate_limits": { 00:16:33.356 "rw_ios_per_sec": 0, 00:16:33.356 "rw_mbytes_per_sec": 0, 00:16:33.356 "r_mbytes_per_sec": 0, 00:16:33.356 "w_mbytes_per_sec": 0 00:16:33.356 }, 00:16:33.356 "claimed": false, 00:16:33.356 "zoned": false, 00:16:33.356 "supported_io_types": { 00:16:33.356 "read": true, 00:16:33.356 "write": true, 00:16:33.356 "unmap": false, 00:16:33.356 "flush": false, 00:16:33.356 "reset": true, 00:16:33.356 "nvme_admin": false, 00:16:33.356 "nvme_io": false, 00:16:33.356 "nvme_io_md": false, 00:16:33.356 "write_zeroes": true, 00:16:33.356 "zcopy": false, 00:16:33.356 "get_zone_info": false, 00:16:33.356 "zone_management": false, 00:16:33.356 "zone_append": false, 00:16:33.356 "compare": false, 00:16:33.356 "compare_and_write": false, 00:16:33.356 "abort": false, 00:16:33.356 "seek_hole": false, 00:16:33.356 "seek_data": false, 00:16:33.356 "copy": false, 00:16:33.356 "nvme_iov_md": false 00:16:33.356 }, 00:16:33.356 "memory_domains": [ 00:16:33.356 { 00:16:33.356 "dma_device_id": "system", 00:16:33.356 "dma_device_type": 1 00:16:33.356 }, 00:16:33.357 { 00:16:33.357 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:33.357 "dma_device_type": 2 00:16:33.357 }, 00:16:33.357 { 00:16:33.357 "dma_device_id": "system", 00:16:33.357 "dma_device_type": 1 00:16:33.357 }, 00:16:33.357 { 00:16:33.357 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:33.357 "dma_device_type": 2 00:16:33.357 } 00:16:33.357 ], 00:16:33.357 "driver_specific": { 00:16:33.357 "raid": { 00:16:33.357 "uuid": "d847b762-0e36-4d7f-9fd7-e7f3360b778d", 00:16:33.357 "strip_size_kb": 0, 00:16:33.357 "state": "online", 00:16:33.357 "raid_level": "raid1", 00:16:33.357 "superblock": true, 00:16:33.357 "num_base_bdevs": 2, 00:16:33.357 "num_base_bdevs_discovered": 2, 00:16:33.357 "num_base_bdevs_operational": 2, 00:16:33.357 "base_bdevs_list": [ 00:16:33.357 { 00:16:33.357 "name": "pt1", 00:16:33.357 "uuid": "00000000-0000-0000-0000-000000000001", 00:16:33.357 "is_configured": true, 00:16:33.357 "data_offset": 256, 00:16:33.357 "data_size": 7936 00:16:33.357 }, 00:16:33.357 { 00:16:33.357 "name": "pt2", 00:16:33.357 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:33.357 "is_configured": true, 00:16:33.357 "data_offset": 256, 00:16:33.357 "data_size": 7936 00:16:33.357 } 00:16:33.357 ] 00:16:33.357 } 00:16:33.357 } 00:16:33.357 }' 00:16:33.357 17:38:04 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:16:33.632 17:38:04 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:16:33.632 pt2' 00:16:33.632 17:38:04 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:33.632 17:38:04 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4096 32 false 0' 00:16:33.632 17:38:04 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:16:33.632 17:38:04 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:33.632 17:38:04 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:16:33.632 17:38:04 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:33.632 17:38:04 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:33.632 17:38:04 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:33.632 17:38:04 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 32 false 0' 00:16:33.632 17:38:04 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@193 -- # [[ 4096 32 false 0 == \4\0\9\6\ \3\2\ \f\a\l\s\e\ \0 ]] 00:16:33.632 17:38:04 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:16:33.632 17:38:04 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:16:33.632 17:38:04 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:33.632 17:38:04 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:33.632 17:38:04 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:33.632 17:38:04 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:33.632 17:38:04 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4096 32 false 0' 00:16:33.632 17:38:04 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@193 -- # [[ 4096 32 false 0 == \4\0\9\6\ \3\2\ \f\a\l\s\e\ \0 ]] 00:16:33.632 17:38:04 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:16:33.632 17:38:04 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:33.632 17:38:04 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:33.632 17:38:04 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:16:33.632 [2024-11-27 17:38:04.711470] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:33.632 17:38:04 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:33.632 17:38:04 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@487 -- # '[' d847b762-0e36-4d7f-9fd7-e7f3360b778d '!=' d847b762-0e36-4d7f-9fd7-e7f3360b778d ']' 00:16:33.632 17:38:04 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@491 -- # has_redundancy raid1 00:16:33.632 17:38:04 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@198 -- # case $1 in 00:16:33.632 17:38:04 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@199 -- # return 0 00:16:33.632 17:38:04 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@493 -- # rpc_cmd bdev_passthru_delete pt1 00:16:33.632 17:38:04 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:33.632 17:38:04 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:33.632 [2024-11-27 17:38:04.755289] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: pt1 00:16:33.632 17:38:04 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:33.632 17:38:04 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@496 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:33.632 17:38:04 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:33.632 17:38:04 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:33.632 17:38:04 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:33.632 17:38:04 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:33.632 17:38:04 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:33.632 17:38:04 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:33.632 17:38:04 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:33.632 17:38:04 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:33.632 17:38:04 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:33.632 17:38:04 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:33.632 17:38:04 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:33.632 17:38:04 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:33.632 17:38:04 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:33.632 17:38:04 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:33.632 17:38:04 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:33.632 "name": "raid_bdev1", 00:16:33.632 "uuid": "d847b762-0e36-4d7f-9fd7-e7f3360b778d", 00:16:33.632 "strip_size_kb": 0, 00:16:33.632 "state": "online", 00:16:33.632 "raid_level": "raid1", 00:16:33.632 "superblock": true, 00:16:33.632 "num_base_bdevs": 2, 00:16:33.632 "num_base_bdevs_discovered": 1, 00:16:33.632 "num_base_bdevs_operational": 1, 00:16:33.632 "base_bdevs_list": [ 00:16:33.632 { 00:16:33.632 "name": null, 00:16:33.632 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:33.632 "is_configured": false, 00:16:33.632 "data_offset": 0, 00:16:33.632 "data_size": 7936 00:16:33.632 }, 00:16:33.632 { 00:16:33.632 "name": "pt2", 00:16:33.632 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:33.632 "is_configured": true, 00:16:33.632 "data_offset": 256, 00:16:33.632 "data_size": 7936 00:16:33.632 } 00:16:33.632 ] 00:16:33.632 }' 00:16:33.632 17:38:04 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:33.632 17:38:04 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:34.241 17:38:05 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@499 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:16:34.241 17:38:05 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:34.241 17:38:05 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:34.241 [2024-11-27 17:38:05.199238] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:16:34.241 [2024-11-27 17:38:05.199270] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:16:34.241 [2024-11-27 17:38:05.199343] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:34.241 [2024-11-27 17:38:05.199389] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:34.241 [2024-11-27 17:38:05.199398] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:16:34.241 17:38:05 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:34.241 17:38:05 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@500 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:34.241 17:38:05 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@500 -- # jq -r '.[]' 00:16:34.241 17:38:05 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:34.241 17:38:05 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:34.241 17:38:05 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:34.241 17:38:05 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@500 -- # raid_bdev= 00:16:34.241 17:38:05 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@501 -- # '[' -n '' ']' 00:16:34.241 17:38:05 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@506 -- # (( i = 1 )) 00:16:34.241 17:38:05 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:16:34.241 17:38:05 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt2 00:16:34.241 17:38:05 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:34.241 17:38:05 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:34.241 17:38:05 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:34.241 17:38:05 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:16:34.241 17:38:05 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:16:34.241 17:38:05 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@511 -- # (( i = 1 )) 00:16:34.241 17:38:05 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:16:34.241 17:38:05 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@519 -- # i=1 00:16:34.241 17:38:05 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@520 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:16:34.241 17:38:05 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:34.241 17:38:05 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:34.241 [2024-11-27 17:38:05.271242] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:16:34.241 [2024-11-27 17:38:05.271302] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:34.241 [2024-11-27 17:38:05.271323] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008780 00:16:34.241 [2024-11-27 17:38:05.271334] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:34.241 [2024-11-27 17:38:05.273198] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:34.241 [2024-11-27 17:38:05.273234] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:16:34.241 [2024-11-27 17:38:05.273302] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:16:34.241 [2024-11-27 17:38:05.273348] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:16:34.241 [2024-11-27 17:38:05.273424] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:16:34.241 [2024-11-27 17:38:05.273432] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:16:34.241 [2024-11-27 17:38:05.273502] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:16:34.241 [2024-11-27 17:38:05.273577] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:16:34.241 [2024-11-27 17:38:05.273587] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001c80 00:16:34.241 [2024-11-27 17:38:05.273649] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:34.241 pt2 00:16:34.241 17:38:05 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:34.241 17:38:05 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@523 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:34.241 17:38:05 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:34.241 17:38:05 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:34.241 17:38:05 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:34.241 17:38:05 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:34.241 17:38:05 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:34.241 17:38:05 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:34.241 17:38:05 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:34.241 17:38:05 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:34.241 17:38:05 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:34.241 17:38:05 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:34.241 17:38:05 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:34.241 17:38:05 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:34.241 17:38:05 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:34.241 17:38:05 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:34.241 17:38:05 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:34.241 "name": "raid_bdev1", 00:16:34.241 "uuid": "d847b762-0e36-4d7f-9fd7-e7f3360b778d", 00:16:34.241 "strip_size_kb": 0, 00:16:34.241 "state": "online", 00:16:34.241 "raid_level": "raid1", 00:16:34.241 "superblock": true, 00:16:34.241 "num_base_bdevs": 2, 00:16:34.241 "num_base_bdevs_discovered": 1, 00:16:34.241 "num_base_bdevs_operational": 1, 00:16:34.241 "base_bdevs_list": [ 00:16:34.241 { 00:16:34.241 "name": null, 00:16:34.241 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:34.241 "is_configured": false, 00:16:34.241 "data_offset": 256, 00:16:34.241 "data_size": 7936 00:16:34.241 }, 00:16:34.241 { 00:16:34.241 "name": "pt2", 00:16:34.241 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:34.241 "is_configured": true, 00:16:34.241 "data_offset": 256, 00:16:34.241 "data_size": 7936 00:16:34.241 } 00:16:34.241 ] 00:16:34.241 }' 00:16:34.241 17:38:05 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:34.241 17:38:05 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:34.510 17:38:05 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@526 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:16:34.510 17:38:05 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:34.510 17:38:05 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:34.510 [2024-11-27 17:38:05.671239] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:16:34.510 [2024-11-27 17:38:05.671264] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:16:34.510 [2024-11-27 17:38:05.671336] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:34.510 [2024-11-27 17:38:05.671397] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:34.510 [2024-11-27 17:38:05.671411] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name raid_bdev1, state offline 00:16:34.510 17:38:05 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:34.510 17:38:05 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@527 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:34.510 17:38:05 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@527 -- # jq -r '.[]' 00:16:34.510 17:38:05 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:34.510 17:38:05 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:34.510 17:38:05 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:34.769 17:38:05 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@527 -- # raid_bdev= 00:16:34.769 17:38:05 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@528 -- # '[' -n '' ']' 00:16:34.769 17:38:05 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@532 -- # '[' 2 -gt 2 ']' 00:16:34.769 17:38:05 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@540 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:16:34.769 17:38:05 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:34.769 17:38:05 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:34.769 [2024-11-27 17:38:05.731255] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:16:34.769 [2024-11-27 17:38:05.731316] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:34.769 [2024-11-27 17:38:05.731337] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008d80 00:16:34.769 [2024-11-27 17:38:05.731352] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:34.769 [2024-11-27 17:38:05.733220] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:34.769 [2024-11-27 17:38:05.733262] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:16:34.769 [2024-11-27 17:38:05.733313] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:16:34.769 [2024-11-27 17:38:05.733355] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:16:34.769 [2024-11-27 17:38:05.733488] bdev_raid.c:3675:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev pt2 (4) greater than existing raid bdev raid_bdev1 (2) 00:16:34.769 [2024-11-27 17:38:05.733510] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:16:34.769 [2024-11-27 17:38:05.733533] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002000 name raid_bdev1, state configuring 00:16:34.769 [2024-11-27 17:38:05.733579] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:16:34.769 [2024-11-27 17:38:05.733647] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000002380 00:16:34.770 [2024-11-27 17:38:05.733663] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:16:34.770 [2024-11-27 17:38:05.733724] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:16:34.770 [2024-11-27 17:38:05.733814] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000002380 00:16:34.770 [2024-11-27 17:38:05.733831] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000002380 00:16:34.770 [2024-11-27 17:38:05.733913] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:34.770 pt1 00:16:34.770 17:38:05 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:34.770 17:38:05 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@542 -- # '[' 2 -gt 2 ']' 00:16:34.770 17:38:05 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@554 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:34.770 17:38:05 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:34.770 17:38:05 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:34.770 17:38:05 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:34.770 17:38:05 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:34.770 17:38:05 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:34.770 17:38:05 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:34.770 17:38:05 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:34.770 17:38:05 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:34.770 17:38:05 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:34.770 17:38:05 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:34.770 17:38:05 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:34.770 17:38:05 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:34.770 17:38:05 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:34.770 17:38:05 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:34.770 17:38:05 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:34.770 "name": "raid_bdev1", 00:16:34.770 "uuid": "d847b762-0e36-4d7f-9fd7-e7f3360b778d", 00:16:34.770 "strip_size_kb": 0, 00:16:34.770 "state": "online", 00:16:34.770 "raid_level": "raid1", 00:16:34.770 "superblock": true, 00:16:34.770 "num_base_bdevs": 2, 00:16:34.770 "num_base_bdevs_discovered": 1, 00:16:34.770 "num_base_bdevs_operational": 1, 00:16:34.770 "base_bdevs_list": [ 00:16:34.770 { 00:16:34.770 "name": null, 00:16:34.770 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:34.770 "is_configured": false, 00:16:34.770 "data_offset": 256, 00:16:34.770 "data_size": 7936 00:16:34.770 }, 00:16:34.770 { 00:16:34.770 "name": "pt2", 00:16:34.770 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:34.770 "is_configured": true, 00:16:34.770 "data_offset": 256, 00:16:34.770 "data_size": 7936 00:16:34.770 } 00:16:34.770 ] 00:16:34.770 }' 00:16:34.770 17:38:05 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:34.770 17:38:05 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:35.029 17:38:06 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@555 -- # rpc_cmd bdev_raid_get_bdevs online 00:16:35.029 17:38:06 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:35.029 17:38:06 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@555 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:16:35.029 17:38:06 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:35.029 17:38:06 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:35.029 17:38:06 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@555 -- # [[ false == \f\a\l\s\e ]] 00:16:35.029 17:38:06 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@558 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:16:35.029 17:38:06 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:35.029 17:38:06 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:35.029 17:38:06 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@558 -- # jq -r '.[] | .uuid' 00:16:35.029 [2024-11-27 17:38:06.207466] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:35.029 17:38:06 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:35.288 17:38:06 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@558 -- # '[' d847b762-0e36-4d7f-9fd7-e7f3360b778d '!=' d847b762-0e36-4d7f-9fd7-e7f3360b778d ']' 00:16:35.288 17:38:06 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@563 -- # killprocess 97643 00:16:35.288 17:38:06 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@950 -- # '[' -z 97643 ']' 00:16:35.288 17:38:06 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@954 -- # kill -0 97643 00:16:35.288 17:38:06 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@955 -- # uname 00:16:35.288 17:38:06 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:16:35.288 17:38:06 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 97643 00:16:35.288 17:38:06 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:16:35.288 17:38:06 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:16:35.288 killing process with pid 97643 00:16:35.288 17:38:06 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@968 -- # echo 'killing process with pid 97643' 00:16:35.288 17:38:06 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@969 -- # kill 97643 00:16:35.288 [2024-11-27 17:38:06.287413] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:16:35.288 [2024-11-27 17:38:06.287484] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:35.288 [2024-11-27 17:38:06.287529] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:35.288 [2024-11-27 17:38:06.287538] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002380 name raid_bdev1, state offline 00:16:35.288 17:38:06 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@974 -- # wait 97643 00:16:35.288 [2024-11-27 17:38:06.311576] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:16:35.548 17:38:06 bdev_raid.raid_superblock_test_md_separate -- bdev/bdev_raid.sh@565 -- # return 0 00:16:35.548 00:16:35.548 real 0m4.809s 00:16:35.548 user 0m7.786s 00:16:35.548 sys 0m1.019s 00:16:35.548 17:38:06 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@1126 -- # xtrace_disable 00:16:35.548 17:38:06 bdev_raid.raid_superblock_test_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:35.548 ************************************ 00:16:35.548 END TEST raid_superblock_test_md_separate 00:16:35.549 ************************************ 00:16:35.549 17:38:06 bdev_raid -- bdev/bdev_raid.sh@1006 -- # '[' true = true ']' 00:16:35.549 17:38:06 bdev_raid -- bdev/bdev_raid.sh@1007 -- # run_test raid_rebuild_test_sb_md_separate raid_rebuild_test raid1 2 true false true 00:16:35.549 17:38:06 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 7 -le 1 ']' 00:16:35.549 17:38:06 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:16:35.549 17:38:06 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:16:35.549 ************************************ 00:16:35.549 START TEST raid_rebuild_test_sb_md_separate 00:16:35.549 ************************************ 00:16:35.549 17:38:06 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@1125 -- # raid_rebuild_test raid1 2 true false true 00:16:35.549 17:38:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:16:35.549 17:38:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=2 00:16:35.549 17:38:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:16:35.549 17:38:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:16:35.549 17:38:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@573 -- # local verify=true 00:16:35.549 17:38:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:16:35.549 17:38:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:16:35.549 17:38:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:16:35.549 17:38:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:16:35.549 17:38:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:16:35.549 17:38:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:16:35.549 17:38:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:16:35.549 17:38:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:16:35.549 17:38:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:16:35.549 17:38:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:16:35.549 17:38:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:16:35.549 17:38:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@576 -- # local strip_size 00:16:35.549 17:38:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@577 -- # local create_arg 00:16:35.549 17:38:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:16:35.549 17:38:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@579 -- # local data_offset 00:16:35.549 17:38:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:16:35.549 17:38:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:16:35.549 17:38:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:16:35.549 17:38:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:16:35.549 17:38:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@597 -- # raid_pid=97960 00:16:35.549 17:38:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:16:35.549 17:38:06 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@598 -- # waitforlisten 97960 00:16:35.549 17:38:06 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@831 -- # '[' -z 97960 ']' 00:16:35.549 17:38:06 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:16:35.549 17:38:06 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@836 -- # local max_retries=100 00:16:35.549 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:16:35.549 17:38:06 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:16:35.549 17:38:06 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@840 -- # xtrace_disable 00:16:35.549 17:38:06 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:35.549 I/O size of 3145728 is greater than zero copy threshold (65536). 00:16:35.549 Zero copy mechanism will not be used. 00:16:35.549 [2024-11-27 17:38:06.731168] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:16:35.549 [2024-11-27 17:38:06.731314] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid97960 ] 00:16:35.809 [2024-11-27 17:38:06.857744] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:16:35.809 [2024-11-27 17:38:06.901064] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:16:35.809 [2024-11-27 17:38:06.943790] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:16:35.809 [2024-11-27 17:38:06.943838] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:16:36.377 17:38:07 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:16:36.377 17:38:07 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@864 -- # return 0 00:16:36.377 17:38:07 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:16:36.377 17:38:07 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -b BaseBdev1_malloc 00:16:36.377 17:38:07 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:36.377 17:38:07 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:36.377 BaseBdev1_malloc 00:16:36.377 17:38:07 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:36.377 17:38:07 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:16:36.377 17:38:07 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:36.377 17:38:07 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:36.377 [2024-11-27 17:38:07.566740] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:16:36.377 [2024-11-27 17:38:07.566840] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:36.377 [2024-11-27 17:38:07.566871] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:16:36.377 [2024-11-27 17:38:07.566895] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:36.636 [2024-11-27 17:38:07.568753] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:36.636 [2024-11-27 17:38:07.568796] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:16:36.636 BaseBdev1 00:16:36.636 17:38:07 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:36.636 17:38:07 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:16:36.636 17:38:07 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -b BaseBdev2_malloc 00:16:36.636 17:38:07 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:36.636 17:38:07 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:36.636 BaseBdev2_malloc 00:16:36.636 17:38:07 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:36.636 17:38:07 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:16:36.636 17:38:07 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:36.636 17:38:07 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:36.636 [2024-11-27 17:38:07.612509] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:16:36.636 [2024-11-27 17:38:07.612623] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:36.636 [2024-11-27 17:38:07.612680] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:16:36.636 [2024-11-27 17:38:07.612709] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:36.636 [2024-11-27 17:38:07.616178] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:36.636 [2024-11-27 17:38:07.616233] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:16:36.636 BaseBdev2 00:16:36.636 17:38:07 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:36.636 17:38:07 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -b spare_malloc 00:16:36.636 17:38:07 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:36.636 17:38:07 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:36.636 spare_malloc 00:16:36.636 17:38:07 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:36.636 17:38:07 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:16:36.636 17:38:07 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:36.636 17:38:07 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:36.636 spare_delay 00:16:36.636 17:38:07 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:36.636 17:38:07 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:16:36.636 17:38:07 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:36.636 17:38:07 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:36.636 [2024-11-27 17:38:07.654586] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:16:36.636 [2024-11-27 17:38:07.654654] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:36.636 [2024-11-27 17:38:07.654677] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:16:36.636 [2024-11-27 17:38:07.654690] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:36.636 [2024-11-27 17:38:07.656549] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:36.636 [2024-11-27 17:38:07.656590] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:16:36.636 spare 00:16:36.636 17:38:07 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:36.636 17:38:07 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 00:16:36.636 17:38:07 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:36.636 17:38:07 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:36.636 [2024-11-27 17:38:07.666628] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:16:36.636 [2024-11-27 17:38:07.668411] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:16:36.636 [2024-11-27 17:38:07.668576] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:16:36.636 [2024-11-27 17:38:07.668589] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:16:36.636 [2024-11-27 17:38:07.668674] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:16:36.636 [2024-11-27 17:38:07.668781] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:16:36.636 [2024-11-27 17:38:07.668806] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:16:36.636 [2024-11-27 17:38:07.668893] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:36.636 17:38:07 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:36.636 17:38:07 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:16:36.636 17:38:07 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:36.636 17:38:07 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:36.636 17:38:07 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:36.636 17:38:07 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:36.636 17:38:07 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:36.636 17:38:07 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:36.636 17:38:07 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:36.636 17:38:07 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:36.636 17:38:07 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:36.636 17:38:07 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:36.636 17:38:07 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:36.636 17:38:07 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:36.636 17:38:07 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:36.636 17:38:07 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:36.636 17:38:07 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:36.636 "name": "raid_bdev1", 00:16:36.636 "uuid": "73f0ad94-d8a4-4ed9-97dc-c66b90d469e0", 00:16:36.636 "strip_size_kb": 0, 00:16:36.636 "state": "online", 00:16:36.636 "raid_level": "raid1", 00:16:36.636 "superblock": true, 00:16:36.636 "num_base_bdevs": 2, 00:16:36.636 "num_base_bdevs_discovered": 2, 00:16:36.636 "num_base_bdevs_operational": 2, 00:16:36.636 "base_bdevs_list": [ 00:16:36.636 { 00:16:36.636 "name": "BaseBdev1", 00:16:36.636 "uuid": "c4359f5b-ae82-5bb3-ac72-325ea1269387", 00:16:36.636 "is_configured": true, 00:16:36.636 "data_offset": 256, 00:16:36.636 "data_size": 7936 00:16:36.636 }, 00:16:36.636 { 00:16:36.636 "name": "BaseBdev2", 00:16:36.636 "uuid": "3f16235d-acd4-5104-a066-8085bfe43bf7", 00:16:36.636 "is_configured": true, 00:16:36.636 "data_offset": 256, 00:16:36.636 "data_size": 7936 00:16:36.636 } 00:16:36.636 ] 00:16:36.636 }' 00:16:36.636 17:38:07 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:36.636 17:38:07 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:37.204 17:38:08 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:16:37.204 17:38:08 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:16:37.204 17:38:08 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:37.204 17:38:08 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:37.204 [2024-11-27 17:38:08.102086] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:37.204 17:38:08 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:37.204 17:38:08 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=7936 00:16:37.204 17:38:08 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:16:37.204 17:38:08 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:37.204 17:38:08 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:37.205 17:38:08 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:37.205 17:38:08 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:37.205 17:38:08 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@619 -- # data_offset=256 00:16:37.205 17:38:08 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:16:37.205 17:38:08 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@624 -- # '[' true = true ']' 00:16:37.205 17:38:08 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@625 -- # local write_unit_size 00:16:37.205 17:38:08 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@628 -- # nbd_start_disks /var/tmp/spdk.sock raid_bdev1 /dev/nbd0 00:16:37.205 17:38:08 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:16:37.205 17:38:08 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@10 -- # bdev_list=('raid_bdev1') 00:16:37.205 17:38:08 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@10 -- # local bdev_list 00:16:37.205 17:38:08 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:16:37.205 17:38:08 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@11 -- # local nbd_list 00:16:37.205 17:38:08 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@12 -- # local i 00:16:37.205 17:38:08 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:16:37.205 17:38:08 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:16:37.205 17:38:08 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk raid_bdev1 /dev/nbd0 00:16:37.205 [2024-11-27 17:38:08.349487] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:16:37.205 /dev/nbd0 00:16:37.205 17:38:08 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:16:37.205 17:38:08 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:16:37.205 17:38:08 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:16:37.205 17:38:08 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@869 -- # local i 00:16:37.205 17:38:08 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:16:37.205 17:38:08 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:16:37.205 17:38:08 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:16:37.205 17:38:08 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@873 -- # break 00:16:37.205 17:38:08 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:16:37.205 17:38:08 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:16:37.205 17:38:08 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:16:37.464 1+0 records in 00:16:37.464 1+0 records out 00:16:37.464 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000390856 s, 10.5 MB/s 00:16:37.464 17:38:08 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:16:37.464 17:38:08 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@886 -- # size=4096 00:16:37.464 17:38:08 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:16:37.464 17:38:08 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:16:37.464 17:38:08 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@889 -- # return 0 00:16:37.464 17:38:08 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:16:37.464 17:38:08 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:16:37.464 17:38:08 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@629 -- # '[' raid1 = raid5f ']' 00:16:37.464 17:38:08 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@633 -- # write_unit_size=1 00:16:37.464 17:38:08 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@635 -- # dd if=/dev/urandom of=/dev/nbd0 bs=4096 count=7936 oflag=direct 00:16:38.033 7936+0 records in 00:16:38.033 7936+0 records out 00:16:38.033 32505856 bytes (33 MB, 31 MiB) copied, 0.650538 s, 50.0 MB/s 00:16:38.033 17:38:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@636 -- # nbd_stop_disks /var/tmp/spdk.sock /dev/nbd0 00:16:38.033 17:38:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:16:38.033 17:38:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:16:38.033 17:38:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@50 -- # local nbd_list 00:16:38.033 17:38:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@51 -- # local i 00:16:38.033 17:38:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:16:38.033 17:38:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:16:38.292 [2024-11-27 17:38:09.246774] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:38.292 17:38:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:16:38.292 17:38:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:16:38.292 17:38:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:16:38.292 17:38:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:16:38.292 17:38:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:16:38.292 17:38:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:16:38.292 17:38:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@41 -- # break 00:16:38.292 17:38:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@45 -- # return 0 00:16:38.292 17:38:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:16:38.292 17:38:09 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:38.292 17:38:09 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:38.292 [2024-11-27 17:38:09.278806] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:16:38.292 17:38:09 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:38.292 17:38:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:38.292 17:38:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:38.292 17:38:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:38.292 17:38:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:38.292 17:38:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:38.292 17:38:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:38.292 17:38:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:38.292 17:38:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:38.292 17:38:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:38.292 17:38:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:38.292 17:38:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:38.292 17:38:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:38.292 17:38:09 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:38.292 17:38:09 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:38.292 17:38:09 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:38.292 17:38:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:38.292 "name": "raid_bdev1", 00:16:38.292 "uuid": "73f0ad94-d8a4-4ed9-97dc-c66b90d469e0", 00:16:38.292 "strip_size_kb": 0, 00:16:38.292 "state": "online", 00:16:38.292 "raid_level": "raid1", 00:16:38.292 "superblock": true, 00:16:38.292 "num_base_bdevs": 2, 00:16:38.292 "num_base_bdevs_discovered": 1, 00:16:38.292 "num_base_bdevs_operational": 1, 00:16:38.292 "base_bdevs_list": [ 00:16:38.292 { 00:16:38.292 "name": null, 00:16:38.292 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:38.292 "is_configured": false, 00:16:38.292 "data_offset": 0, 00:16:38.292 "data_size": 7936 00:16:38.292 }, 00:16:38.292 { 00:16:38.292 "name": "BaseBdev2", 00:16:38.292 "uuid": "3f16235d-acd4-5104-a066-8085bfe43bf7", 00:16:38.292 "is_configured": true, 00:16:38.292 "data_offset": 256, 00:16:38.292 "data_size": 7936 00:16:38.292 } 00:16:38.292 ] 00:16:38.292 }' 00:16:38.292 17:38:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:38.292 17:38:09 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:38.552 17:38:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:16:38.552 17:38:09 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:38.552 17:38:09 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:38.552 [2024-11-27 17:38:09.686096] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:16:38.552 [2024-11-27 17:38:09.687942] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d00019c960 00:16:38.552 [2024-11-27 17:38:09.689701] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:16:38.552 17:38:09 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:38.552 17:38:09 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@647 -- # sleep 1 00:16:39.936 17:38:10 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:39.936 17:38:10 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:39.936 17:38:10 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:39.936 17:38:10 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:39.936 17:38:10 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:39.936 17:38:10 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:39.936 17:38:10 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:39.936 17:38:10 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:39.936 17:38:10 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:39.936 17:38:10 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:39.936 17:38:10 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:39.936 "name": "raid_bdev1", 00:16:39.936 "uuid": "73f0ad94-d8a4-4ed9-97dc-c66b90d469e0", 00:16:39.936 "strip_size_kb": 0, 00:16:39.936 "state": "online", 00:16:39.936 "raid_level": "raid1", 00:16:39.936 "superblock": true, 00:16:39.936 "num_base_bdevs": 2, 00:16:39.936 "num_base_bdevs_discovered": 2, 00:16:39.936 "num_base_bdevs_operational": 2, 00:16:39.936 "process": { 00:16:39.936 "type": "rebuild", 00:16:39.936 "target": "spare", 00:16:39.936 "progress": { 00:16:39.936 "blocks": 2560, 00:16:39.936 "percent": 32 00:16:39.936 } 00:16:39.936 }, 00:16:39.936 "base_bdevs_list": [ 00:16:39.936 { 00:16:39.936 "name": "spare", 00:16:39.936 "uuid": "521bd60f-f352-50a4-b567-d339f52247a6", 00:16:39.936 "is_configured": true, 00:16:39.936 "data_offset": 256, 00:16:39.936 "data_size": 7936 00:16:39.936 }, 00:16:39.936 { 00:16:39.936 "name": "BaseBdev2", 00:16:39.936 "uuid": "3f16235d-acd4-5104-a066-8085bfe43bf7", 00:16:39.936 "is_configured": true, 00:16:39.936 "data_offset": 256, 00:16:39.936 "data_size": 7936 00:16:39.936 } 00:16:39.936 ] 00:16:39.936 }' 00:16:39.936 17:38:10 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:39.936 17:38:10 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:39.936 17:38:10 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:39.936 17:38:10 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:39.936 17:38:10 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:16:39.936 17:38:10 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:39.936 17:38:10 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:39.936 [2024-11-27 17:38:10.853071] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:39.936 [2024-11-27 17:38:10.894406] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:16:39.936 [2024-11-27 17:38:10.894469] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:39.936 [2024-11-27 17:38:10.894506] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:39.936 [2024-11-27 17:38:10.894517] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:16:39.936 17:38:10 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:39.936 17:38:10 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:39.936 17:38:10 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:39.936 17:38:10 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:39.936 17:38:10 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:39.936 17:38:10 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:39.936 17:38:10 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:39.936 17:38:10 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:39.936 17:38:10 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:39.936 17:38:10 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:39.936 17:38:10 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:39.936 17:38:10 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:39.936 17:38:10 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:39.936 17:38:10 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:39.936 17:38:10 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:39.936 17:38:10 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:39.936 17:38:10 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:39.936 "name": "raid_bdev1", 00:16:39.936 "uuid": "73f0ad94-d8a4-4ed9-97dc-c66b90d469e0", 00:16:39.936 "strip_size_kb": 0, 00:16:39.936 "state": "online", 00:16:39.936 "raid_level": "raid1", 00:16:39.936 "superblock": true, 00:16:39.936 "num_base_bdevs": 2, 00:16:39.936 "num_base_bdevs_discovered": 1, 00:16:39.936 "num_base_bdevs_operational": 1, 00:16:39.936 "base_bdevs_list": [ 00:16:39.936 { 00:16:39.936 "name": null, 00:16:39.936 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:39.936 "is_configured": false, 00:16:39.936 "data_offset": 0, 00:16:39.936 "data_size": 7936 00:16:39.936 }, 00:16:39.936 { 00:16:39.936 "name": "BaseBdev2", 00:16:39.936 "uuid": "3f16235d-acd4-5104-a066-8085bfe43bf7", 00:16:39.936 "is_configured": true, 00:16:39.936 "data_offset": 256, 00:16:39.936 "data_size": 7936 00:16:39.936 } 00:16:39.936 ] 00:16:39.936 }' 00:16:39.936 17:38:10 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:39.936 17:38:10 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:40.197 17:38:11 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:16:40.197 17:38:11 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:40.197 17:38:11 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:16:40.197 17:38:11 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=none 00:16:40.197 17:38:11 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:40.197 17:38:11 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:40.197 17:38:11 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:40.197 17:38:11 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:40.197 17:38:11 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:40.458 17:38:11 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:40.458 17:38:11 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:40.458 "name": "raid_bdev1", 00:16:40.458 "uuid": "73f0ad94-d8a4-4ed9-97dc-c66b90d469e0", 00:16:40.458 "strip_size_kb": 0, 00:16:40.458 "state": "online", 00:16:40.458 "raid_level": "raid1", 00:16:40.458 "superblock": true, 00:16:40.458 "num_base_bdevs": 2, 00:16:40.458 "num_base_bdevs_discovered": 1, 00:16:40.458 "num_base_bdevs_operational": 1, 00:16:40.458 "base_bdevs_list": [ 00:16:40.458 { 00:16:40.458 "name": null, 00:16:40.458 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:40.458 "is_configured": false, 00:16:40.458 "data_offset": 0, 00:16:40.458 "data_size": 7936 00:16:40.458 }, 00:16:40.458 { 00:16:40.458 "name": "BaseBdev2", 00:16:40.458 "uuid": "3f16235d-acd4-5104-a066-8085bfe43bf7", 00:16:40.458 "is_configured": true, 00:16:40.458 "data_offset": 256, 00:16:40.458 "data_size": 7936 00:16:40.458 } 00:16:40.458 ] 00:16:40.458 }' 00:16:40.458 17:38:11 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:40.458 17:38:11 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:16:40.458 17:38:11 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:40.458 17:38:11 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:16:40.458 17:38:11 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:16:40.458 17:38:11 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:40.458 17:38:11 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:40.458 [2024-11-27 17:38:11.488315] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:16:40.458 [2024-11-27 17:38:11.490196] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d00019ca30 00:16:40.458 [2024-11-27 17:38:11.491996] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:16:40.458 17:38:11 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:40.458 17:38:11 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@663 -- # sleep 1 00:16:41.398 17:38:12 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:41.398 17:38:12 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:41.398 17:38:12 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:41.398 17:38:12 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:41.398 17:38:12 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:41.398 17:38:12 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:41.398 17:38:12 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:41.398 17:38:12 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:41.398 17:38:12 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:41.398 17:38:12 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:41.398 17:38:12 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:41.398 "name": "raid_bdev1", 00:16:41.398 "uuid": "73f0ad94-d8a4-4ed9-97dc-c66b90d469e0", 00:16:41.398 "strip_size_kb": 0, 00:16:41.398 "state": "online", 00:16:41.398 "raid_level": "raid1", 00:16:41.398 "superblock": true, 00:16:41.398 "num_base_bdevs": 2, 00:16:41.398 "num_base_bdevs_discovered": 2, 00:16:41.398 "num_base_bdevs_operational": 2, 00:16:41.398 "process": { 00:16:41.398 "type": "rebuild", 00:16:41.398 "target": "spare", 00:16:41.398 "progress": { 00:16:41.398 "blocks": 2560, 00:16:41.398 "percent": 32 00:16:41.398 } 00:16:41.398 }, 00:16:41.398 "base_bdevs_list": [ 00:16:41.398 { 00:16:41.398 "name": "spare", 00:16:41.398 "uuid": "521bd60f-f352-50a4-b567-d339f52247a6", 00:16:41.399 "is_configured": true, 00:16:41.399 "data_offset": 256, 00:16:41.399 "data_size": 7936 00:16:41.399 }, 00:16:41.399 { 00:16:41.399 "name": "BaseBdev2", 00:16:41.399 "uuid": "3f16235d-acd4-5104-a066-8085bfe43bf7", 00:16:41.399 "is_configured": true, 00:16:41.399 "data_offset": 256, 00:16:41.399 "data_size": 7936 00:16:41.399 } 00:16:41.399 ] 00:16:41.399 }' 00:16:41.399 17:38:12 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:41.658 17:38:12 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:41.658 17:38:12 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:41.658 17:38:12 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:41.658 17:38:12 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:16:41.658 17:38:12 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:16:41.658 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:16:41.658 17:38:12 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=2 00:16:41.658 17:38:12 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:16:41.658 17:38:12 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@693 -- # '[' 2 -gt 2 ']' 00:16:41.658 17:38:12 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@706 -- # local timeout=600 00:16:41.658 17:38:12 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:16:41.658 17:38:12 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:41.658 17:38:12 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:41.658 17:38:12 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:41.658 17:38:12 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:41.658 17:38:12 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:41.658 17:38:12 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:41.658 17:38:12 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:41.658 17:38:12 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:41.658 17:38:12 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:41.658 17:38:12 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:41.658 17:38:12 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:41.658 "name": "raid_bdev1", 00:16:41.658 "uuid": "73f0ad94-d8a4-4ed9-97dc-c66b90d469e0", 00:16:41.658 "strip_size_kb": 0, 00:16:41.658 "state": "online", 00:16:41.659 "raid_level": "raid1", 00:16:41.659 "superblock": true, 00:16:41.659 "num_base_bdevs": 2, 00:16:41.659 "num_base_bdevs_discovered": 2, 00:16:41.659 "num_base_bdevs_operational": 2, 00:16:41.659 "process": { 00:16:41.659 "type": "rebuild", 00:16:41.659 "target": "spare", 00:16:41.659 "progress": { 00:16:41.659 "blocks": 2816, 00:16:41.659 "percent": 35 00:16:41.659 } 00:16:41.659 }, 00:16:41.659 "base_bdevs_list": [ 00:16:41.659 { 00:16:41.659 "name": "spare", 00:16:41.659 "uuid": "521bd60f-f352-50a4-b567-d339f52247a6", 00:16:41.659 "is_configured": true, 00:16:41.659 "data_offset": 256, 00:16:41.659 "data_size": 7936 00:16:41.659 }, 00:16:41.659 { 00:16:41.659 "name": "BaseBdev2", 00:16:41.659 "uuid": "3f16235d-acd4-5104-a066-8085bfe43bf7", 00:16:41.659 "is_configured": true, 00:16:41.659 "data_offset": 256, 00:16:41.659 "data_size": 7936 00:16:41.659 } 00:16:41.659 ] 00:16:41.659 }' 00:16:41.659 17:38:12 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:41.659 17:38:12 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:41.659 17:38:12 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:41.659 17:38:12 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:41.659 17:38:12 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@711 -- # sleep 1 00:16:42.598 17:38:13 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:16:42.598 17:38:13 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:42.598 17:38:13 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:42.598 17:38:13 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:42.598 17:38:13 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:42.598 17:38:13 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:42.598 17:38:13 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:42.858 17:38:13 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:42.858 17:38:13 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:42.859 17:38:13 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:42.859 17:38:13 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:42.859 17:38:13 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:42.859 "name": "raid_bdev1", 00:16:42.859 "uuid": "73f0ad94-d8a4-4ed9-97dc-c66b90d469e0", 00:16:42.859 "strip_size_kb": 0, 00:16:42.859 "state": "online", 00:16:42.859 "raid_level": "raid1", 00:16:42.859 "superblock": true, 00:16:42.859 "num_base_bdevs": 2, 00:16:42.859 "num_base_bdevs_discovered": 2, 00:16:42.859 "num_base_bdevs_operational": 2, 00:16:42.859 "process": { 00:16:42.859 "type": "rebuild", 00:16:42.859 "target": "spare", 00:16:42.859 "progress": { 00:16:42.859 "blocks": 5632, 00:16:42.859 "percent": 70 00:16:42.859 } 00:16:42.859 }, 00:16:42.859 "base_bdevs_list": [ 00:16:42.859 { 00:16:42.859 "name": "spare", 00:16:42.859 "uuid": "521bd60f-f352-50a4-b567-d339f52247a6", 00:16:42.859 "is_configured": true, 00:16:42.859 "data_offset": 256, 00:16:42.859 "data_size": 7936 00:16:42.859 }, 00:16:42.859 { 00:16:42.859 "name": "BaseBdev2", 00:16:42.859 "uuid": "3f16235d-acd4-5104-a066-8085bfe43bf7", 00:16:42.859 "is_configured": true, 00:16:42.859 "data_offset": 256, 00:16:42.859 "data_size": 7936 00:16:42.859 } 00:16:42.859 ] 00:16:42.859 }' 00:16:42.859 17:38:13 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:42.859 17:38:13 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:42.859 17:38:13 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:42.859 17:38:13 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:42.859 17:38:13 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@711 -- # sleep 1 00:16:43.428 [2024-11-27 17:38:14.608683] bdev_raid.c:2896:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:16:43.428 [2024-11-27 17:38:14.608769] bdev_raid.c:2558:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:16:43.428 [2024-11-27 17:38:14.608907] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:44.000 17:38:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:16:44.000 17:38:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:44.000 17:38:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:44.000 17:38:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:44.000 17:38:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:44.000 17:38:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:44.000 17:38:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:44.000 17:38:14 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:44.000 17:38:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:44.000 17:38:14 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:44.000 17:38:14 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:44.000 17:38:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:44.000 "name": "raid_bdev1", 00:16:44.000 "uuid": "73f0ad94-d8a4-4ed9-97dc-c66b90d469e0", 00:16:44.000 "strip_size_kb": 0, 00:16:44.000 "state": "online", 00:16:44.000 "raid_level": "raid1", 00:16:44.000 "superblock": true, 00:16:44.000 "num_base_bdevs": 2, 00:16:44.000 "num_base_bdevs_discovered": 2, 00:16:44.000 "num_base_bdevs_operational": 2, 00:16:44.000 "base_bdevs_list": [ 00:16:44.000 { 00:16:44.000 "name": "spare", 00:16:44.000 "uuid": "521bd60f-f352-50a4-b567-d339f52247a6", 00:16:44.000 "is_configured": true, 00:16:44.000 "data_offset": 256, 00:16:44.000 "data_size": 7936 00:16:44.000 }, 00:16:44.000 { 00:16:44.000 "name": "BaseBdev2", 00:16:44.000 "uuid": "3f16235d-acd4-5104-a066-8085bfe43bf7", 00:16:44.001 "is_configured": true, 00:16:44.001 "data_offset": 256, 00:16:44.001 "data_size": 7936 00:16:44.001 } 00:16:44.001 ] 00:16:44.001 }' 00:16:44.001 17:38:14 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:44.001 17:38:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:16:44.001 17:38:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:44.001 17:38:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:16:44.001 17:38:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@709 -- # break 00:16:44.001 17:38:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:16:44.001 17:38:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:44.001 17:38:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:16:44.001 17:38:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=none 00:16:44.001 17:38:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:44.001 17:38:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:44.001 17:38:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:44.001 17:38:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:44.001 17:38:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:44.001 17:38:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:44.001 17:38:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:44.001 "name": "raid_bdev1", 00:16:44.001 "uuid": "73f0ad94-d8a4-4ed9-97dc-c66b90d469e0", 00:16:44.001 "strip_size_kb": 0, 00:16:44.001 "state": "online", 00:16:44.001 "raid_level": "raid1", 00:16:44.001 "superblock": true, 00:16:44.001 "num_base_bdevs": 2, 00:16:44.001 "num_base_bdevs_discovered": 2, 00:16:44.001 "num_base_bdevs_operational": 2, 00:16:44.001 "base_bdevs_list": [ 00:16:44.001 { 00:16:44.001 "name": "spare", 00:16:44.001 "uuid": "521bd60f-f352-50a4-b567-d339f52247a6", 00:16:44.001 "is_configured": true, 00:16:44.001 "data_offset": 256, 00:16:44.001 "data_size": 7936 00:16:44.001 }, 00:16:44.001 { 00:16:44.001 "name": "BaseBdev2", 00:16:44.001 "uuid": "3f16235d-acd4-5104-a066-8085bfe43bf7", 00:16:44.001 "is_configured": true, 00:16:44.001 "data_offset": 256, 00:16:44.001 "data_size": 7936 00:16:44.001 } 00:16:44.001 ] 00:16:44.001 }' 00:16:44.001 17:38:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:44.001 17:38:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:16:44.001 17:38:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:44.262 17:38:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:16:44.262 17:38:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:16:44.262 17:38:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:44.262 17:38:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:44.262 17:38:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:44.262 17:38:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:44.262 17:38:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:44.262 17:38:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:44.262 17:38:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:44.262 17:38:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:44.262 17:38:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:44.262 17:38:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:44.262 17:38:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:44.262 17:38:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:44.262 17:38:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:44.262 17:38:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:44.262 17:38:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:44.262 "name": "raid_bdev1", 00:16:44.262 "uuid": "73f0ad94-d8a4-4ed9-97dc-c66b90d469e0", 00:16:44.262 "strip_size_kb": 0, 00:16:44.262 "state": "online", 00:16:44.262 "raid_level": "raid1", 00:16:44.262 "superblock": true, 00:16:44.262 "num_base_bdevs": 2, 00:16:44.262 "num_base_bdevs_discovered": 2, 00:16:44.262 "num_base_bdevs_operational": 2, 00:16:44.262 "base_bdevs_list": [ 00:16:44.262 { 00:16:44.262 "name": "spare", 00:16:44.262 "uuid": "521bd60f-f352-50a4-b567-d339f52247a6", 00:16:44.262 "is_configured": true, 00:16:44.262 "data_offset": 256, 00:16:44.262 "data_size": 7936 00:16:44.262 }, 00:16:44.262 { 00:16:44.262 "name": "BaseBdev2", 00:16:44.262 "uuid": "3f16235d-acd4-5104-a066-8085bfe43bf7", 00:16:44.262 "is_configured": true, 00:16:44.262 "data_offset": 256, 00:16:44.262 "data_size": 7936 00:16:44.262 } 00:16:44.262 ] 00:16:44.262 }' 00:16:44.262 17:38:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:44.262 17:38:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:44.523 17:38:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:16:44.523 17:38:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:44.523 17:38:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:44.523 [2024-11-27 17:38:15.600297] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:16:44.523 [2024-11-27 17:38:15.600377] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:16:44.523 [2024-11-27 17:38:15.600499] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:44.523 [2024-11-27 17:38:15.600616] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:44.523 [2024-11-27 17:38:15.600701] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:16:44.523 17:38:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:44.523 17:38:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:44.523 17:38:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:44.523 17:38:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:44.523 17:38:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@720 -- # jq length 00:16:44.523 17:38:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:44.523 17:38:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:16:44.523 17:38:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@722 -- # '[' true = true ']' 00:16:44.523 17:38:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@723 -- # '[' false = true ']' 00:16:44.523 17:38:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@737 -- # nbd_start_disks /var/tmp/spdk.sock 'BaseBdev1 spare' '/dev/nbd0 /dev/nbd1' 00:16:44.523 17:38:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk.sock 00:16:44.523 17:38:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@10 -- # bdev_list=('BaseBdev1' 'spare') 00:16:44.523 17:38:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@10 -- # local bdev_list 00:16:44.523 17:38:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:16:44.523 17:38:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@11 -- # local nbd_list 00:16:44.523 17:38:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@12 -- # local i 00:16:44.523 17:38:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:16:44.523 17:38:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:16:44.523 17:38:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk BaseBdev1 /dev/nbd0 00:16:44.784 /dev/nbd0 00:16:44.784 17:38:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:16:44.784 17:38:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:16:44.784 17:38:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:16:44.784 17:38:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@869 -- # local i 00:16:44.784 17:38:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:16:44.784 17:38:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:16:44.784 17:38:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:16:44.784 17:38:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@873 -- # break 00:16:44.784 17:38:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:16:44.784 17:38:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:16:44.784 17:38:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:16:44.784 1+0 records in 00:16:44.784 1+0 records out 00:16:44.784 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000325073 s, 12.6 MB/s 00:16:44.784 17:38:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:16:44.784 17:38:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@886 -- # size=4096 00:16:44.784 17:38:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:16:44.784 17:38:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:16:44.784 17:38:15 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@889 -- # return 0 00:16:44.784 17:38:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:16:44.784 17:38:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:16:44.784 17:38:15 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_start_disk spare /dev/nbd1 00:16:45.044 /dev/nbd1 00:16:45.044 17:38:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:16:45.044 17:38:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:16:45.044 17:38:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@868 -- # local nbd_name=nbd1 00:16:45.044 17:38:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@869 -- # local i 00:16:45.044 17:38:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:16:45.044 17:38:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:16:45.044 17:38:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@872 -- # grep -q -w nbd1 /proc/partitions 00:16:45.044 17:38:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@873 -- # break 00:16:45.044 17:38:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:16:45.044 17:38:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:16:45.044 17:38:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@885 -- # dd if=/dev/nbd1 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:16:45.044 1+0 records in 00:16:45.044 1+0 records out 00:16:45.044 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000399318 s, 10.3 MB/s 00:16:45.044 17:38:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:16:45.044 17:38:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@886 -- # size=4096 00:16:45.044 17:38:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:16:45.044 17:38:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:16:45.044 17:38:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@889 -- # return 0 00:16:45.044 17:38:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:16:45.044 17:38:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:16:45.044 17:38:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@738 -- # cmp -i 1048576 /dev/nbd0 /dev/nbd1 00:16:45.044 17:38:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@739 -- # nbd_stop_disks /var/tmp/spdk.sock '/dev/nbd0 /dev/nbd1' 00:16:45.044 17:38:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk.sock 00:16:45.044 17:38:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:16:45.044 17:38:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@50 -- # local nbd_list 00:16:45.044 17:38:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@51 -- # local i 00:16:45.044 17:38:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:16:45.044 17:38:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd0 00:16:45.304 17:38:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:16:45.304 17:38:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:16:45.304 17:38:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:16:45.304 17:38:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:16:45.304 17:38:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:16:45.304 17:38:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:16:45.304 17:38:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@41 -- # break 00:16:45.304 17:38:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@45 -- # return 0 00:16:45.304 17:38:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:16:45.304 17:38:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk.sock nbd_stop_disk /dev/nbd1 00:16:45.565 17:38:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:16:45.565 17:38:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:16:45.565 17:38:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:16:45.565 17:38:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:16:45.565 17:38:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:16:45.565 17:38:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:16:45.565 17:38:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@41 -- # break 00:16:45.565 17:38:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/nbd_common.sh@45 -- # return 0 00:16:45.565 17:38:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:16:45.565 17:38:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:16:45.565 17:38:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:45.565 17:38:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:45.565 17:38:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:45.565 17:38:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:16:45.565 17:38:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:45.565 17:38:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:45.565 [2024-11-27 17:38:16.632060] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:16:45.565 [2024-11-27 17:38:16.632121] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:45.565 [2024-11-27 17:38:16.632169] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009680 00:16:45.565 [2024-11-27 17:38:16.632183] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:45.565 [2024-11-27 17:38:16.634006] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:45.565 [2024-11-27 17:38:16.634046] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:16:45.565 [2024-11-27 17:38:16.634104] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:16:45.565 [2024-11-27 17:38:16.634167] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:16:45.565 [2024-11-27 17:38:16.634277] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:16:45.565 spare 00:16:45.565 17:38:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:45.565 17:38:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:16:45.565 17:38:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:45.565 17:38:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:45.565 [2024-11-27 17:38:16.734165] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001580 00:16:45.565 [2024-11-27 17:38:16.734188] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4096 00:16:45.565 [2024-11-27 17:38:16.734305] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0001bb1b0 00:16:45.565 [2024-11-27 17:38:16.734401] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001580 00:16:45.565 [2024-11-27 17:38:16.734411] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001580 00:16:45.565 [2024-11-27 17:38:16.734496] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:45.565 17:38:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:45.565 17:38:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:16:45.565 17:38:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:45.565 17:38:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:45.565 17:38:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:45.565 17:38:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:45.565 17:38:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:45.565 17:38:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:45.565 17:38:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:45.565 17:38:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:45.565 17:38:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:45.565 17:38:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:45.565 17:38:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:45.565 17:38:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:45.565 17:38:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:45.825 17:38:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:45.825 17:38:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:45.825 "name": "raid_bdev1", 00:16:45.825 "uuid": "73f0ad94-d8a4-4ed9-97dc-c66b90d469e0", 00:16:45.825 "strip_size_kb": 0, 00:16:45.825 "state": "online", 00:16:45.825 "raid_level": "raid1", 00:16:45.825 "superblock": true, 00:16:45.825 "num_base_bdevs": 2, 00:16:45.825 "num_base_bdevs_discovered": 2, 00:16:45.825 "num_base_bdevs_operational": 2, 00:16:45.825 "base_bdevs_list": [ 00:16:45.825 { 00:16:45.825 "name": "spare", 00:16:45.825 "uuid": "521bd60f-f352-50a4-b567-d339f52247a6", 00:16:45.825 "is_configured": true, 00:16:45.825 "data_offset": 256, 00:16:45.825 "data_size": 7936 00:16:45.825 }, 00:16:45.825 { 00:16:45.825 "name": "BaseBdev2", 00:16:45.825 "uuid": "3f16235d-acd4-5104-a066-8085bfe43bf7", 00:16:45.825 "is_configured": true, 00:16:45.825 "data_offset": 256, 00:16:45.825 "data_size": 7936 00:16:45.825 } 00:16:45.825 ] 00:16:45.825 }' 00:16:45.825 17:38:16 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:45.826 17:38:16 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:46.086 17:38:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:16:46.086 17:38:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:46.086 17:38:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:16:46.086 17:38:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=none 00:16:46.086 17:38:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:46.086 17:38:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:46.086 17:38:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:46.086 17:38:17 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:46.086 17:38:17 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:46.086 17:38:17 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:46.086 17:38:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:46.086 "name": "raid_bdev1", 00:16:46.086 "uuid": "73f0ad94-d8a4-4ed9-97dc-c66b90d469e0", 00:16:46.086 "strip_size_kb": 0, 00:16:46.086 "state": "online", 00:16:46.086 "raid_level": "raid1", 00:16:46.086 "superblock": true, 00:16:46.086 "num_base_bdevs": 2, 00:16:46.086 "num_base_bdevs_discovered": 2, 00:16:46.086 "num_base_bdevs_operational": 2, 00:16:46.086 "base_bdevs_list": [ 00:16:46.086 { 00:16:46.086 "name": "spare", 00:16:46.086 "uuid": "521bd60f-f352-50a4-b567-d339f52247a6", 00:16:46.086 "is_configured": true, 00:16:46.086 "data_offset": 256, 00:16:46.086 "data_size": 7936 00:16:46.086 }, 00:16:46.086 { 00:16:46.086 "name": "BaseBdev2", 00:16:46.086 "uuid": "3f16235d-acd4-5104-a066-8085bfe43bf7", 00:16:46.086 "is_configured": true, 00:16:46.086 "data_offset": 256, 00:16:46.086 "data_size": 7936 00:16:46.086 } 00:16:46.086 ] 00:16:46.086 }' 00:16:46.086 17:38:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:46.086 17:38:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:16:46.086 17:38:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:46.347 17:38:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:16:46.347 17:38:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:16:46.347 17:38:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:46.347 17:38:17 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:46.347 17:38:17 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:46.347 17:38:17 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:46.347 17:38:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:16:46.347 17:38:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:16:46.347 17:38:17 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:46.347 17:38:17 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:46.347 [2024-11-27 17:38:17.347074] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:46.347 17:38:17 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:46.347 17:38:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:46.347 17:38:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:46.347 17:38:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:46.347 17:38:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:46.347 17:38:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:46.347 17:38:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:46.347 17:38:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:46.347 17:38:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:46.347 17:38:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:46.347 17:38:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:46.347 17:38:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:46.347 17:38:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:46.347 17:38:17 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:46.347 17:38:17 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:46.347 17:38:17 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:46.347 17:38:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:46.347 "name": "raid_bdev1", 00:16:46.347 "uuid": "73f0ad94-d8a4-4ed9-97dc-c66b90d469e0", 00:16:46.347 "strip_size_kb": 0, 00:16:46.347 "state": "online", 00:16:46.347 "raid_level": "raid1", 00:16:46.347 "superblock": true, 00:16:46.347 "num_base_bdevs": 2, 00:16:46.347 "num_base_bdevs_discovered": 1, 00:16:46.347 "num_base_bdevs_operational": 1, 00:16:46.347 "base_bdevs_list": [ 00:16:46.347 { 00:16:46.347 "name": null, 00:16:46.347 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:46.347 "is_configured": false, 00:16:46.347 "data_offset": 0, 00:16:46.347 "data_size": 7936 00:16:46.348 }, 00:16:46.348 { 00:16:46.348 "name": "BaseBdev2", 00:16:46.348 "uuid": "3f16235d-acd4-5104-a066-8085bfe43bf7", 00:16:46.348 "is_configured": true, 00:16:46.348 "data_offset": 256, 00:16:46.348 "data_size": 7936 00:16:46.348 } 00:16:46.348 ] 00:16:46.348 }' 00:16:46.348 17:38:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:46.348 17:38:17 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:46.918 17:38:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:16:46.918 17:38:17 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:46.918 17:38:17 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:46.918 [2024-11-27 17:38:17.806292] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:16:46.918 [2024-11-27 17:38:17.806430] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:16:46.918 [2024-11-27 17:38:17.806443] bdev_raid.c:3748:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:16:46.918 [2024-11-27 17:38:17.806503] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:16:46.918 [2024-11-27 17:38:17.808133] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0001bb280 00:16:46.918 [2024-11-27 17:38:17.809885] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:16:46.918 17:38:17 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:46.918 17:38:17 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@757 -- # sleep 1 00:16:47.859 17:38:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:47.859 17:38:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:47.859 17:38:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:47.859 17:38:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:47.859 17:38:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:47.859 17:38:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:47.859 17:38:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:47.859 17:38:18 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:47.859 17:38:18 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:47.859 17:38:18 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:47.859 17:38:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:47.859 "name": "raid_bdev1", 00:16:47.859 "uuid": "73f0ad94-d8a4-4ed9-97dc-c66b90d469e0", 00:16:47.859 "strip_size_kb": 0, 00:16:47.859 "state": "online", 00:16:47.859 "raid_level": "raid1", 00:16:47.859 "superblock": true, 00:16:47.859 "num_base_bdevs": 2, 00:16:47.859 "num_base_bdevs_discovered": 2, 00:16:47.859 "num_base_bdevs_operational": 2, 00:16:47.859 "process": { 00:16:47.859 "type": "rebuild", 00:16:47.859 "target": "spare", 00:16:47.859 "progress": { 00:16:47.859 "blocks": 2560, 00:16:47.859 "percent": 32 00:16:47.859 } 00:16:47.859 }, 00:16:47.859 "base_bdevs_list": [ 00:16:47.859 { 00:16:47.859 "name": "spare", 00:16:47.859 "uuid": "521bd60f-f352-50a4-b567-d339f52247a6", 00:16:47.859 "is_configured": true, 00:16:47.859 "data_offset": 256, 00:16:47.859 "data_size": 7936 00:16:47.859 }, 00:16:47.859 { 00:16:47.859 "name": "BaseBdev2", 00:16:47.859 "uuid": "3f16235d-acd4-5104-a066-8085bfe43bf7", 00:16:47.859 "is_configured": true, 00:16:47.859 "data_offset": 256, 00:16:47.859 "data_size": 7936 00:16:47.859 } 00:16:47.859 ] 00:16:47.859 }' 00:16:47.859 17:38:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:47.859 17:38:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:47.859 17:38:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:47.859 17:38:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:47.859 17:38:18 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:16:47.859 17:38:18 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:47.859 17:38:18 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:47.859 [2024-11-27 17:38:18.981422] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:47.859 [2024-11-27 17:38:19.013737] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:16:47.859 [2024-11-27 17:38:19.013789] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:47.859 [2024-11-27 17:38:19.013806] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:47.859 [2024-11-27 17:38:19.013812] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:16:47.859 17:38:19 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:47.859 17:38:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:47.859 17:38:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:47.859 17:38:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:47.859 17:38:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:47.859 17:38:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:47.859 17:38:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:47.859 17:38:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:47.859 17:38:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:47.859 17:38:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:47.859 17:38:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:47.859 17:38:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:47.859 17:38:19 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:47.859 17:38:19 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:47.859 17:38:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:47.859 17:38:19 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:48.119 17:38:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:48.119 "name": "raid_bdev1", 00:16:48.119 "uuid": "73f0ad94-d8a4-4ed9-97dc-c66b90d469e0", 00:16:48.119 "strip_size_kb": 0, 00:16:48.119 "state": "online", 00:16:48.119 "raid_level": "raid1", 00:16:48.119 "superblock": true, 00:16:48.119 "num_base_bdevs": 2, 00:16:48.119 "num_base_bdevs_discovered": 1, 00:16:48.119 "num_base_bdevs_operational": 1, 00:16:48.119 "base_bdevs_list": [ 00:16:48.119 { 00:16:48.119 "name": null, 00:16:48.119 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:48.119 "is_configured": false, 00:16:48.119 "data_offset": 0, 00:16:48.119 "data_size": 7936 00:16:48.119 }, 00:16:48.119 { 00:16:48.119 "name": "BaseBdev2", 00:16:48.119 "uuid": "3f16235d-acd4-5104-a066-8085bfe43bf7", 00:16:48.119 "is_configured": true, 00:16:48.119 "data_offset": 256, 00:16:48.119 "data_size": 7936 00:16:48.119 } 00:16:48.119 ] 00:16:48.119 }' 00:16:48.119 17:38:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:48.119 17:38:19 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:48.379 17:38:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:16:48.379 17:38:19 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:48.379 17:38:19 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:48.379 [2024-11-27 17:38:19.464588] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:16:48.379 [2024-11-27 17:38:19.464692] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:48.379 [2024-11-27 17:38:19.464735] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009f80 00:16:48.379 [2024-11-27 17:38:19.464766] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:48.379 [2024-11-27 17:38:19.464980] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:48.379 [2024-11-27 17:38:19.465033] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:16:48.379 [2024-11-27 17:38:19.465109] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:16:48.379 [2024-11-27 17:38:19.465161] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:16:48.379 [2024-11-27 17:38:19.465212] bdev_raid.c:3748:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:16:48.379 [2024-11-27 17:38:19.465263] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:16:48.379 [2024-11-27 17:38:19.466698] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0001bb350 00:16:48.379 [2024-11-27 17:38:19.468551] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:16:48.379 spare 00:16:48.379 17:38:19 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:48.379 17:38:19 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@764 -- # sleep 1 00:16:49.320 17:38:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:16:49.320 17:38:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:49.320 17:38:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:16:49.320 17:38:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=spare 00:16:49.320 17:38:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:49.320 17:38:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:49.320 17:38:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:49.320 17:38:20 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:49.320 17:38:20 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:49.320 17:38:20 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:49.580 17:38:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:49.580 "name": "raid_bdev1", 00:16:49.580 "uuid": "73f0ad94-d8a4-4ed9-97dc-c66b90d469e0", 00:16:49.580 "strip_size_kb": 0, 00:16:49.580 "state": "online", 00:16:49.581 "raid_level": "raid1", 00:16:49.581 "superblock": true, 00:16:49.581 "num_base_bdevs": 2, 00:16:49.581 "num_base_bdevs_discovered": 2, 00:16:49.581 "num_base_bdevs_operational": 2, 00:16:49.581 "process": { 00:16:49.581 "type": "rebuild", 00:16:49.581 "target": "spare", 00:16:49.581 "progress": { 00:16:49.581 "blocks": 2560, 00:16:49.581 "percent": 32 00:16:49.581 } 00:16:49.581 }, 00:16:49.581 "base_bdevs_list": [ 00:16:49.581 { 00:16:49.581 "name": "spare", 00:16:49.581 "uuid": "521bd60f-f352-50a4-b567-d339f52247a6", 00:16:49.581 "is_configured": true, 00:16:49.581 "data_offset": 256, 00:16:49.581 "data_size": 7936 00:16:49.581 }, 00:16:49.581 { 00:16:49.581 "name": "BaseBdev2", 00:16:49.581 "uuid": "3f16235d-acd4-5104-a066-8085bfe43bf7", 00:16:49.581 "is_configured": true, 00:16:49.581 "data_offset": 256, 00:16:49.581 "data_size": 7936 00:16:49.581 } 00:16:49.581 ] 00:16:49.581 }' 00:16:49.581 17:38:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:49.581 17:38:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:16:49.581 17:38:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:49.581 17:38:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:16:49.581 17:38:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:16:49.581 17:38:20 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:49.581 17:38:20 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:49.581 [2024-11-27 17:38:20.607615] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:49.581 [2024-11-27 17:38:20.672355] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:16:49.581 [2024-11-27 17:38:20.672414] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:49.581 [2024-11-27 17:38:20.672428] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:16:49.581 [2024-11-27 17:38:20.672437] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:16:49.581 17:38:20 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:49.581 17:38:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:49.581 17:38:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:49.581 17:38:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:49.581 17:38:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:49.581 17:38:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:49.581 17:38:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:49.581 17:38:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:49.581 17:38:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:49.581 17:38:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:49.581 17:38:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:49.581 17:38:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:49.581 17:38:20 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:49.581 17:38:20 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:49.581 17:38:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:49.581 17:38:20 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:49.581 17:38:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:49.581 "name": "raid_bdev1", 00:16:49.581 "uuid": "73f0ad94-d8a4-4ed9-97dc-c66b90d469e0", 00:16:49.581 "strip_size_kb": 0, 00:16:49.581 "state": "online", 00:16:49.581 "raid_level": "raid1", 00:16:49.581 "superblock": true, 00:16:49.581 "num_base_bdevs": 2, 00:16:49.581 "num_base_bdevs_discovered": 1, 00:16:49.581 "num_base_bdevs_operational": 1, 00:16:49.581 "base_bdevs_list": [ 00:16:49.581 { 00:16:49.581 "name": null, 00:16:49.581 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:49.581 "is_configured": false, 00:16:49.581 "data_offset": 0, 00:16:49.581 "data_size": 7936 00:16:49.581 }, 00:16:49.581 { 00:16:49.581 "name": "BaseBdev2", 00:16:49.581 "uuid": "3f16235d-acd4-5104-a066-8085bfe43bf7", 00:16:49.581 "is_configured": true, 00:16:49.581 "data_offset": 256, 00:16:49.581 "data_size": 7936 00:16:49.581 } 00:16:49.581 ] 00:16:49.581 }' 00:16:49.581 17:38:20 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:49.581 17:38:20 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:50.151 17:38:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:16:50.151 17:38:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:50.151 17:38:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:16:50.151 17:38:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=none 00:16:50.151 17:38:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:50.151 17:38:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:50.151 17:38:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:50.151 17:38:21 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:50.151 17:38:21 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:50.151 17:38:21 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:50.151 17:38:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:50.151 "name": "raid_bdev1", 00:16:50.151 "uuid": "73f0ad94-d8a4-4ed9-97dc-c66b90d469e0", 00:16:50.151 "strip_size_kb": 0, 00:16:50.151 "state": "online", 00:16:50.151 "raid_level": "raid1", 00:16:50.151 "superblock": true, 00:16:50.151 "num_base_bdevs": 2, 00:16:50.151 "num_base_bdevs_discovered": 1, 00:16:50.151 "num_base_bdevs_operational": 1, 00:16:50.151 "base_bdevs_list": [ 00:16:50.151 { 00:16:50.151 "name": null, 00:16:50.151 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:50.151 "is_configured": false, 00:16:50.151 "data_offset": 0, 00:16:50.151 "data_size": 7936 00:16:50.151 }, 00:16:50.151 { 00:16:50.151 "name": "BaseBdev2", 00:16:50.151 "uuid": "3f16235d-acd4-5104-a066-8085bfe43bf7", 00:16:50.151 "is_configured": true, 00:16:50.151 "data_offset": 256, 00:16:50.151 "data_size": 7936 00:16:50.151 } 00:16:50.151 ] 00:16:50.151 }' 00:16:50.151 17:38:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:50.151 17:38:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:16:50.151 17:38:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:50.151 17:38:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:16:50.151 17:38:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:16:50.151 17:38:21 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:50.151 17:38:21 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:50.151 17:38:21 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:50.151 17:38:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:16:50.151 17:38:21 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:50.152 17:38:21 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:50.152 [2024-11-27 17:38:21.247247] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:16:50.152 [2024-11-27 17:38:21.247301] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:50.152 [2024-11-27 17:38:21.247319] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a580 00:16:50.152 [2024-11-27 17:38:21.247328] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:50.152 [2024-11-27 17:38:21.247494] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:50.152 [2024-11-27 17:38:21.247510] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:16:50.152 [2024-11-27 17:38:21.247551] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:16:50.152 [2024-11-27 17:38:21.247569] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:16:50.152 [2024-11-27 17:38:21.247576] bdev_raid.c:3709:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:16:50.152 [2024-11-27 17:38:21.247594] bdev_raid.c:3884:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:16:50.152 BaseBdev1 00:16:50.152 17:38:21 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:50.152 17:38:21 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@775 -- # sleep 1 00:16:51.092 17:38:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:51.092 17:38:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:51.092 17:38:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:51.092 17:38:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:51.092 17:38:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:51.092 17:38:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:51.092 17:38:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:51.092 17:38:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:51.093 17:38:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:51.093 17:38:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:51.093 17:38:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:51.093 17:38:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:51.093 17:38:22 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:51.093 17:38:22 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:51.093 17:38:22 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:51.353 17:38:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:51.353 "name": "raid_bdev1", 00:16:51.353 "uuid": "73f0ad94-d8a4-4ed9-97dc-c66b90d469e0", 00:16:51.353 "strip_size_kb": 0, 00:16:51.353 "state": "online", 00:16:51.353 "raid_level": "raid1", 00:16:51.353 "superblock": true, 00:16:51.353 "num_base_bdevs": 2, 00:16:51.353 "num_base_bdevs_discovered": 1, 00:16:51.353 "num_base_bdevs_operational": 1, 00:16:51.353 "base_bdevs_list": [ 00:16:51.353 { 00:16:51.353 "name": null, 00:16:51.353 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:51.353 "is_configured": false, 00:16:51.353 "data_offset": 0, 00:16:51.353 "data_size": 7936 00:16:51.353 }, 00:16:51.353 { 00:16:51.353 "name": "BaseBdev2", 00:16:51.353 "uuid": "3f16235d-acd4-5104-a066-8085bfe43bf7", 00:16:51.353 "is_configured": true, 00:16:51.353 "data_offset": 256, 00:16:51.353 "data_size": 7936 00:16:51.353 } 00:16:51.353 ] 00:16:51.353 }' 00:16:51.353 17:38:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:51.353 17:38:22 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:51.613 17:38:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:16:51.613 17:38:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:51.613 17:38:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:16:51.613 17:38:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=none 00:16:51.613 17:38:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:51.613 17:38:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:51.613 17:38:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:51.613 17:38:22 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:51.613 17:38:22 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:51.613 17:38:22 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:51.613 17:38:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:51.613 "name": "raid_bdev1", 00:16:51.613 "uuid": "73f0ad94-d8a4-4ed9-97dc-c66b90d469e0", 00:16:51.613 "strip_size_kb": 0, 00:16:51.613 "state": "online", 00:16:51.613 "raid_level": "raid1", 00:16:51.613 "superblock": true, 00:16:51.613 "num_base_bdevs": 2, 00:16:51.613 "num_base_bdevs_discovered": 1, 00:16:51.613 "num_base_bdevs_operational": 1, 00:16:51.613 "base_bdevs_list": [ 00:16:51.613 { 00:16:51.613 "name": null, 00:16:51.613 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:51.613 "is_configured": false, 00:16:51.613 "data_offset": 0, 00:16:51.613 "data_size": 7936 00:16:51.613 }, 00:16:51.613 { 00:16:51.613 "name": "BaseBdev2", 00:16:51.613 "uuid": "3f16235d-acd4-5104-a066-8085bfe43bf7", 00:16:51.613 "is_configured": true, 00:16:51.613 "data_offset": 256, 00:16:51.613 "data_size": 7936 00:16:51.613 } 00:16:51.613 ] 00:16:51.613 }' 00:16:51.613 17:38:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:51.873 17:38:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:16:51.873 17:38:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:51.873 17:38:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:16:51.873 17:38:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:16:51.873 17:38:22 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@650 -- # local es=0 00:16:51.873 17:38:22 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:16:51.873 17:38:22 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:16:51.873 17:38:22 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:16:51.873 17:38:22 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:16:51.873 17:38:22 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:16:51.873 17:38:22 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:16:51.873 17:38:22 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:51.873 17:38:22 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:51.873 [2024-11-27 17:38:22.867276] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:16:51.873 [2024-11-27 17:38:22.867456] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:16:51.873 [2024-11-27 17:38:22.867474] bdev_raid.c:3709:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:16:51.873 request: 00:16:51.873 { 00:16:51.873 "base_bdev": "BaseBdev1", 00:16:51.873 "raid_bdev": "raid_bdev1", 00:16:51.873 "method": "bdev_raid_add_base_bdev", 00:16:51.873 "req_id": 1 00:16:51.873 } 00:16:51.873 Got JSON-RPC error response 00:16:51.874 response: 00:16:51.874 { 00:16:51.874 "code": -22, 00:16:51.874 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:16:51.874 } 00:16:51.874 17:38:22 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:16:51.874 17:38:22 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@653 -- # es=1 00:16:51.874 17:38:22 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:16:51.874 17:38:22 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:16:51.874 17:38:22 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:16:51.874 17:38:22 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@779 -- # sleep 1 00:16:52.814 17:38:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:16:52.814 17:38:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:52.814 17:38:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:52.814 17:38:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:52.814 17:38:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:52.814 17:38:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:52.814 17:38:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:52.814 17:38:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:52.814 17:38:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:52.814 17:38:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:52.814 17:38:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:52.814 17:38:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:52.814 17:38:23 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:52.814 17:38:23 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:52.814 17:38:23 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:52.814 17:38:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:52.814 "name": "raid_bdev1", 00:16:52.814 "uuid": "73f0ad94-d8a4-4ed9-97dc-c66b90d469e0", 00:16:52.814 "strip_size_kb": 0, 00:16:52.814 "state": "online", 00:16:52.814 "raid_level": "raid1", 00:16:52.814 "superblock": true, 00:16:52.814 "num_base_bdevs": 2, 00:16:52.814 "num_base_bdevs_discovered": 1, 00:16:52.814 "num_base_bdevs_operational": 1, 00:16:52.814 "base_bdevs_list": [ 00:16:52.814 { 00:16:52.814 "name": null, 00:16:52.814 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:52.814 "is_configured": false, 00:16:52.814 "data_offset": 0, 00:16:52.814 "data_size": 7936 00:16:52.814 }, 00:16:52.814 { 00:16:52.814 "name": "BaseBdev2", 00:16:52.814 "uuid": "3f16235d-acd4-5104-a066-8085bfe43bf7", 00:16:52.814 "is_configured": true, 00:16:52.814 "data_offset": 256, 00:16:52.814 "data_size": 7936 00:16:52.814 } 00:16:52.814 ] 00:16:52.814 }' 00:16:52.814 17:38:23 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:52.814 17:38:23 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:53.385 17:38:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:16:53.386 17:38:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:16:53.386 17:38:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:16:53.386 17:38:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@171 -- # local target=none 00:16:53.386 17:38:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:16:53.386 17:38:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:53.386 17:38:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:53.386 17:38:24 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:53.386 17:38:24 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:53.386 17:38:24 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:53.386 17:38:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:16:53.386 "name": "raid_bdev1", 00:16:53.386 "uuid": "73f0ad94-d8a4-4ed9-97dc-c66b90d469e0", 00:16:53.386 "strip_size_kb": 0, 00:16:53.386 "state": "online", 00:16:53.386 "raid_level": "raid1", 00:16:53.386 "superblock": true, 00:16:53.386 "num_base_bdevs": 2, 00:16:53.386 "num_base_bdevs_discovered": 1, 00:16:53.386 "num_base_bdevs_operational": 1, 00:16:53.386 "base_bdevs_list": [ 00:16:53.386 { 00:16:53.386 "name": null, 00:16:53.386 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:53.386 "is_configured": false, 00:16:53.386 "data_offset": 0, 00:16:53.386 "data_size": 7936 00:16:53.386 }, 00:16:53.386 { 00:16:53.386 "name": "BaseBdev2", 00:16:53.386 "uuid": "3f16235d-acd4-5104-a066-8085bfe43bf7", 00:16:53.386 "is_configured": true, 00:16:53.386 "data_offset": 256, 00:16:53.386 "data_size": 7936 00:16:53.386 } 00:16:53.386 ] 00:16:53.386 }' 00:16:53.386 17:38:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:16:53.386 17:38:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:16:53.386 17:38:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:16:53.386 17:38:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:16:53.386 17:38:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@784 -- # killprocess 97960 00:16:53.386 17:38:24 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@950 -- # '[' -z 97960 ']' 00:16:53.386 17:38:24 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@954 -- # kill -0 97960 00:16:53.386 17:38:24 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@955 -- # uname 00:16:53.386 17:38:24 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:16:53.386 17:38:24 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 97960 00:16:53.386 17:38:24 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:16:53.386 17:38:24 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:16:53.386 killing process with pid 97960 00:16:53.386 Received shutdown signal, test time was about 60.000000 seconds 00:16:53.386 00:16:53.386 Latency(us) 00:16:53.386 [2024-11-27T17:38:24.578Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:16:53.386 [2024-11-27T17:38:24.578Z] =================================================================================================================== 00:16:53.386 [2024-11-27T17:38:24.578Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:16:53.386 17:38:24 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@968 -- # echo 'killing process with pid 97960' 00:16:53.386 17:38:24 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@969 -- # kill 97960 00:16:53.386 [2024-11-27 17:38:24.561773] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:16:53.386 [2024-11-27 17:38:24.561874] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:53.386 [2024-11-27 17:38:24.561918] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:53.386 [2024-11-27 17:38:24.561927] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state offline 00:16:53.386 17:38:24 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@974 -- # wait 97960 00:16:53.645 [2024-11-27 17:38:24.594693] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:16:53.645 17:38:24 bdev_raid.raid_rebuild_test_sb_md_separate -- bdev/bdev_raid.sh@786 -- # return 0 00:16:53.645 00:16:53.645 real 0m18.189s 00:16:53.645 user 0m23.968s 00:16:53.645 sys 0m2.534s 00:16:53.645 17:38:24 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@1126 -- # xtrace_disable 00:16:53.645 ************************************ 00:16:53.645 END TEST raid_rebuild_test_sb_md_separate 00:16:53.645 ************************************ 00:16:53.645 17:38:24 bdev_raid.raid_rebuild_test_sb_md_separate -- common/autotest_common.sh@10 -- # set +x 00:16:53.905 17:38:24 bdev_raid -- bdev/bdev_raid.sh@1010 -- # base_malloc_params='-m 32 -i' 00:16:53.905 17:38:24 bdev_raid -- bdev/bdev_raid.sh@1011 -- # run_test raid_state_function_test_sb_md_interleaved raid_state_function_test raid1 2 true 00:16:53.905 17:38:24 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:16:53.905 17:38:24 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:16:53.905 17:38:24 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:16:53.905 ************************************ 00:16:53.905 START TEST raid_state_function_test_sb_md_interleaved 00:16:53.905 ************************************ 00:16:53.905 17:38:24 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@1125 -- # raid_state_function_test raid1 2 true 00:16:53.905 17:38:24 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@205 -- # local raid_level=raid1 00:16:53.905 17:38:24 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@206 -- # local num_base_bdevs=2 00:16:53.905 17:38:24 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@207 -- # local superblock=true 00:16:53.905 17:38:24 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@208 -- # local raid_bdev 00:16:53.905 17:38:24 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@209 -- # (( i = 1 )) 00:16:53.905 17:38:24 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:16:53.905 17:38:24 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@211 -- # echo BaseBdev1 00:16:53.905 17:38:24 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:16:53.905 17:38:24 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:16:53.905 17:38:24 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@211 -- # echo BaseBdev2 00:16:53.905 17:38:24 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@209 -- # (( i++ )) 00:16:53.905 17:38:24 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@209 -- # (( i <= num_base_bdevs )) 00:16:53.905 17:38:24 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@209 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:16:53.905 17:38:24 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@209 -- # local base_bdevs 00:16:53.905 17:38:24 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@210 -- # local raid_bdev_name=Existed_Raid 00:16:53.905 17:38:24 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@211 -- # local strip_size 00:16:53.905 17:38:24 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@212 -- # local strip_size_create_arg 00:16:53.905 17:38:24 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@213 -- # local superblock_create_arg 00:16:53.905 17:38:24 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@215 -- # '[' raid1 '!=' raid1 ']' 00:16:53.905 17:38:24 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@219 -- # strip_size=0 00:16:53.905 17:38:24 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@222 -- # '[' true = true ']' 00:16:53.905 17:38:24 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@223 -- # superblock_create_arg=-s 00:16:53.905 17:38:24 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@229 -- # raid_pid=98662 00:16:53.905 17:38:24 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@228 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -i 0 -L bdev_raid 00:16:53.905 17:38:24 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@230 -- # echo 'Process raid pid: 98662' 00:16:53.905 Process raid pid: 98662 00:16:53.905 17:38:24 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@231 -- # waitforlisten 98662 00:16:53.905 17:38:24 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@831 -- # '[' -z 98662 ']' 00:16:53.905 17:38:24 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:16:53.905 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:16:53.905 17:38:24 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@836 -- # local max_retries=100 00:16:53.905 17:38:24 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:16:53.905 17:38:24 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@840 -- # xtrace_disable 00:16:53.905 17:38:24 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:53.905 [2024-11-27 17:38:24.997380] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:16:53.905 [2024-11-27 17:38:24.997517] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:16:54.165 [2024-11-27 17:38:25.143580] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:16:54.165 [2024-11-27 17:38:25.187861] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:16:54.165 [2024-11-27 17:38:25.229903] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:16:54.165 [2024-11-27 17:38:25.229940] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:16:54.735 17:38:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:16:54.735 17:38:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@864 -- # return 0 00:16:54.735 17:38:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@235 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:16:54.735 17:38:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:54.735 17:38:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:54.735 [2024-11-27 17:38:25.811323] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:16:54.735 [2024-11-27 17:38:25.811378] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:16:54.735 [2024-11-27 17:38:25.811406] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:16:54.735 [2024-11-27 17:38:25.811416] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:16:54.735 17:38:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:54.735 17:38:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@236 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:16:54.735 17:38:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:16:54.735 17:38:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:16:54.735 17:38:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:54.735 17:38:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:54.735 17:38:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:54.735 17:38:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:54.735 17:38:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:54.735 17:38:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:54.735 17:38:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:54.735 17:38:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:54.735 17:38:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:54.735 17:38:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:16:54.735 17:38:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:54.735 17:38:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:54.735 17:38:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:54.735 "name": "Existed_Raid", 00:16:54.735 "uuid": "58754d2f-4549-43c3-a9c4-5b363e3a9c38", 00:16:54.735 "strip_size_kb": 0, 00:16:54.735 "state": "configuring", 00:16:54.735 "raid_level": "raid1", 00:16:54.735 "superblock": true, 00:16:54.735 "num_base_bdevs": 2, 00:16:54.735 "num_base_bdevs_discovered": 0, 00:16:54.735 "num_base_bdevs_operational": 2, 00:16:54.735 "base_bdevs_list": [ 00:16:54.735 { 00:16:54.735 "name": "BaseBdev1", 00:16:54.735 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:54.735 "is_configured": false, 00:16:54.735 "data_offset": 0, 00:16:54.735 "data_size": 0 00:16:54.735 }, 00:16:54.735 { 00:16:54.735 "name": "BaseBdev2", 00:16:54.735 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:54.735 "is_configured": false, 00:16:54.735 "data_offset": 0, 00:16:54.735 "data_size": 0 00:16:54.735 } 00:16:54.735 ] 00:16:54.735 }' 00:16:54.735 17:38:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:54.735 17:38:25 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:55.336 17:38:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@237 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:16:55.336 17:38:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:55.336 17:38:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:55.336 [2024-11-27 17:38:26.215242] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:16:55.336 [2024-11-27 17:38:26.215330] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name Existed_Raid, state configuring 00:16:55.336 17:38:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:55.336 17:38:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@241 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:16:55.336 17:38:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:55.336 17:38:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:55.336 [2024-11-27 17:38:26.223266] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev1 00:16:55.336 [2024-11-27 17:38:26.223345] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev1 doesn't exist now 00:16:55.336 [2024-11-27 17:38:26.223381] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:16:55.336 [2024-11-27 17:38:26.223403] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:16:55.336 17:38:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:55.336 17:38:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@242 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -i -b BaseBdev1 00:16:55.336 17:38:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:55.336 17:38:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:55.336 [2024-11-27 17:38:26.240043] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:16:55.336 BaseBdev1 00:16:55.336 17:38:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:55.336 17:38:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@243 -- # waitforbdev BaseBdev1 00:16:55.336 17:38:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev1 00:16:55.336 17:38:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:16:55.336 17:38:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@901 -- # local i 00:16:55.336 17:38:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:16:55.336 17:38:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:16:55.336 17:38:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:16:55.336 17:38:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:55.336 17:38:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:55.336 17:38:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:55.336 17:38:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 -t 2000 00:16:55.336 17:38:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:55.336 17:38:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:55.336 [ 00:16:55.336 { 00:16:55.336 "name": "BaseBdev1", 00:16:55.336 "aliases": [ 00:16:55.336 "990a5d61-d39e-4b90-bfdc-454d94b0986f" 00:16:55.336 ], 00:16:55.336 "product_name": "Malloc disk", 00:16:55.336 "block_size": 4128, 00:16:55.336 "num_blocks": 8192, 00:16:55.336 "uuid": "990a5d61-d39e-4b90-bfdc-454d94b0986f", 00:16:55.336 "md_size": 32, 00:16:55.336 "md_interleave": true, 00:16:55.336 "dif_type": 0, 00:16:55.336 "assigned_rate_limits": { 00:16:55.336 "rw_ios_per_sec": 0, 00:16:55.336 "rw_mbytes_per_sec": 0, 00:16:55.336 "r_mbytes_per_sec": 0, 00:16:55.336 "w_mbytes_per_sec": 0 00:16:55.336 }, 00:16:55.336 "claimed": true, 00:16:55.336 "claim_type": "exclusive_write", 00:16:55.336 "zoned": false, 00:16:55.336 "supported_io_types": { 00:16:55.336 "read": true, 00:16:55.336 "write": true, 00:16:55.336 "unmap": true, 00:16:55.336 "flush": true, 00:16:55.336 "reset": true, 00:16:55.336 "nvme_admin": false, 00:16:55.336 "nvme_io": false, 00:16:55.336 "nvme_io_md": false, 00:16:55.336 "write_zeroes": true, 00:16:55.336 "zcopy": true, 00:16:55.336 "get_zone_info": false, 00:16:55.336 "zone_management": false, 00:16:55.336 "zone_append": false, 00:16:55.336 "compare": false, 00:16:55.336 "compare_and_write": false, 00:16:55.336 "abort": true, 00:16:55.336 "seek_hole": false, 00:16:55.336 "seek_data": false, 00:16:55.336 "copy": true, 00:16:55.336 "nvme_iov_md": false 00:16:55.336 }, 00:16:55.336 "memory_domains": [ 00:16:55.336 { 00:16:55.336 "dma_device_id": "system", 00:16:55.336 "dma_device_type": 1 00:16:55.336 }, 00:16:55.336 { 00:16:55.336 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:55.336 "dma_device_type": 2 00:16:55.336 } 00:16:55.336 ], 00:16:55.336 "driver_specific": {} 00:16:55.336 } 00:16:55.336 ] 00:16:55.336 17:38:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:55.336 17:38:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@907 -- # return 0 00:16:55.337 17:38:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@244 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:16:55.337 17:38:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:16:55.337 17:38:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:16:55.337 17:38:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:55.337 17:38:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:55.337 17:38:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:55.337 17:38:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:55.337 17:38:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:55.337 17:38:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:55.337 17:38:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:55.337 17:38:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:55.337 17:38:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:16:55.337 17:38:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:55.337 17:38:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:55.337 17:38:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:55.337 17:38:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:55.337 "name": "Existed_Raid", 00:16:55.337 "uuid": "cd591d2a-d780-44e2-ba77-1fcbfab21fd7", 00:16:55.337 "strip_size_kb": 0, 00:16:55.337 "state": "configuring", 00:16:55.337 "raid_level": "raid1", 00:16:55.337 "superblock": true, 00:16:55.337 "num_base_bdevs": 2, 00:16:55.337 "num_base_bdevs_discovered": 1, 00:16:55.337 "num_base_bdevs_operational": 2, 00:16:55.337 "base_bdevs_list": [ 00:16:55.337 { 00:16:55.337 "name": "BaseBdev1", 00:16:55.337 "uuid": "990a5d61-d39e-4b90-bfdc-454d94b0986f", 00:16:55.337 "is_configured": true, 00:16:55.337 "data_offset": 256, 00:16:55.337 "data_size": 7936 00:16:55.337 }, 00:16:55.337 { 00:16:55.337 "name": "BaseBdev2", 00:16:55.337 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:55.337 "is_configured": false, 00:16:55.337 "data_offset": 0, 00:16:55.337 "data_size": 0 00:16:55.337 } 00:16:55.337 ] 00:16:55.337 }' 00:16:55.337 17:38:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:55.337 17:38:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:55.597 17:38:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@245 -- # rpc_cmd bdev_raid_delete Existed_Raid 00:16:55.597 17:38:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:55.597 17:38:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:55.597 [2024-11-27 17:38:26.767237] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: Existed_Raid 00:16:55.597 [2024-11-27 17:38:26.767276] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name Existed_Raid, state configuring 00:16:55.597 17:38:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:55.597 17:38:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@249 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n Existed_Raid 00:16:55.597 17:38:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:55.597 17:38:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:55.597 [2024-11-27 17:38:26.779262] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:16:55.597 [2024-11-27 17:38:26.781016] bdev.c:8272:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: BaseBdev2 00:16:55.597 [2024-11-27 17:38:26.781058] bdev_raid_rpc.c: 311:rpc_bdev_raid_create: *DEBUG*: base bdev BaseBdev2 doesn't exist now 00:16:55.856 17:38:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:55.856 17:38:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@250 -- # (( i = 1 )) 00:16:55.856 17:38:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:16:55.856 17:38:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@251 -- # verify_raid_bdev_state Existed_Raid configuring raid1 0 2 00:16:55.856 17:38:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:16:55.856 17:38:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:16:55.856 17:38:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:55.856 17:38:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:55.856 17:38:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:55.856 17:38:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:55.856 17:38:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:55.856 17:38:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:55.856 17:38:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:55.856 17:38:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:55.856 17:38:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:16:55.856 17:38:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:55.856 17:38:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:55.856 17:38:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:55.856 17:38:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:55.856 "name": "Existed_Raid", 00:16:55.856 "uuid": "27b2ecde-fa18-4057-940e-998c7b3e789d", 00:16:55.856 "strip_size_kb": 0, 00:16:55.856 "state": "configuring", 00:16:55.856 "raid_level": "raid1", 00:16:55.856 "superblock": true, 00:16:55.856 "num_base_bdevs": 2, 00:16:55.856 "num_base_bdevs_discovered": 1, 00:16:55.856 "num_base_bdevs_operational": 2, 00:16:55.856 "base_bdevs_list": [ 00:16:55.856 { 00:16:55.856 "name": "BaseBdev1", 00:16:55.856 "uuid": "990a5d61-d39e-4b90-bfdc-454d94b0986f", 00:16:55.856 "is_configured": true, 00:16:55.856 "data_offset": 256, 00:16:55.856 "data_size": 7936 00:16:55.856 }, 00:16:55.856 { 00:16:55.856 "name": "BaseBdev2", 00:16:55.857 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:55.857 "is_configured": false, 00:16:55.857 "data_offset": 0, 00:16:55.857 "data_size": 0 00:16:55.857 } 00:16:55.857 ] 00:16:55.857 }' 00:16:55.857 17:38:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:55.857 17:38:26 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:56.117 17:38:27 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@252 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -i -b BaseBdev2 00:16:56.117 17:38:27 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:56.117 17:38:27 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:56.117 [2024-11-27 17:38:27.229691] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:16:56.117 [2024-11-27 17:38:27.230341] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:16:56.117 [2024-11-27 17:38:27.230497] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4128 00:16:56.117 BaseBdev2 00:16:56.117 [2024-11-27 17:38:27.230892] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:16:56.117 [2024-11-27 17:38:27.231196] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:16:56.117 [2024-11-27 17:38:27.231357] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name Existed_Raid, raid_bdev 0x617000001900 00:16:56.117 [2024-11-27 17:38:27.231754] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:56.117 17:38:27 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:56.117 17:38:27 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@253 -- # waitforbdev BaseBdev2 00:16:56.117 17:38:27 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@899 -- # local bdev_name=BaseBdev2 00:16:56.117 17:38:27 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:16:56.117 17:38:27 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@901 -- # local i 00:16:56.117 17:38:27 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:16:56.117 17:38:27 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:16:56.117 17:38:27 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@904 -- # rpc_cmd bdev_wait_for_examine 00:16:56.117 17:38:27 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:56.117 17:38:27 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:56.117 17:38:27 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:56.117 17:38:27 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@906 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 -t 2000 00:16:56.117 17:38:27 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:56.117 17:38:27 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:56.117 [ 00:16:56.117 { 00:16:56.117 "name": "BaseBdev2", 00:16:56.117 "aliases": [ 00:16:56.117 "0cf817ef-2632-420e-93cb-8801b1c923af" 00:16:56.117 ], 00:16:56.117 "product_name": "Malloc disk", 00:16:56.117 "block_size": 4128, 00:16:56.117 "num_blocks": 8192, 00:16:56.117 "uuid": "0cf817ef-2632-420e-93cb-8801b1c923af", 00:16:56.117 "md_size": 32, 00:16:56.117 "md_interleave": true, 00:16:56.117 "dif_type": 0, 00:16:56.117 "assigned_rate_limits": { 00:16:56.117 "rw_ios_per_sec": 0, 00:16:56.117 "rw_mbytes_per_sec": 0, 00:16:56.117 "r_mbytes_per_sec": 0, 00:16:56.117 "w_mbytes_per_sec": 0 00:16:56.117 }, 00:16:56.117 "claimed": true, 00:16:56.117 "claim_type": "exclusive_write", 00:16:56.117 "zoned": false, 00:16:56.117 "supported_io_types": { 00:16:56.117 "read": true, 00:16:56.117 "write": true, 00:16:56.117 "unmap": true, 00:16:56.117 "flush": true, 00:16:56.117 "reset": true, 00:16:56.117 "nvme_admin": false, 00:16:56.117 "nvme_io": false, 00:16:56.117 "nvme_io_md": false, 00:16:56.117 "write_zeroes": true, 00:16:56.117 "zcopy": true, 00:16:56.117 "get_zone_info": false, 00:16:56.117 "zone_management": false, 00:16:56.117 "zone_append": false, 00:16:56.117 "compare": false, 00:16:56.117 "compare_and_write": false, 00:16:56.117 "abort": true, 00:16:56.117 "seek_hole": false, 00:16:56.117 "seek_data": false, 00:16:56.117 "copy": true, 00:16:56.117 "nvme_iov_md": false 00:16:56.117 }, 00:16:56.117 "memory_domains": [ 00:16:56.117 { 00:16:56.117 "dma_device_id": "system", 00:16:56.117 "dma_device_type": 1 00:16:56.117 }, 00:16:56.117 { 00:16:56.117 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:56.117 "dma_device_type": 2 00:16:56.117 } 00:16:56.117 ], 00:16:56.117 "driver_specific": {} 00:16:56.117 } 00:16:56.117 ] 00:16:56.117 17:38:27 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:56.117 17:38:27 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@907 -- # return 0 00:16:56.117 17:38:27 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@250 -- # (( i++ )) 00:16:56.117 17:38:27 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@250 -- # (( i < num_base_bdevs )) 00:16:56.117 17:38:27 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@255 -- # verify_raid_bdev_state Existed_Raid online raid1 0 2 00:16:56.117 17:38:27 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:16:56.117 17:38:27 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:56.117 17:38:27 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:56.117 17:38:27 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:56.117 17:38:27 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:56.117 17:38:27 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:56.117 17:38:27 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:56.117 17:38:27 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:56.117 17:38:27 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:56.117 17:38:27 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:56.117 17:38:27 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:16:56.117 17:38:27 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:56.117 17:38:27 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:56.117 17:38:27 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:56.377 17:38:27 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:56.377 "name": "Existed_Raid", 00:16:56.377 "uuid": "27b2ecde-fa18-4057-940e-998c7b3e789d", 00:16:56.377 "strip_size_kb": 0, 00:16:56.377 "state": "online", 00:16:56.377 "raid_level": "raid1", 00:16:56.377 "superblock": true, 00:16:56.377 "num_base_bdevs": 2, 00:16:56.377 "num_base_bdevs_discovered": 2, 00:16:56.377 "num_base_bdevs_operational": 2, 00:16:56.377 "base_bdevs_list": [ 00:16:56.377 { 00:16:56.377 "name": "BaseBdev1", 00:16:56.377 "uuid": "990a5d61-d39e-4b90-bfdc-454d94b0986f", 00:16:56.377 "is_configured": true, 00:16:56.377 "data_offset": 256, 00:16:56.377 "data_size": 7936 00:16:56.377 }, 00:16:56.377 { 00:16:56.377 "name": "BaseBdev2", 00:16:56.377 "uuid": "0cf817ef-2632-420e-93cb-8801b1c923af", 00:16:56.377 "is_configured": true, 00:16:56.377 "data_offset": 256, 00:16:56.377 "data_size": 7936 00:16:56.377 } 00:16:56.377 ] 00:16:56.377 }' 00:16:56.377 17:38:27 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:56.377 17:38:27 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:56.638 17:38:27 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@256 -- # verify_raid_bdev_properties Existed_Raid 00:16:56.638 17:38:27 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=Existed_Raid 00:16:56.638 17:38:27 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:16:56.638 17:38:27 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:16:56.638 17:38:27 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@184 -- # local name 00:16:56.638 17:38:27 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:16:56.638 17:38:27 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b Existed_Raid 00:16:56.638 17:38:27 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:16:56.638 17:38:27 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:56.638 17:38:27 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:56.638 [2024-11-27 17:38:27.769060] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:56.638 17:38:27 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:56.638 17:38:27 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:16:56.638 "name": "Existed_Raid", 00:16:56.638 "aliases": [ 00:16:56.638 "27b2ecde-fa18-4057-940e-998c7b3e789d" 00:16:56.638 ], 00:16:56.638 "product_name": "Raid Volume", 00:16:56.638 "block_size": 4128, 00:16:56.638 "num_blocks": 7936, 00:16:56.638 "uuid": "27b2ecde-fa18-4057-940e-998c7b3e789d", 00:16:56.638 "md_size": 32, 00:16:56.638 "md_interleave": true, 00:16:56.638 "dif_type": 0, 00:16:56.638 "assigned_rate_limits": { 00:16:56.638 "rw_ios_per_sec": 0, 00:16:56.638 "rw_mbytes_per_sec": 0, 00:16:56.638 "r_mbytes_per_sec": 0, 00:16:56.638 "w_mbytes_per_sec": 0 00:16:56.638 }, 00:16:56.638 "claimed": false, 00:16:56.638 "zoned": false, 00:16:56.638 "supported_io_types": { 00:16:56.638 "read": true, 00:16:56.638 "write": true, 00:16:56.638 "unmap": false, 00:16:56.638 "flush": false, 00:16:56.638 "reset": true, 00:16:56.638 "nvme_admin": false, 00:16:56.638 "nvme_io": false, 00:16:56.638 "nvme_io_md": false, 00:16:56.638 "write_zeroes": true, 00:16:56.638 "zcopy": false, 00:16:56.638 "get_zone_info": false, 00:16:56.638 "zone_management": false, 00:16:56.638 "zone_append": false, 00:16:56.638 "compare": false, 00:16:56.638 "compare_and_write": false, 00:16:56.638 "abort": false, 00:16:56.638 "seek_hole": false, 00:16:56.638 "seek_data": false, 00:16:56.638 "copy": false, 00:16:56.638 "nvme_iov_md": false 00:16:56.638 }, 00:16:56.638 "memory_domains": [ 00:16:56.638 { 00:16:56.638 "dma_device_id": "system", 00:16:56.638 "dma_device_type": 1 00:16:56.638 }, 00:16:56.638 { 00:16:56.638 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:56.638 "dma_device_type": 2 00:16:56.638 }, 00:16:56.638 { 00:16:56.638 "dma_device_id": "system", 00:16:56.638 "dma_device_type": 1 00:16:56.638 }, 00:16:56.638 { 00:16:56.638 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:56.638 "dma_device_type": 2 00:16:56.638 } 00:16:56.638 ], 00:16:56.638 "driver_specific": { 00:16:56.638 "raid": { 00:16:56.638 "uuid": "27b2ecde-fa18-4057-940e-998c7b3e789d", 00:16:56.638 "strip_size_kb": 0, 00:16:56.638 "state": "online", 00:16:56.638 "raid_level": "raid1", 00:16:56.638 "superblock": true, 00:16:56.638 "num_base_bdevs": 2, 00:16:56.638 "num_base_bdevs_discovered": 2, 00:16:56.638 "num_base_bdevs_operational": 2, 00:16:56.638 "base_bdevs_list": [ 00:16:56.638 { 00:16:56.638 "name": "BaseBdev1", 00:16:56.638 "uuid": "990a5d61-d39e-4b90-bfdc-454d94b0986f", 00:16:56.638 "is_configured": true, 00:16:56.638 "data_offset": 256, 00:16:56.638 "data_size": 7936 00:16:56.638 }, 00:16:56.638 { 00:16:56.638 "name": "BaseBdev2", 00:16:56.638 "uuid": "0cf817ef-2632-420e-93cb-8801b1c923af", 00:16:56.638 "is_configured": true, 00:16:56.638 "data_offset": 256, 00:16:56.638 "data_size": 7936 00:16:56.638 } 00:16:56.638 ] 00:16:56.638 } 00:16:56.638 } 00:16:56.638 }' 00:16:56.638 17:38:27 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:16:56.899 17:38:27 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@188 -- # base_bdev_names='BaseBdev1 00:16:56.899 BaseBdev2' 00:16:56.899 17:38:27 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:56.899 17:38:27 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4128 32 true 0' 00:16:56.899 17:38:27 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:16:56.899 17:38:27 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev1 00:16:56.899 17:38:27 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:56.899 17:38:27 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:56.899 17:38:27 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:56.899 17:38:27 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:56.899 17:38:27 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4128 32 true 0' 00:16:56.899 17:38:27 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@193 -- # [[ 4128 32 true 0 == \4\1\2\8\ \3\2\ \t\r\u\e\ \0 ]] 00:16:56.899 17:38:27 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:16:56.899 17:38:27 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:56.899 17:38:27 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b BaseBdev2 00:16:56.899 17:38:27 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:56.899 17:38:27 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:56.899 17:38:27 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:56.899 17:38:28 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4128 32 true 0' 00:16:56.899 17:38:28 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@193 -- # [[ 4128 32 true 0 == \4\1\2\8\ \3\2\ \t\r\u\e\ \0 ]] 00:16:56.899 17:38:28 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@259 -- # rpc_cmd bdev_malloc_delete BaseBdev1 00:16:56.899 17:38:28 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:56.899 17:38:28 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:56.899 [2024-11-27 17:38:28.024433] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:16:56.899 17:38:28 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:56.899 17:38:28 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@260 -- # local expected_state 00:16:56.899 17:38:28 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@261 -- # has_redundancy raid1 00:16:56.899 17:38:28 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@198 -- # case $1 in 00:16:56.899 17:38:28 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@199 -- # return 0 00:16:56.899 17:38:28 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@264 -- # expected_state=online 00:16:56.899 17:38:28 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@266 -- # verify_raid_bdev_state Existed_Raid online raid1 0 1 00:16:56.899 17:38:28 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=Existed_Raid 00:16:56.899 17:38:28 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:56.899 17:38:28 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:56.899 17:38:28 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:56.899 17:38:28 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:16:56.899 17:38:28 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:56.899 17:38:28 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:56.899 17:38:28 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:56.899 17:38:28 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:56.899 17:38:28 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:56.899 17:38:28 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "Existed_Raid")' 00:16:56.899 17:38:28 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:56.899 17:38:28 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:56.899 17:38:28 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:57.159 17:38:28 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:57.159 "name": "Existed_Raid", 00:16:57.159 "uuid": "27b2ecde-fa18-4057-940e-998c7b3e789d", 00:16:57.159 "strip_size_kb": 0, 00:16:57.159 "state": "online", 00:16:57.159 "raid_level": "raid1", 00:16:57.159 "superblock": true, 00:16:57.159 "num_base_bdevs": 2, 00:16:57.159 "num_base_bdevs_discovered": 1, 00:16:57.159 "num_base_bdevs_operational": 1, 00:16:57.159 "base_bdevs_list": [ 00:16:57.159 { 00:16:57.159 "name": null, 00:16:57.159 "uuid": "00000000-0000-0000-0000-000000000000", 00:16:57.159 "is_configured": false, 00:16:57.159 "data_offset": 0, 00:16:57.159 "data_size": 7936 00:16:57.159 }, 00:16:57.159 { 00:16:57.159 "name": "BaseBdev2", 00:16:57.159 "uuid": "0cf817ef-2632-420e-93cb-8801b1c923af", 00:16:57.159 "is_configured": true, 00:16:57.159 "data_offset": 256, 00:16:57.159 "data_size": 7936 00:16:57.159 } 00:16:57.159 ] 00:16:57.159 }' 00:16:57.159 17:38:28 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:57.159 17:38:28 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:57.419 17:38:28 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@270 -- # (( i = 1 )) 00:16:57.419 17:38:28 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:16:57.419 17:38:28 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@271 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:57.419 17:38:28 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:57.419 17:38:28 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:57.419 17:38:28 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@271 -- # jq -r '.[0]["name"]' 00:16:57.419 17:38:28 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:57.419 17:38:28 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@271 -- # raid_bdev=Existed_Raid 00:16:57.419 17:38:28 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@272 -- # '[' Existed_Raid '!=' Existed_Raid ']' 00:16:57.419 17:38:28 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@276 -- # rpc_cmd bdev_malloc_delete BaseBdev2 00:16:57.419 17:38:28 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:57.419 17:38:28 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:57.419 [2024-11-27 17:38:28.515300] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev2 00:16:57.419 [2024-11-27 17:38:28.515444] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:16:57.419 [2024-11-27 17:38:28.527301] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:57.419 [2024-11-27 17:38:28.527426] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:57.419 [2024-11-27 17:38:28.527470] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name Existed_Raid, state offline 00:16:57.419 17:38:28 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:57.419 17:38:28 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@270 -- # (( i++ )) 00:16:57.419 17:38:28 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@270 -- # (( i < num_base_bdevs )) 00:16:57.419 17:38:28 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@278 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:57.419 17:38:28 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@278 -- # jq -r '.[0]["name"] | select(.)' 00:16:57.419 17:38:28 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:57.419 17:38:28 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:57.419 17:38:28 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:57.419 17:38:28 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@278 -- # raid_bdev= 00:16:57.419 17:38:28 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@279 -- # '[' -n '' ']' 00:16:57.419 17:38:28 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@284 -- # '[' 2 -gt 2 ']' 00:16:57.419 17:38:28 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@326 -- # killprocess 98662 00:16:57.419 17:38:28 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@950 -- # '[' -z 98662 ']' 00:16:57.419 17:38:28 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@954 -- # kill -0 98662 00:16:57.419 17:38:28 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@955 -- # uname 00:16:57.419 17:38:28 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:16:57.419 17:38:28 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 98662 00:16:57.679 17:38:28 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:16:57.679 killing process with pid 98662 00:16:57.679 17:38:28 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:16:57.679 17:38:28 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@968 -- # echo 'killing process with pid 98662' 00:16:57.679 17:38:28 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@969 -- # kill 98662 00:16:57.679 [2024-11-27 17:38:28.616489] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:16:57.679 17:38:28 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@974 -- # wait 98662 00:16:57.679 [2024-11-27 17:38:28.617449] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:16:57.679 17:38:28 bdev_raid.raid_state_function_test_sb_md_interleaved -- bdev/bdev_raid.sh@328 -- # return 0 00:16:57.679 00:16:57.679 real 0m3.950s 00:16:57.679 user 0m6.177s 00:16:57.679 sys 0m0.858s 00:16:57.679 17:38:28 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@1126 -- # xtrace_disable 00:16:57.679 17:38:28 bdev_raid.raid_state_function_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:57.679 ************************************ 00:16:57.679 END TEST raid_state_function_test_sb_md_interleaved 00:16:57.679 ************************************ 00:16:57.939 17:38:28 bdev_raid -- bdev/bdev_raid.sh@1012 -- # run_test raid_superblock_test_md_interleaved raid_superblock_test raid1 2 00:16:57.939 17:38:28 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:16:57.939 17:38:28 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:16:57.939 17:38:28 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:16:57.939 ************************************ 00:16:57.939 START TEST raid_superblock_test_md_interleaved 00:16:57.939 ************************************ 00:16:57.939 17:38:28 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@1125 -- # raid_superblock_test raid1 2 00:16:57.939 17:38:28 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@393 -- # local raid_level=raid1 00:16:57.939 17:38:28 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@394 -- # local num_base_bdevs=2 00:16:57.939 17:38:28 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@395 -- # base_bdevs_malloc=() 00:16:57.939 17:38:28 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@395 -- # local base_bdevs_malloc 00:16:57.939 17:38:28 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@396 -- # base_bdevs_pt=() 00:16:57.939 17:38:28 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@396 -- # local base_bdevs_pt 00:16:57.939 17:38:28 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@397 -- # base_bdevs_pt_uuid=() 00:16:57.939 17:38:28 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@397 -- # local base_bdevs_pt_uuid 00:16:57.939 17:38:28 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@398 -- # local raid_bdev_name=raid_bdev1 00:16:57.939 17:38:28 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@399 -- # local strip_size 00:16:57.939 17:38:28 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@400 -- # local strip_size_create_arg 00:16:57.939 17:38:28 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@401 -- # local raid_bdev_uuid 00:16:57.939 17:38:28 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@402 -- # local raid_bdev 00:16:57.939 17:38:28 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@404 -- # '[' raid1 '!=' raid1 ']' 00:16:57.939 17:38:28 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@408 -- # strip_size=0 00:16:57.939 17:38:28 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@412 -- # raid_pid=98893 00:16:57.939 17:38:28 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@411 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -L bdev_raid 00:16:57.939 17:38:28 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@413 -- # waitforlisten 98893 00:16:57.939 17:38:28 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@831 -- # '[' -z 98893 ']' 00:16:57.939 17:38:28 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:16:57.939 17:38:28 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@836 -- # local max_retries=100 00:16:57.939 17:38:28 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:16:57.939 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:16:57.939 17:38:28 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@840 -- # xtrace_disable 00:16:57.939 17:38:28 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:57.939 [2024-11-27 17:38:29.022426] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:16:57.939 [2024-11-27 17:38:29.022639] [ DPDK EAL parameters: bdev_svc --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid98893 ] 00:16:58.216 [2024-11-27 17:38:29.169393] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:16:58.216 [2024-11-27 17:38:29.213470] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:16:58.216 [2024-11-27 17:38:29.256713] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:16:58.216 [2024-11-27 17:38:29.256834] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:16:58.793 17:38:29 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:16:58.793 17:38:29 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@864 -- # return 0 00:16:58.793 17:38:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@416 -- # (( i = 1 )) 00:16:58.793 17:38:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:16:58.793 17:38:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc1 00:16:58.793 17:38:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt1 00:16:58.793 17:38:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000001 00:16:58.793 17:38:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:16:58.793 17:38:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:16:58.793 17:38:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:16:58.793 17:38:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -i -b malloc1 00:16:58.793 17:38:29 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:58.793 17:38:29 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:58.793 malloc1 00:16:58.793 17:38:29 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:58.793 17:38:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:16:58.793 17:38:29 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:58.793 17:38:29 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:58.793 [2024-11-27 17:38:29.862267] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:16:58.794 [2024-11-27 17:38:29.862393] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:58.794 [2024-11-27 17:38:29.862448] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:16:58.794 [2024-11-27 17:38:29.862478] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:58.794 [2024-11-27 17:38:29.864341] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:58.794 [2024-11-27 17:38:29.864407] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:16:58.794 pt1 00:16:58.794 17:38:29 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:58.794 17:38:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:16:58.794 17:38:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:16:58.794 17:38:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@417 -- # local bdev_malloc=malloc2 00:16:58.794 17:38:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@418 -- # local bdev_pt=pt2 00:16:58.794 17:38:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@419 -- # local bdev_pt_uuid=00000000-0000-0000-0000-000000000002 00:16:58.794 17:38:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@421 -- # base_bdevs_malloc+=($bdev_malloc) 00:16:58.794 17:38:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@422 -- # base_bdevs_pt+=($bdev_pt) 00:16:58.794 17:38:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@423 -- # base_bdevs_pt_uuid+=($bdev_pt_uuid) 00:16:58.794 17:38:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@425 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -i -b malloc2 00:16:58.794 17:38:29 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:58.794 17:38:29 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:58.794 malloc2 00:16:58.794 17:38:29 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:58.794 17:38:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@426 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:16:58.794 17:38:29 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:58.794 17:38:29 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:58.794 [2024-11-27 17:38:29.902463] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:16:58.794 [2024-11-27 17:38:29.902518] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:58.794 [2024-11-27 17:38:29.902532] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:16:58.794 [2024-11-27 17:38:29.902542] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:58.794 [2024-11-27 17:38:29.904324] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:58.794 [2024-11-27 17:38:29.904360] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:16:58.794 pt2 00:16:58.794 17:38:29 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:58.794 17:38:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@416 -- # (( i++ )) 00:16:58.794 17:38:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@416 -- # (( i <= num_base_bdevs )) 00:16:58.794 17:38:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@430 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''pt1 pt2'\''' -n raid_bdev1 -s 00:16:58.794 17:38:29 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:58.794 17:38:29 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:58.794 [2024-11-27 17:38:29.914491] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:16:58.794 [2024-11-27 17:38:29.916172] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:16:58.794 [2024-11-27 17:38:29.916385] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:16:58.794 [2024-11-27 17:38:29.916405] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4128 00:16:58.794 [2024-11-27 17:38:29.916485] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002390 00:16:58.794 [2024-11-27 17:38:29.916541] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:16:58.794 [2024-11-27 17:38:29.916552] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:16:58.794 [2024-11-27 17:38:29.916608] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:16:58.794 17:38:29 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:58.794 17:38:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@431 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:16:58.794 17:38:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:58.794 17:38:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:16:58.794 17:38:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:58.794 17:38:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:58.794 17:38:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:58.794 17:38:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:58.794 17:38:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:58.794 17:38:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:58.794 17:38:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:58.794 17:38:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:58.794 17:38:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:58.794 17:38:29 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:58.794 17:38:29 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:58.794 17:38:29 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:58.794 17:38:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:58.794 "name": "raid_bdev1", 00:16:58.794 "uuid": "10ce1a84-858f-49e1-bd9b-62e61299cb58", 00:16:58.794 "strip_size_kb": 0, 00:16:58.794 "state": "online", 00:16:58.794 "raid_level": "raid1", 00:16:58.794 "superblock": true, 00:16:58.794 "num_base_bdevs": 2, 00:16:58.794 "num_base_bdevs_discovered": 2, 00:16:58.794 "num_base_bdevs_operational": 2, 00:16:58.794 "base_bdevs_list": [ 00:16:58.794 { 00:16:58.794 "name": "pt1", 00:16:58.794 "uuid": "00000000-0000-0000-0000-000000000001", 00:16:58.794 "is_configured": true, 00:16:58.794 "data_offset": 256, 00:16:58.794 "data_size": 7936 00:16:58.794 }, 00:16:58.794 { 00:16:58.794 "name": "pt2", 00:16:58.794 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:58.794 "is_configured": true, 00:16:58.794 "data_offset": 256, 00:16:58.794 "data_size": 7936 00:16:58.794 } 00:16:58.794 ] 00:16:58.794 }' 00:16:58.794 17:38:29 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:58.794 17:38:29 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:59.365 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@432 -- # verify_raid_bdev_properties raid_bdev1 00:16:59.365 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:16:59.365 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:16:59.365 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:16:59.365 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@184 -- # local name 00:16:59.365 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:16:59.365 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:16:59.365 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:59.365 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:59.365 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:16:59.365 [2024-11-27 17:38:30.413907] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:59.365 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:59.365 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:16:59.365 "name": "raid_bdev1", 00:16:59.365 "aliases": [ 00:16:59.365 "10ce1a84-858f-49e1-bd9b-62e61299cb58" 00:16:59.365 ], 00:16:59.365 "product_name": "Raid Volume", 00:16:59.365 "block_size": 4128, 00:16:59.365 "num_blocks": 7936, 00:16:59.365 "uuid": "10ce1a84-858f-49e1-bd9b-62e61299cb58", 00:16:59.365 "md_size": 32, 00:16:59.365 "md_interleave": true, 00:16:59.365 "dif_type": 0, 00:16:59.365 "assigned_rate_limits": { 00:16:59.365 "rw_ios_per_sec": 0, 00:16:59.365 "rw_mbytes_per_sec": 0, 00:16:59.365 "r_mbytes_per_sec": 0, 00:16:59.365 "w_mbytes_per_sec": 0 00:16:59.365 }, 00:16:59.365 "claimed": false, 00:16:59.365 "zoned": false, 00:16:59.365 "supported_io_types": { 00:16:59.365 "read": true, 00:16:59.365 "write": true, 00:16:59.365 "unmap": false, 00:16:59.365 "flush": false, 00:16:59.365 "reset": true, 00:16:59.365 "nvme_admin": false, 00:16:59.365 "nvme_io": false, 00:16:59.365 "nvme_io_md": false, 00:16:59.365 "write_zeroes": true, 00:16:59.365 "zcopy": false, 00:16:59.365 "get_zone_info": false, 00:16:59.365 "zone_management": false, 00:16:59.365 "zone_append": false, 00:16:59.365 "compare": false, 00:16:59.365 "compare_and_write": false, 00:16:59.365 "abort": false, 00:16:59.365 "seek_hole": false, 00:16:59.365 "seek_data": false, 00:16:59.365 "copy": false, 00:16:59.365 "nvme_iov_md": false 00:16:59.365 }, 00:16:59.365 "memory_domains": [ 00:16:59.365 { 00:16:59.365 "dma_device_id": "system", 00:16:59.365 "dma_device_type": 1 00:16:59.365 }, 00:16:59.365 { 00:16:59.365 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:59.365 "dma_device_type": 2 00:16:59.365 }, 00:16:59.365 { 00:16:59.365 "dma_device_id": "system", 00:16:59.365 "dma_device_type": 1 00:16:59.365 }, 00:16:59.365 { 00:16:59.365 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:16:59.365 "dma_device_type": 2 00:16:59.365 } 00:16:59.365 ], 00:16:59.365 "driver_specific": { 00:16:59.365 "raid": { 00:16:59.365 "uuid": "10ce1a84-858f-49e1-bd9b-62e61299cb58", 00:16:59.365 "strip_size_kb": 0, 00:16:59.365 "state": "online", 00:16:59.365 "raid_level": "raid1", 00:16:59.365 "superblock": true, 00:16:59.365 "num_base_bdevs": 2, 00:16:59.365 "num_base_bdevs_discovered": 2, 00:16:59.365 "num_base_bdevs_operational": 2, 00:16:59.365 "base_bdevs_list": [ 00:16:59.365 { 00:16:59.365 "name": "pt1", 00:16:59.365 "uuid": "00000000-0000-0000-0000-000000000001", 00:16:59.365 "is_configured": true, 00:16:59.365 "data_offset": 256, 00:16:59.365 "data_size": 7936 00:16:59.365 }, 00:16:59.365 { 00:16:59.365 "name": "pt2", 00:16:59.365 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:59.365 "is_configured": true, 00:16:59.365 "data_offset": 256, 00:16:59.365 "data_size": 7936 00:16:59.365 } 00:16:59.365 ] 00:16:59.365 } 00:16:59.365 } 00:16:59.365 }' 00:16:59.365 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:16:59.365 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:16:59.365 pt2' 00:16:59.365 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:59.365 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4128 32 true 0' 00:16:59.626 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:16:59.626 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:16:59.626 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:59.626 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:59.626 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:59.626 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:59.626 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4128 32 true 0' 00:16:59.626 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@193 -- # [[ 4128 32 true 0 == \4\1\2\8\ \3\2\ \t\r\u\e\ \0 ]] 00:16:59.626 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:16:59.626 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:16:59.626 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:59.626 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:16:59.626 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:59.626 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:59.626 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4128 32 true 0' 00:16:59.626 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@193 -- # [[ 4128 32 true 0 == \4\1\2\8\ \3\2\ \t\r\u\e\ \0 ]] 00:16:59.626 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@435 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:16:59.626 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@435 -- # jq -r '.[] | .uuid' 00:16:59.626 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:59.626 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:59.626 [2024-11-27 17:38:30.669375] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:16:59.626 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:59.626 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@435 -- # raid_bdev_uuid=10ce1a84-858f-49e1-bd9b-62e61299cb58 00:16:59.626 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@436 -- # '[' -z 10ce1a84-858f-49e1-bd9b-62e61299cb58 ']' 00:16:59.626 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@441 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:16:59.626 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:59.626 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:59.626 [2024-11-27 17:38:30.713061] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:16:59.626 [2024-11-27 17:38:30.713127] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:16:59.626 [2024-11-27 17:38:30.713212] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:16:59.626 [2024-11-27 17:38:30.713286] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:16:59.626 [2024-11-27 17:38:30.713295] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:16:59.626 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:59.626 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@442 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:59.626 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:59.626 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@442 -- # jq -r '.[]' 00:16:59.626 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:59.626 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:59.626 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@442 -- # raid_bdev= 00:16:59.626 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@443 -- # '[' -n '' ']' 00:16:59.626 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:16:59.626 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt1 00:16:59.626 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:59.626 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:59.626 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:59.626 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@448 -- # for i in "${base_bdevs_pt[@]}" 00:16:59.626 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@449 -- # rpc_cmd bdev_passthru_delete pt2 00:16:59.626 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:59.626 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:59.626 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:59.626 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@451 -- # rpc_cmd bdev_get_bdevs 00:16:59.626 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@451 -- # jq -r '[.[] | select(.product_name == "passthru")] | any' 00:16:59.626 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:59.627 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:59.887 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:59.887 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@451 -- # '[' false == true ']' 00:16:59.887 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@457 -- # NOT rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:16:59.887 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@650 -- # local es=0 00:16:59.887 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:16:59.887 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:16:59.887 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:16:59.887 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:16:59.887 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:16:59.887 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_create -r raid1 -b ''\''malloc1 malloc2'\''' -n raid_bdev1 00:16:59.887 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:59.887 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:59.887 [2024-11-27 17:38:30.856830] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc1 is claimed 00:16:59.888 [2024-11-27 17:38:30.858593] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev malloc2 is claimed 00:16:59.888 [2024-11-27 17:38:30.858654] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc1 00:16:59.888 [2024-11-27 17:38:30.858696] bdev_raid.c:3229:raid_bdev_configure_base_bdev_check_sb_cb: *ERROR*: Superblock of a different raid bdev found on bdev malloc2 00:16:59.888 [2024-11-27 17:38:30.858727] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:16:59.888 [2024-11-27 17:38:30.858746] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state configuring 00:16:59.888 request: 00:16:59.888 { 00:16:59.888 "name": "raid_bdev1", 00:16:59.888 "raid_level": "raid1", 00:16:59.888 "base_bdevs": [ 00:16:59.888 "malloc1", 00:16:59.888 "malloc2" 00:16:59.888 ], 00:16:59.888 "superblock": false, 00:16:59.888 "method": "bdev_raid_create", 00:16:59.888 "req_id": 1 00:16:59.888 } 00:16:59.888 Got JSON-RPC error response 00:16:59.888 response: 00:16:59.888 { 00:16:59.888 "code": -17, 00:16:59.888 "message": "Failed to create RAID bdev raid_bdev1: File exists" 00:16:59.888 } 00:16:59.888 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:16:59.888 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@653 -- # es=1 00:16:59.888 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:16:59.888 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:16:59.888 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:16:59.888 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@459 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:59.888 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:59.888 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:59.888 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@459 -- # jq -r '.[]' 00:16:59.888 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:59.888 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@459 -- # raid_bdev= 00:16:59.888 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@460 -- # '[' -n '' ']' 00:16:59.888 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@465 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:16:59.888 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:59.888 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:59.888 [2024-11-27 17:38:30.920690] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:16:59.888 [2024-11-27 17:38:30.920782] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:16:59.888 [2024-11-27 17:38:30.920831] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007e80 00:16:59.888 [2024-11-27 17:38:30.920858] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:16:59.888 [2024-11-27 17:38:30.922685] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:16:59.888 [2024-11-27 17:38:30.922746] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:16:59.888 [2024-11-27 17:38:30.922803] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:16:59.888 [2024-11-27 17:38:30.922845] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:16:59.888 pt1 00:16:59.888 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:59.888 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@468 -- # verify_raid_bdev_state raid_bdev1 configuring raid1 0 2 00:16:59.888 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:16:59.888 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=configuring 00:16:59.888 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:16:59.888 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:16:59.888 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:16:59.888 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:16:59.888 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:16:59.888 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:16:59.888 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:16:59.888 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:16:59.888 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:16:59.888 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:59.888 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:16:59.888 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:59.888 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:16:59.888 "name": "raid_bdev1", 00:16:59.888 "uuid": "10ce1a84-858f-49e1-bd9b-62e61299cb58", 00:16:59.888 "strip_size_kb": 0, 00:16:59.888 "state": "configuring", 00:16:59.888 "raid_level": "raid1", 00:16:59.888 "superblock": true, 00:16:59.888 "num_base_bdevs": 2, 00:16:59.888 "num_base_bdevs_discovered": 1, 00:16:59.888 "num_base_bdevs_operational": 2, 00:16:59.888 "base_bdevs_list": [ 00:16:59.888 { 00:16:59.888 "name": "pt1", 00:16:59.888 "uuid": "00000000-0000-0000-0000-000000000001", 00:16:59.888 "is_configured": true, 00:16:59.888 "data_offset": 256, 00:16:59.888 "data_size": 7936 00:16:59.888 }, 00:16:59.888 { 00:16:59.888 "name": null, 00:16:59.888 "uuid": "00000000-0000-0000-0000-000000000002", 00:16:59.888 "is_configured": false, 00:16:59.888 "data_offset": 256, 00:16:59.888 "data_size": 7936 00:16:59.888 } 00:16:59.888 ] 00:16:59.888 }' 00:16:59.888 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:16:59.888 17:38:30 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:00.459 17:38:31 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@470 -- # '[' 2 -gt 2 ']' 00:17:00.459 17:38:31 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@478 -- # (( i = 1 )) 00:17:00.459 17:38:31 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:17:00.459 17:38:31 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@479 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:17:00.459 17:38:31 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:00.459 17:38:31 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:00.459 [2024-11-27 17:38:31.367950] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:17:00.459 [2024-11-27 17:38:31.368039] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:17:00.459 [2024-11-27 17:38:31.368075] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:17:00.459 [2024-11-27 17:38:31.368083] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:17:00.459 [2024-11-27 17:38:31.368219] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:17:00.459 [2024-11-27 17:38:31.368233] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:17:00.459 [2024-11-27 17:38:31.368271] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:17:00.459 [2024-11-27 17:38:31.368293] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:17:00.459 [2024-11-27 17:38:31.368360] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001900 00:17:00.459 [2024-11-27 17:38:31.368367] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4128 00:17:00.459 [2024-11-27 17:38:31.368433] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:17:00.459 [2024-11-27 17:38:31.368485] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001900 00:17:00.459 [2024-11-27 17:38:31.368496] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001900 00:17:00.459 [2024-11-27 17:38:31.368542] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:17:00.459 pt2 00:17:00.459 17:38:31 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:00.459 17:38:31 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@478 -- # (( i++ )) 00:17:00.459 17:38:31 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@478 -- # (( i < num_base_bdevs )) 00:17:00.459 17:38:31 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@483 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:17:00.459 17:38:31 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:17:00.459 17:38:31 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:17:00.459 17:38:31 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:00.459 17:38:31 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:00.459 17:38:31 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:17:00.459 17:38:31 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:00.459 17:38:31 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:00.459 17:38:31 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:00.459 17:38:31 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:00.459 17:38:31 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:00.459 17:38:31 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:00.459 17:38:31 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:00.459 17:38:31 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:00.459 17:38:31 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:00.459 17:38:31 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:00.459 "name": "raid_bdev1", 00:17:00.459 "uuid": "10ce1a84-858f-49e1-bd9b-62e61299cb58", 00:17:00.459 "strip_size_kb": 0, 00:17:00.459 "state": "online", 00:17:00.459 "raid_level": "raid1", 00:17:00.459 "superblock": true, 00:17:00.459 "num_base_bdevs": 2, 00:17:00.459 "num_base_bdevs_discovered": 2, 00:17:00.459 "num_base_bdevs_operational": 2, 00:17:00.459 "base_bdevs_list": [ 00:17:00.459 { 00:17:00.459 "name": "pt1", 00:17:00.459 "uuid": "00000000-0000-0000-0000-000000000001", 00:17:00.459 "is_configured": true, 00:17:00.459 "data_offset": 256, 00:17:00.459 "data_size": 7936 00:17:00.459 }, 00:17:00.459 { 00:17:00.459 "name": "pt2", 00:17:00.459 "uuid": "00000000-0000-0000-0000-000000000002", 00:17:00.459 "is_configured": true, 00:17:00.459 "data_offset": 256, 00:17:00.459 "data_size": 7936 00:17:00.459 } 00:17:00.459 ] 00:17:00.459 }' 00:17:00.459 17:38:31 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:00.459 17:38:31 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:00.720 17:38:31 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@484 -- # verify_raid_bdev_properties raid_bdev1 00:17:00.720 17:38:31 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@181 -- # local raid_bdev_name=raid_bdev1 00:17:00.720 17:38:31 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@182 -- # local raid_bdev_info 00:17:00.720 17:38:31 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@183 -- # local base_bdev_names 00:17:00.720 17:38:31 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@184 -- # local name 00:17:00.720 17:38:31 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@185 -- # local cmp_raid_bdev cmp_base_bdev 00:17:00.720 17:38:31 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@187 -- # jq '.[]' 00:17:00.720 17:38:31 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@187 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:17:00.720 17:38:31 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:00.720 17:38:31 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:00.720 [2024-11-27 17:38:31.811509] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:17:00.720 17:38:31 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:00.720 17:38:31 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@187 -- # raid_bdev_info='{ 00:17:00.720 "name": "raid_bdev1", 00:17:00.720 "aliases": [ 00:17:00.720 "10ce1a84-858f-49e1-bd9b-62e61299cb58" 00:17:00.720 ], 00:17:00.720 "product_name": "Raid Volume", 00:17:00.720 "block_size": 4128, 00:17:00.720 "num_blocks": 7936, 00:17:00.720 "uuid": "10ce1a84-858f-49e1-bd9b-62e61299cb58", 00:17:00.720 "md_size": 32, 00:17:00.720 "md_interleave": true, 00:17:00.720 "dif_type": 0, 00:17:00.720 "assigned_rate_limits": { 00:17:00.720 "rw_ios_per_sec": 0, 00:17:00.720 "rw_mbytes_per_sec": 0, 00:17:00.720 "r_mbytes_per_sec": 0, 00:17:00.720 "w_mbytes_per_sec": 0 00:17:00.720 }, 00:17:00.720 "claimed": false, 00:17:00.720 "zoned": false, 00:17:00.720 "supported_io_types": { 00:17:00.720 "read": true, 00:17:00.720 "write": true, 00:17:00.720 "unmap": false, 00:17:00.720 "flush": false, 00:17:00.720 "reset": true, 00:17:00.720 "nvme_admin": false, 00:17:00.720 "nvme_io": false, 00:17:00.720 "nvme_io_md": false, 00:17:00.720 "write_zeroes": true, 00:17:00.720 "zcopy": false, 00:17:00.720 "get_zone_info": false, 00:17:00.720 "zone_management": false, 00:17:00.720 "zone_append": false, 00:17:00.720 "compare": false, 00:17:00.720 "compare_and_write": false, 00:17:00.720 "abort": false, 00:17:00.720 "seek_hole": false, 00:17:00.720 "seek_data": false, 00:17:00.720 "copy": false, 00:17:00.720 "nvme_iov_md": false 00:17:00.720 }, 00:17:00.720 "memory_domains": [ 00:17:00.720 { 00:17:00.720 "dma_device_id": "system", 00:17:00.720 "dma_device_type": 1 00:17:00.720 }, 00:17:00.720 { 00:17:00.720 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:17:00.720 "dma_device_type": 2 00:17:00.720 }, 00:17:00.720 { 00:17:00.720 "dma_device_id": "system", 00:17:00.720 "dma_device_type": 1 00:17:00.720 }, 00:17:00.720 { 00:17:00.720 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:17:00.720 "dma_device_type": 2 00:17:00.720 } 00:17:00.720 ], 00:17:00.720 "driver_specific": { 00:17:00.720 "raid": { 00:17:00.720 "uuid": "10ce1a84-858f-49e1-bd9b-62e61299cb58", 00:17:00.720 "strip_size_kb": 0, 00:17:00.720 "state": "online", 00:17:00.720 "raid_level": "raid1", 00:17:00.720 "superblock": true, 00:17:00.720 "num_base_bdevs": 2, 00:17:00.720 "num_base_bdevs_discovered": 2, 00:17:00.720 "num_base_bdevs_operational": 2, 00:17:00.720 "base_bdevs_list": [ 00:17:00.720 { 00:17:00.720 "name": "pt1", 00:17:00.720 "uuid": "00000000-0000-0000-0000-000000000001", 00:17:00.720 "is_configured": true, 00:17:00.720 "data_offset": 256, 00:17:00.720 "data_size": 7936 00:17:00.720 }, 00:17:00.720 { 00:17:00.720 "name": "pt2", 00:17:00.720 "uuid": "00000000-0000-0000-0000-000000000002", 00:17:00.720 "is_configured": true, 00:17:00.720 "data_offset": 256, 00:17:00.720 "data_size": 7936 00:17:00.720 } 00:17:00.720 ] 00:17:00.720 } 00:17:00.720 } 00:17:00.720 }' 00:17:00.720 17:38:31 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@188 -- # jq -r '.driver_specific.raid.base_bdevs_list[] | select(.is_configured == true).name' 00:17:00.720 17:38:31 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@188 -- # base_bdev_names='pt1 00:17:00.720 pt2' 00:17:00.720 17:38:31 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@189 -- # jq -r '[.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:17:00.981 17:38:31 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@189 -- # cmp_raid_bdev='4128 32 true 0' 00:17:00.981 17:38:31 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:17:00.981 17:38:31 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:17:00.981 17:38:31 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt1 00:17:00.981 17:38:31 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:00.981 17:38:31 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:00.981 17:38:31 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:00.981 17:38:31 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4128 32 true 0' 00:17:00.981 17:38:31 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@193 -- # [[ 4128 32 true 0 == \4\1\2\8\ \3\2\ \t\r\u\e\ \0 ]] 00:17:00.981 17:38:31 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@191 -- # for name in $base_bdev_names 00:17:00.981 17:38:31 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # rpc_cmd bdev_get_bdevs -b pt2 00:17:00.981 17:38:31 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # jq -r '.[] | [.block_size, .md_size, .md_interleave, .dif_type] | join(" ")' 00:17:00.981 17:38:31 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:00.981 17:38:31 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:00.981 17:38:31 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:00.981 17:38:32 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@192 -- # cmp_base_bdev='4128 32 true 0' 00:17:00.981 17:38:32 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@193 -- # [[ 4128 32 true 0 == \4\1\2\8\ \3\2\ \t\r\u\e\ \0 ]] 00:17:00.981 17:38:32 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@487 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:17:00.981 17:38:32 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:00.981 17:38:32 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@487 -- # jq -r '.[] | .uuid' 00:17:00.981 17:38:32 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:00.981 [2024-11-27 17:38:32.027432] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:17:00.981 17:38:32 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:00.981 17:38:32 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@487 -- # '[' 10ce1a84-858f-49e1-bd9b-62e61299cb58 '!=' 10ce1a84-858f-49e1-bd9b-62e61299cb58 ']' 00:17:00.981 17:38:32 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@491 -- # has_redundancy raid1 00:17:00.981 17:38:32 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@198 -- # case $1 in 00:17:00.981 17:38:32 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@199 -- # return 0 00:17:00.981 17:38:32 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@493 -- # rpc_cmd bdev_passthru_delete pt1 00:17:00.981 17:38:32 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:00.981 17:38:32 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:00.981 [2024-11-27 17:38:32.075135] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: pt1 00:17:00.981 17:38:32 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:00.981 17:38:32 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@496 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:17:00.981 17:38:32 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:17:00.981 17:38:32 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:17:00.981 17:38:32 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:00.981 17:38:32 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:00.981 17:38:32 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:17:00.981 17:38:32 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:00.981 17:38:32 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:00.981 17:38:32 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:00.981 17:38:32 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:00.981 17:38:32 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:00.981 17:38:32 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:00.981 17:38:32 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:00.981 17:38:32 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:00.981 17:38:32 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:00.981 17:38:32 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:00.981 "name": "raid_bdev1", 00:17:00.981 "uuid": "10ce1a84-858f-49e1-bd9b-62e61299cb58", 00:17:00.981 "strip_size_kb": 0, 00:17:00.981 "state": "online", 00:17:00.981 "raid_level": "raid1", 00:17:00.981 "superblock": true, 00:17:00.981 "num_base_bdevs": 2, 00:17:00.981 "num_base_bdevs_discovered": 1, 00:17:00.981 "num_base_bdevs_operational": 1, 00:17:00.981 "base_bdevs_list": [ 00:17:00.981 { 00:17:00.981 "name": null, 00:17:00.981 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:00.981 "is_configured": false, 00:17:00.981 "data_offset": 0, 00:17:00.981 "data_size": 7936 00:17:00.981 }, 00:17:00.981 { 00:17:00.981 "name": "pt2", 00:17:00.981 "uuid": "00000000-0000-0000-0000-000000000002", 00:17:00.981 "is_configured": true, 00:17:00.981 "data_offset": 256, 00:17:00.981 "data_size": 7936 00:17:00.981 } 00:17:00.982 ] 00:17:00.982 }' 00:17:00.982 17:38:32 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:00.982 17:38:32 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:01.553 17:38:32 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@499 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:17:01.553 17:38:32 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:01.553 17:38:32 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:01.553 [2024-11-27 17:38:32.506347] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:17:01.553 [2024-11-27 17:38:32.506417] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:17:01.553 [2024-11-27 17:38:32.506485] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:17:01.553 [2024-11-27 17:38:32.506540] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:17:01.553 [2024-11-27 17:38:32.506569] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001900 name raid_bdev1, state offline 00:17:01.553 17:38:32 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:01.553 17:38:32 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@500 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:01.553 17:38:32 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:01.553 17:38:32 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:01.553 17:38:32 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@500 -- # jq -r '.[]' 00:17:01.553 17:38:32 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:01.553 17:38:32 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@500 -- # raid_bdev= 00:17:01.553 17:38:32 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@501 -- # '[' -n '' ']' 00:17:01.553 17:38:32 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@506 -- # (( i = 1 )) 00:17:01.553 17:38:32 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:17:01.553 17:38:32 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@507 -- # rpc_cmd bdev_passthru_delete pt2 00:17:01.553 17:38:32 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:01.553 17:38:32 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:01.553 17:38:32 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:01.553 17:38:32 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@506 -- # (( i++ )) 00:17:01.553 17:38:32 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@506 -- # (( i < num_base_bdevs )) 00:17:01.553 17:38:32 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@511 -- # (( i = 1 )) 00:17:01.553 17:38:32 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@511 -- # (( i < num_base_bdevs - 1 )) 00:17:01.553 17:38:32 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@519 -- # i=1 00:17:01.553 17:38:32 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@520 -- # rpc_cmd bdev_passthru_create -b malloc2 -p pt2 -u 00000000-0000-0000-0000-000000000002 00:17:01.553 17:38:32 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:01.553 17:38:32 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:01.553 [2024-11-27 17:38:32.582224] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc2 00:17:01.553 [2024-11-27 17:38:32.582277] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:17:01.553 [2024-11-27 17:38:32.582296] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008780 00:17:01.553 [2024-11-27 17:38:32.582304] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:17:01.553 [2024-11-27 17:38:32.584078] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:17:01.553 [2024-11-27 17:38:32.584118] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt2 00:17:01.553 [2024-11-27 17:38:32.584172] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt2 00:17:01.553 [2024-11-27 17:38:32.584201] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:17:01.553 [2024-11-27 17:38:32.584253] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001c80 00:17:01.553 [2024-11-27 17:38:32.584262] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4128 00:17:01.553 [2024-11-27 17:38:32.584335] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:17:01.553 [2024-11-27 17:38:32.584401] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001c80 00:17:01.553 [2024-11-27 17:38:32.584410] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001c80 00:17:01.553 [2024-11-27 17:38:32.584459] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:17:01.553 pt2 00:17:01.553 17:38:32 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:01.553 17:38:32 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@523 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:17:01.553 17:38:32 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:17:01.553 17:38:32 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:17:01.553 17:38:32 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:01.553 17:38:32 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:01.553 17:38:32 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:17:01.553 17:38:32 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:01.553 17:38:32 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:01.553 17:38:32 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:01.553 17:38:32 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:01.553 17:38:32 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:01.553 17:38:32 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:01.553 17:38:32 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:01.553 17:38:32 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:01.553 17:38:32 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:01.553 17:38:32 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:01.553 "name": "raid_bdev1", 00:17:01.553 "uuid": "10ce1a84-858f-49e1-bd9b-62e61299cb58", 00:17:01.553 "strip_size_kb": 0, 00:17:01.553 "state": "online", 00:17:01.554 "raid_level": "raid1", 00:17:01.554 "superblock": true, 00:17:01.554 "num_base_bdevs": 2, 00:17:01.554 "num_base_bdevs_discovered": 1, 00:17:01.554 "num_base_bdevs_operational": 1, 00:17:01.554 "base_bdevs_list": [ 00:17:01.554 { 00:17:01.554 "name": null, 00:17:01.554 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:01.554 "is_configured": false, 00:17:01.554 "data_offset": 256, 00:17:01.554 "data_size": 7936 00:17:01.554 }, 00:17:01.554 { 00:17:01.554 "name": "pt2", 00:17:01.554 "uuid": "00000000-0000-0000-0000-000000000002", 00:17:01.554 "is_configured": true, 00:17:01.554 "data_offset": 256, 00:17:01.554 "data_size": 7936 00:17:01.554 } 00:17:01.554 ] 00:17:01.554 }' 00:17:01.554 17:38:32 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:01.554 17:38:32 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:01.814 17:38:32 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@526 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:17:01.814 17:38:32 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:01.814 17:38:32 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:01.814 [2024-11-27 17:38:32.953581] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:17:01.814 [2024-11-27 17:38:32.953654] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:17:01.814 [2024-11-27 17:38:32.953737] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:17:01.814 [2024-11-27 17:38:32.953797] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:17:01.814 [2024-11-27 17:38:32.953829] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001c80 name raid_bdev1, state offline 00:17:01.814 17:38:32 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:01.814 17:38:32 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@527 -- # jq -r '.[]' 00:17:01.814 17:38:32 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@527 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:01.814 17:38:32 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:01.814 17:38:32 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:01.814 17:38:32 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:01.814 17:38:32 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@527 -- # raid_bdev= 00:17:01.814 17:38:32 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@528 -- # '[' -n '' ']' 00:17:01.814 17:38:32 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@532 -- # '[' 2 -gt 2 ']' 00:17:01.814 17:38:32 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@540 -- # rpc_cmd bdev_passthru_create -b malloc1 -p pt1 -u 00000000-0000-0000-0000-000000000001 00:17:01.814 17:38:32 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:01.815 17:38:32 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:01.815 [2024-11-27 17:38:32.997523] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on malloc1 00:17:01.815 [2024-11-27 17:38:32.997620] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:17:01.815 [2024-11-27 17:38:32.997654] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008d80 00:17:01.815 [2024-11-27 17:38:32.997685] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:17:01.815 [2024-11-27 17:38:32.999487] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:17:01.815 [2024-11-27 17:38:32.999555] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: pt1 00:17:01.815 [2024-11-27 17:38:32.999615] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev pt1 00:17:01.815 [2024-11-27 17:38:32.999662] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt1 is claimed 00:17:01.815 [2024-11-27 17:38:32.999766] bdev_raid.c:3675:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev pt2 (4) greater than existing raid bdev raid_bdev1 (2) 00:17:01.815 [2024-11-27 17:38:32.999816] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:17:01.815 [2024-11-27 17:38:32.999856] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002000 name raid_bdev1, state configuring 00:17:01.815 [2024-11-27 17:38:32.999918] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev pt2 is claimed 00:17:01.815 [2024-11-27 17:38:33.000001] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000002380 00:17:01.815 [2024-11-27 17:38:33.000037] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4128 00:17:01.815 [2024-11-27 17:38:33.000127] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:17:01.815 [2024-11-27 17:38:33.000226] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000002380 00:17:01.815 [2024-11-27 17:38:33.000264] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000002380 00:17:01.815 [2024-11-27 17:38:33.000359] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:17:01.815 pt1 00:17:01.815 17:38:33 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:02.075 17:38:33 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@542 -- # '[' 2 -gt 2 ']' 00:17:02.075 17:38:33 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@554 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:17:02.075 17:38:33 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:17:02.075 17:38:33 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:17:02.075 17:38:33 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:02.075 17:38:33 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:02.075 17:38:33 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:17:02.075 17:38:33 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:02.075 17:38:33 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:02.075 17:38:33 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:02.075 17:38:33 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:02.075 17:38:33 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:02.075 17:38:33 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:02.075 17:38:33 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:02.075 17:38:33 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:02.075 17:38:33 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:02.075 17:38:33 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:02.075 "name": "raid_bdev1", 00:17:02.075 "uuid": "10ce1a84-858f-49e1-bd9b-62e61299cb58", 00:17:02.075 "strip_size_kb": 0, 00:17:02.075 "state": "online", 00:17:02.075 "raid_level": "raid1", 00:17:02.075 "superblock": true, 00:17:02.075 "num_base_bdevs": 2, 00:17:02.075 "num_base_bdevs_discovered": 1, 00:17:02.075 "num_base_bdevs_operational": 1, 00:17:02.075 "base_bdevs_list": [ 00:17:02.075 { 00:17:02.075 "name": null, 00:17:02.075 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:02.075 "is_configured": false, 00:17:02.075 "data_offset": 256, 00:17:02.075 "data_size": 7936 00:17:02.075 }, 00:17:02.075 { 00:17:02.075 "name": "pt2", 00:17:02.075 "uuid": "00000000-0000-0000-0000-000000000002", 00:17:02.075 "is_configured": true, 00:17:02.075 "data_offset": 256, 00:17:02.075 "data_size": 7936 00:17:02.075 } 00:17:02.075 ] 00:17:02.075 }' 00:17:02.075 17:38:33 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:02.075 17:38:33 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:02.336 17:38:33 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@555 -- # rpc_cmd bdev_raid_get_bdevs online 00:17:02.336 17:38:33 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:02.336 17:38:33 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:02.336 17:38:33 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@555 -- # jq -r '.[].base_bdevs_list[0].is_configured' 00:17:02.336 17:38:33 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:02.336 17:38:33 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@555 -- # [[ false == \f\a\l\s\e ]] 00:17:02.336 17:38:33 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@558 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:17:02.336 17:38:33 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@558 -- # jq -r '.[] | .uuid' 00:17:02.336 17:38:33 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:02.336 17:38:33 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:02.336 [2024-11-27 17:38:33.452926] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:17:02.336 17:38:33 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:02.336 17:38:33 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@558 -- # '[' 10ce1a84-858f-49e1-bd9b-62e61299cb58 '!=' 10ce1a84-858f-49e1-bd9b-62e61299cb58 ']' 00:17:02.336 17:38:33 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@563 -- # killprocess 98893 00:17:02.336 17:38:33 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@950 -- # '[' -z 98893 ']' 00:17:02.336 17:38:33 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@954 -- # kill -0 98893 00:17:02.336 17:38:33 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@955 -- # uname 00:17:02.336 17:38:33 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:17:02.336 17:38:33 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 98893 00:17:02.336 17:38:33 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:17:02.336 17:38:33 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:17:02.336 17:38:33 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@968 -- # echo 'killing process with pid 98893' 00:17:02.336 killing process with pid 98893 00:17:02.336 17:38:33 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@969 -- # kill 98893 00:17:02.336 [2024-11-27 17:38:33.525007] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:17:02.336 [2024-11-27 17:38:33.525066] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:17:02.336 [2024-11-27 17:38:33.525103] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:17:02.336 [2024-11-27 17:38:33.525111] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000002380 name raid_bdev1, state offline 00:17:02.336 17:38:33 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@974 -- # wait 98893 00:17:02.597 [2024-11-27 17:38:33.548306] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:17:02.597 17:38:33 bdev_raid.raid_superblock_test_md_interleaved -- bdev/bdev_raid.sh@565 -- # return 0 00:17:02.597 00:17:02.597 real 0m4.852s 00:17:02.597 user 0m7.889s 00:17:02.597 sys 0m1.045s 00:17:02.597 17:38:33 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@1126 -- # xtrace_disable 00:17:02.597 ************************************ 00:17:02.597 END TEST raid_superblock_test_md_interleaved 00:17:02.597 ************************************ 00:17:02.597 17:38:33 bdev_raid.raid_superblock_test_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:02.857 17:38:33 bdev_raid -- bdev/bdev_raid.sh@1013 -- # run_test raid_rebuild_test_sb_md_interleaved raid_rebuild_test raid1 2 true false false 00:17:02.857 17:38:33 bdev_raid -- common/autotest_common.sh@1101 -- # '[' 7 -le 1 ']' 00:17:02.857 17:38:33 bdev_raid -- common/autotest_common.sh@1107 -- # xtrace_disable 00:17:02.857 17:38:33 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:17:02.857 ************************************ 00:17:02.857 START TEST raid_rebuild_test_sb_md_interleaved 00:17:02.857 ************************************ 00:17:02.857 17:38:33 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@1125 -- # raid_rebuild_test raid1 2 true false false 00:17:02.857 17:38:33 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@569 -- # local raid_level=raid1 00:17:02.857 17:38:33 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@570 -- # local num_base_bdevs=2 00:17:02.857 17:38:33 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@571 -- # local superblock=true 00:17:02.857 17:38:33 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@572 -- # local background_io=false 00:17:02.857 17:38:33 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@573 -- # local verify=false 00:17:02.857 17:38:33 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@574 -- # (( i = 1 )) 00:17:02.857 17:38:33 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:17:02.857 17:38:33 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@576 -- # echo BaseBdev1 00:17:02.857 17:38:33 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:17:02.857 17:38:33 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:17:02.857 17:38:33 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@576 -- # echo BaseBdev2 00:17:02.857 17:38:33 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@574 -- # (( i++ )) 00:17:02.857 17:38:33 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@574 -- # (( i <= num_base_bdevs )) 00:17:02.857 17:38:33 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@574 -- # base_bdevs=('BaseBdev1' 'BaseBdev2') 00:17:02.857 17:38:33 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@574 -- # local base_bdevs 00:17:02.858 17:38:33 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@575 -- # local raid_bdev_name=raid_bdev1 00:17:02.858 17:38:33 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@576 -- # local strip_size 00:17:02.858 17:38:33 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@577 -- # local create_arg 00:17:02.858 17:38:33 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@578 -- # local raid_bdev_size 00:17:02.858 17:38:33 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@579 -- # local data_offset 00:17:02.858 17:38:33 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@581 -- # '[' raid1 '!=' raid1 ']' 00:17:02.858 17:38:33 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@589 -- # strip_size=0 00:17:02.858 17:38:33 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@592 -- # '[' true = true ']' 00:17:02.858 17:38:33 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@593 -- # create_arg+=' -s' 00:17:02.858 17:38:33 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@597 -- # raid_pid=99211 00:17:02.858 17:38:33 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@596 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf -T raid_bdev1 -t 60 -w randrw -M 50 -o 3M -q 2 -U -z -L bdev_raid 00:17:02.858 17:38:33 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@598 -- # waitforlisten 99211 00:17:02.858 17:38:33 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@831 -- # '[' -z 99211 ']' 00:17:02.858 17:38:33 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:17:02.858 17:38:33 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@836 -- # local max_retries=100 00:17:02.858 17:38:33 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:17:02.858 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:17:02.858 17:38:33 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@840 -- # xtrace_disable 00:17:02.858 17:38:33 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:02.858 I/O size of 3145728 is greater than zero copy threshold (65536). 00:17:02.858 Zero copy mechanism will not be used. 00:17:02.858 [2024-11-27 17:38:33.963047] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:17:02.858 [2024-11-27 17:38:33.963214] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid99211 ] 00:17:03.118 [2024-11-27 17:38:34.109601] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:17:03.118 [2024-11-27 17:38:34.155273] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:17:03.118 [2024-11-27 17:38:34.198287] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:17:03.118 [2024-11-27 17:38:34.198401] bdev_raid.c:1452:raid_bdev_get_ctx_size: *DEBUG*: raid_bdev_get_ctx_size 00:17:03.689 17:38:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:17:03.689 17:38:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@864 -- # return 0 00:17:03.689 17:38:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:17:03.689 17:38:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -i -b BaseBdev1_malloc 00:17:03.689 17:38:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:03.689 17:38:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:03.689 BaseBdev1_malloc 00:17:03.689 17:38:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:03.689 17:38:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:17:03.689 17:38:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:03.689 17:38:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:03.689 [2024-11-27 17:38:34.808368] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:17:03.689 [2024-11-27 17:38:34.808433] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:17:03.689 [2024-11-27 17:38:34.808462] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000006680 00:17:03.689 [2024-11-27 17:38:34.808478] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:17:03.689 [2024-11-27 17:38:34.810285] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:17:03.689 [2024-11-27 17:38:34.810325] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:17:03.689 BaseBdev1 00:17:03.689 17:38:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:03.689 17:38:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@601 -- # for bdev in "${base_bdevs[@]}" 00:17:03.689 17:38:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@602 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -i -b BaseBdev2_malloc 00:17:03.689 17:38:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:03.689 17:38:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:03.689 BaseBdev2_malloc 00:17:03.689 17:38:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:03.690 17:38:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@603 -- # rpc_cmd bdev_passthru_create -b BaseBdev2_malloc -p BaseBdev2 00:17:03.690 17:38:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:03.690 17:38:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:03.690 [2024-11-27 17:38:34.853316] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev2_malloc 00:17:03.690 [2024-11-27 17:38:34.853423] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:17:03.690 [2024-11-27 17:38:34.853473] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000007280 00:17:03.690 [2024-11-27 17:38:34.853497] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:17:03.690 [2024-11-27 17:38:34.857331] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:17:03.690 [2024-11-27 17:38:34.857384] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev2 00:17:03.690 BaseBdev2 00:17:03.690 17:38:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:03.690 17:38:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@607 -- # rpc_cmd bdev_malloc_create 32 4096 -m 32 -i -b spare_malloc 00:17:03.690 17:38:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:03.690 17:38:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:03.690 spare_malloc 00:17:03.690 17:38:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:03.690 17:38:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@608 -- # rpc_cmd bdev_delay_create -b spare_malloc -d spare_delay -r 0 -t 0 -w 100000 -n 100000 00:17:03.690 17:38:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:03.690 17:38:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:03.950 spare_delay 00:17:03.950 17:38:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:03.950 17:38:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@609 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:17:03.950 17:38:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:03.950 17:38:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:03.950 [2024-11-27 17:38:34.895078] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:17:03.950 [2024-11-27 17:38:34.895214] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:17:03.950 [2024-11-27 17:38:34.895243] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000008480 00:17:03.950 [2024-11-27 17:38:34.895252] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:17:03.950 [2024-11-27 17:38:34.897076] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:17:03.950 [2024-11-27 17:38:34.897125] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:17:03.950 spare 00:17:03.950 17:38:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:03.950 17:38:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@612 -- # rpc_cmd bdev_raid_create -s -r raid1 -b ''\''BaseBdev1 BaseBdev2'\''' -n raid_bdev1 00:17:03.950 17:38:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:03.950 17:38:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:03.950 [2024-11-27 17:38:34.907126] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:17:03.950 [2024-11-27 17:38:34.908990] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:17:03.950 [2024-11-27 17:38:34.909158] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001200 00:17:03.950 [2024-11-27 17:38:34.909176] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4128 00:17:03.950 [2024-11-27 17:38:34.909249] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002460 00:17:03.950 [2024-11-27 17:38:34.909317] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001200 00:17:03.950 [2024-11-27 17:38:34.909327] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001200 00:17:03.950 [2024-11-27 17:38:34.909393] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:17:03.950 17:38:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:03.950 17:38:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@613 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:17:03.950 17:38:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:17:03.950 17:38:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:17:03.950 17:38:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:03.950 17:38:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:03.951 17:38:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:17:03.951 17:38:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:03.951 17:38:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:03.951 17:38:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:03.951 17:38:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:03.951 17:38:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:03.951 17:38:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:03.951 17:38:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:03.951 17:38:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:03.951 17:38:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:03.951 17:38:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:03.951 "name": "raid_bdev1", 00:17:03.951 "uuid": "ce088f1b-e78a-4f92-83c0-f1bc98379766", 00:17:03.951 "strip_size_kb": 0, 00:17:03.951 "state": "online", 00:17:03.951 "raid_level": "raid1", 00:17:03.951 "superblock": true, 00:17:03.951 "num_base_bdevs": 2, 00:17:03.951 "num_base_bdevs_discovered": 2, 00:17:03.951 "num_base_bdevs_operational": 2, 00:17:03.951 "base_bdevs_list": [ 00:17:03.951 { 00:17:03.951 "name": "BaseBdev1", 00:17:03.951 "uuid": "9f7e88be-06f8-5984-a096-8cd17448aac9", 00:17:03.951 "is_configured": true, 00:17:03.951 "data_offset": 256, 00:17:03.951 "data_size": 7936 00:17:03.951 }, 00:17:03.951 { 00:17:03.951 "name": "BaseBdev2", 00:17:03.951 "uuid": "84f144e7-7f26-5250-8bec-51172e731063", 00:17:03.951 "is_configured": true, 00:17:03.951 "data_offset": 256, 00:17:03.951 "data_size": 7936 00:17:03.951 } 00:17:03.951 ] 00:17:03.951 }' 00:17:03.951 17:38:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:03.951 17:38:34 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:04.211 17:38:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@616 -- # rpc_cmd bdev_get_bdevs -b raid_bdev1 00:17:04.211 17:38:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:04.211 17:38:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:04.211 17:38:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@616 -- # jq -r '.[].num_blocks' 00:17:04.211 [2024-11-27 17:38:35.342527] bdev_raid.c:1129:raid_bdev_dump_info_json: *DEBUG*: raid_bdev_dump_config_json 00:17:04.211 17:38:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:04.211 17:38:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@616 -- # raid_bdev_size=7936 00:17:04.211 17:38:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@619 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:04.211 17:38:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@619 -- # jq -r '.[].base_bdevs_list[0].data_offset' 00:17:04.211 17:38:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:04.211 17:38:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:04.472 17:38:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:04.472 17:38:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@619 -- # data_offset=256 00:17:04.472 17:38:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@621 -- # '[' false = true ']' 00:17:04.472 17:38:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@624 -- # '[' false = true ']' 00:17:04.472 17:38:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@640 -- # rpc_cmd bdev_raid_remove_base_bdev BaseBdev1 00:17:04.472 17:38:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:04.472 17:38:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:04.472 [2024-11-27 17:38:35.442131] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: BaseBdev1 00:17:04.472 17:38:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:04.472 17:38:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@643 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:17:04.472 17:38:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:17:04.472 17:38:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:17:04.472 17:38:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:04.472 17:38:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:04.472 17:38:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:17:04.472 17:38:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:04.472 17:38:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:04.472 17:38:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:04.472 17:38:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:04.472 17:38:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:04.472 17:38:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:04.472 17:38:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:04.472 17:38:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:04.472 17:38:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:04.472 17:38:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:04.472 "name": "raid_bdev1", 00:17:04.472 "uuid": "ce088f1b-e78a-4f92-83c0-f1bc98379766", 00:17:04.472 "strip_size_kb": 0, 00:17:04.472 "state": "online", 00:17:04.472 "raid_level": "raid1", 00:17:04.472 "superblock": true, 00:17:04.472 "num_base_bdevs": 2, 00:17:04.472 "num_base_bdevs_discovered": 1, 00:17:04.472 "num_base_bdevs_operational": 1, 00:17:04.472 "base_bdevs_list": [ 00:17:04.472 { 00:17:04.472 "name": null, 00:17:04.472 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:04.472 "is_configured": false, 00:17:04.472 "data_offset": 0, 00:17:04.472 "data_size": 7936 00:17:04.472 }, 00:17:04.472 { 00:17:04.472 "name": "BaseBdev2", 00:17:04.472 "uuid": "84f144e7-7f26-5250-8bec-51172e731063", 00:17:04.472 "is_configured": true, 00:17:04.472 "data_offset": 256, 00:17:04.472 "data_size": 7936 00:17:04.472 } 00:17:04.472 ] 00:17:04.472 }' 00:17:04.472 17:38:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:04.472 17:38:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:04.732 17:38:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@646 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:17:04.732 17:38:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:04.732 17:38:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:04.732 [2024-11-27 17:38:35.901349] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:17:04.732 [2024-11-27 17:38:35.904270] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002530 00:17:04.732 [2024-11-27 17:38:35.906029] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:17:04.732 17:38:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:04.732 17:38:35 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@647 -- # sleep 1 00:17:06.115 17:38:36 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@650 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:17:06.115 17:38:36 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:17:06.115 17:38:36 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:17:06.115 17:38:36 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=spare 00:17:06.115 17:38:36 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:17:06.115 17:38:36 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:06.115 17:38:36 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:06.115 17:38:36 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:06.115 17:38:36 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:06.115 17:38:36 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:06.115 17:38:36 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:17:06.115 "name": "raid_bdev1", 00:17:06.115 "uuid": "ce088f1b-e78a-4f92-83c0-f1bc98379766", 00:17:06.115 "strip_size_kb": 0, 00:17:06.115 "state": "online", 00:17:06.115 "raid_level": "raid1", 00:17:06.115 "superblock": true, 00:17:06.115 "num_base_bdevs": 2, 00:17:06.115 "num_base_bdevs_discovered": 2, 00:17:06.115 "num_base_bdevs_operational": 2, 00:17:06.115 "process": { 00:17:06.115 "type": "rebuild", 00:17:06.115 "target": "spare", 00:17:06.115 "progress": { 00:17:06.115 "blocks": 2560, 00:17:06.115 "percent": 32 00:17:06.115 } 00:17:06.115 }, 00:17:06.115 "base_bdevs_list": [ 00:17:06.115 { 00:17:06.115 "name": "spare", 00:17:06.115 "uuid": "1ef6695d-3b3a-53fd-a2c8-512ff743a86e", 00:17:06.115 "is_configured": true, 00:17:06.115 "data_offset": 256, 00:17:06.115 "data_size": 7936 00:17:06.115 }, 00:17:06.115 { 00:17:06.115 "name": "BaseBdev2", 00:17:06.115 "uuid": "84f144e7-7f26-5250-8bec-51172e731063", 00:17:06.115 "is_configured": true, 00:17:06.115 "data_offset": 256, 00:17:06.115 "data_size": 7936 00:17:06.115 } 00:17:06.115 ] 00:17:06.115 }' 00:17:06.115 17:38:36 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:17:06.115 17:38:36 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:17:06.115 17:38:36 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:17:06.115 17:38:37 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:17:06.115 17:38:37 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@653 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:17:06.115 17:38:37 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:06.115 17:38:37 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:06.115 [2024-11-27 17:38:37.049242] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:17:06.115 [2024-11-27 17:38:37.110634] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:17:06.115 [2024-11-27 17:38:37.110703] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:17:06.115 [2024-11-27 17:38:37.110720] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:17:06.115 [2024-11-27 17:38:37.110727] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:17:06.115 17:38:37 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:06.115 17:38:37 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@656 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:17:06.115 17:38:37 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:17:06.115 17:38:37 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:17:06.115 17:38:37 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:06.115 17:38:37 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:06.116 17:38:37 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:17:06.116 17:38:37 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:06.116 17:38:37 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:06.116 17:38:37 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:06.116 17:38:37 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:06.116 17:38:37 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:06.116 17:38:37 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:06.116 17:38:37 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:06.116 17:38:37 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:06.116 17:38:37 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:06.116 17:38:37 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:06.116 "name": "raid_bdev1", 00:17:06.116 "uuid": "ce088f1b-e78a-4f92-83c0-f1bc98379766", 00:17:06.116 "strip_size_kb": 0, 00:17:06.116 "state": "online", 00:17:06.116 "raid_level": "raid1", 00:17:06.116 "superblock": true, 00:17:06.116 "num_base_bdevs": 2, 00:17:06.116 "num_base_bdevs_discovered": 1, 00:17:06.116 "num_base_bdevs_operational": 1, 00:17:06.116 "base_bdevs_list": [ 00:17:06.116 { 00:17:06.116 "name": null, 00:17:06.116 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:06.116 "is_configured": false, 00:17:06.116 "data_offset": 0, 00:17:06.116 "data_size": 7936 00:17:06.116 }, 00:17:06.116 { 00:17:06.116 "name": "BaseBdev2", 00:17:06.116 "uuid": "84f144e7-7f26-5250-8bec-51172e731063", 00:17:06.116 "is_configured": true, 00:17:06.116 "data_offset": 256, 00:17:06.116 "data_size": 7936 00:17:06.116 } 00:17:06.116 ] 00:17:06.116 }' 00:17:06.116 17:38:37 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:06.116 17:38:37 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:06.686 17:38:37 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@659 -- # verify_raid_bdev_process raid_bdev1 none none 00:17:06.686 17:38:37 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:17:06.686 17:38:37 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:17:06.686 17:38:37 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=none 00:17:06.686 17:38:37 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:17:06.686 17:38:37 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:06.686 17:38:37 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:06.686 17:38:37 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:06.686 17:38:37 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:06.686 17:38:37 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:06.686 17:38:37 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:17:06.686 "name": "raid_bdev1", 00:17:06.687 "uuid": "ce088f1b-e78a-4f92-83c0-f1bc98379766", 00:17:06.687 "strip_size_kb": 0, 00:17:06.687 "state": "online", 00:17:06.687 "raid_level": "raid1", 00:17:06.687 "superblock": true, 00:17:06.687 "num_base_bdevs": 2, 00:17:06.687 "num_base_bdevs_discovered": 1, 00:17:06.687 "num_base_bdevs_operational": 1, 00:17:06.687 "base_bdevs_list": [ 00:17:06.687 { 00:17:06.687 "name": null, 00:17:06.687 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:06.687 "is_configured": false, 00:17:06.687 "data_offset": 0, 00:17:06.687 "data_size": 7936 00:17:06.687 }, 00:17:06.687 { 00:17:06.687 "name": "BaseBdev2", 00:17:06.687 "uuid": "84f144e7-7f26-5250-8bec-51172e731063", 00:17:06.687 "is_configured": true, 00:17:06.687 "data_offset": 256, 00:17:06.687 "data_size": 7936 00:17:06.687 } 00:17:06.687 ] 00:17:06.687 }' 00:17:06.687 17:38:37 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:17:06.687 17:38:37 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:17:06.687 17:38:37 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:17:06.687 17:38:37 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:17:06.687 17:38:37 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@662 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:17:06.687 17:38:37 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:06.687 17:38:37 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:06.687 [2024-11-27 17:38:37.701269] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:17:06.687 [2024-11-27 17:38:37.703956] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002600 00:17:06.687 [2024-11-27 17:38:37.705660] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:17:06.687 17:38:37 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:06.687 17:38:37 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@663 -- # sleep 1 00:17:07.627 17:38:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@664 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:17:07.627 17:38:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:17:07.627 17:38:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:17:07.627 17:38:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=spare 00:17:07.627 17:38:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:17:07.627 17:38:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:07.627 17:38:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:07.627 17:38:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:07.627 17:38:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:07.627 17:38:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:07.627 17:38:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:17:07.627 "name": "raid_bdev1", 00:17:07.627 "uuid": "ce088f1b-e78a-4f92-83c0-f1bc98379766", 00:17:07.627 "strip_size_kb": 0, 00:17:07.627 "state": "online", 00:17:07.627 "raid_level": "raid1", 00:17:07.627 "superblock": true, 00:17:07.627 "num_base_bdevs": 2, 00:17:07.627 "num_base_bdevs_discovered": 2, 00:17:07.627 "num_base_bdevs_operational": 2, 00:17:07.627 "process": { 00:17:07.627 "type": "rebuild", 00:17:07.627 "target": "spare", 00:17:07.627 "progress": { 00:17:07.627 "blocks": 2560, 00:17:07.627 "percent": 32 00:17:07.627 } 00:17:07.627 }, 00:17:07.627 "base_bdevs_list": [ 00:17:07.627 { 00:17:07.627 "name": "spare", 00:17:07.627 "uuid": "1ef6695d-3b3a-53fd-a2c8-512ff743a86e", 00:17:07.627 "is_configured": true, 00:17:07.627 "data_offset": 256, 00:17:07.627 "data_size": 7936 00:17:07.627 }, 00:17:07.627 { 00:17:07.627 "name": "BaseBdev2", 00:17:07.627 "uuid": "84f144e7-7f26-5250-8bec-51172e731063", 00:17:07.627 "is_configured": true, 00:17:07.627 "data_offset": 256, 00:17:07.627 "data_size": 7936 00:17:07.627 } 00:17:07.627 ] 00:17:07.627 }' 00:17:07.627 17:38:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:17:07.627 17:38:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:17:07.627 17:38:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:17:07.887 17:38:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:17:07.887 17:38:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@666 -- # '[' true = true ']' 00:17:07.887 17:38:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@666 -- # '[' = false ']' 00:17:07.887 /home/vagrant/spdk_repo/spdk/test/bdev/bdev_raid.sh: line 666: [: =: unary operator expected 00:17:07.887 17:38:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@691 -- # local num_base_bdevs_operational=2 00:17:07.887 17:38:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@693 -- # '[' raid1 = raid1 ']' 00:17:07.887 17:38:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@693 -- # '[' 2 -gt 2 ']' 00:17:07.887 17:38:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@706 -- # local timeout=626 00:17:07.887 17:38:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:17:07.887 17:38:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:17:07.887 17:38:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:17:07.887 17:38:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:17:07.887 17:38:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=spare 00:17:07.887 17:38:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:17:07.887 17:38:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:07.887 17:38:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:07.887 17:38:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:07.887 17:38:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:07.888 17:38:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:07.888 17:38:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:17:07.888 "name": "raid_bdev1", 00:17:07.888 "uuid": "ce088f1b-e78a-4f92-83c0-f1bc98379766", 00:17:07.888 "strip_size_kb": 0, 00:17:07.888 "state": "online", 00:17:07.888 "raid_level": "raid1", 00:17:07.888 "superblock": true, 00:17:07.888 "num_base_bdevs": 2, 00:17:07.888 "num_base_bdevs_discovered": 2, 00:17:07.888 "num_base_bdevs_operational": 2, 00:17:07.888 "process": { 00:17:07.888 "type": "rebuild", 00:17:07.888 "target": "spare", 00:17:07.888 "progress": { 00:17:07.888 "blocks": 2816, 00:17:07.888 "percent": 35 00:17:07.888 } 00:17:07.888 }, 00:17:07.888 "base_bdevs_list": [ 00:17:07.888 { 00:17:07.888 "name": "spare", 00:17:07.888 "uuid": "1ef6695d-3b3a-53fd-a2c8-512ff743a86e", 00:17:07.888 "is_configured": true, 00:17:07.888 "data_offset": 256, 00:17:07.888 "data_size": 7936 00:17:07.888 }, 00:17:07.888 { 00:17:07.888 "name": "BaseBdev2", 00:17:07.888 "uuid": "84f144e7-7f26-5250-8bec-51172e731063", 00:17:07.888 "is_configured": true, 00:17:07.888 "data_offset": 256, 00:17:07.888 "data_size": 7936 00:17:07.888 } 00:17:07.888 ] 00:17:07.888 }' 00:17:07.888 17:38:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:17:07.888 17:38:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:17:07.888 17:38:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:17:07.888 17:38:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:17:07.888 17:38:38 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@711 -- # sleep 1 00:17:08.827 17:38:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:17:08.827 17:38:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:17:08.827 17:38:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:17:08.827 17:38:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:17:08.827 17:38:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=spare 00:17:08.827 17:38:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:17:08.827 17:38:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:08.827 17:38:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:08.827 17:38:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:08.827 17:38:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:08.827 17:38:39 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:08.827 17:38:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:17:08.827 "name": "raid_bdev1", 00:17:08.827 "uuid": "ce088f1b-e78a-4f92-83c0-f1bc98379766", 00:17:08.827 "strip_size_kb": 0, 00:17:08.827 "state": "online", 00:17:08.827 "raid_level": "raid1", 00:17:08.827 "superblock": true, 00:17:08.827 "num_base_bdevs": 2, 00:17:08.827 "num_base_bdevs_discovered": 2, 00:17:08.827 "num_base_bdevs_operational": 2, 00:17:08.827 "process": { 00:17:08.827 "type": "rebuild", 00:17:08.827 "target": "spare", 00:17:08.827 "progress": { 00:17:08.827 "blocks": 5632, 00:17:08.827 "percent": 70 00:17:08.827 } 00:17:08.827 }, 00:17:08.827 "base_bdevs_list": [ 00:17:08.827 { 00:17:08.827 "name": "spare", 00:17:08.827 "uuid": "1ef6695d-3b3a-53fd-a2c8-512ff743a86e", 00:17:08.827 "is_configured": true, 00:17:08.827 "data_offset": 256, 00:17:08.827 "data_size": 7936 00:17:08.827 }, 00:17:08.827 { 00:17:08.827 "name": "BaseBdev2", 00:17:08.827 "uuid": "84f144e7-7f26-5250-8bec-51172e731063", 00:17:08.827 "is_configured": true, 00:17:08.827 "data_offset": 256, 00:17:08.827 "data_size": 7936 00:17:08.827 } 00:17:08.827 ] 00:17:08.827 }' 00:17:08.827 17:38:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:17:09.088 17:38:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:17:09.088 17:38:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:17:09.088 17:38:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:17:09.088 17:38:40 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@711 -- # sleep 1 00:17:09.658 [2024-11-27 17:38:40.815844] bdev_raid.c:2896:raid_bdev_process_thread_run: *DEBUG*: process completed on raid_bdev1 00:17:09.658 [2024-11-27 17:38:40.815911] bdev_raid.c:2558:raid_bdev_process_finish_done: *NOTICE*: Finished rebuild on raid bdev raid_bdev1 00:17:09.658 [2024-11-27 17:38:40.816011] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:17:10.228 17:38:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@707 -- # (( SECONDS < timeout )) 00:17:10.228 17:38:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@708 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:17:10.228 17:38:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:17:10.228 17:38:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:17:10.228 17:38:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=spare 00:17:10.228 17:38:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:17:10.229 17:38:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:10.229 17:38:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:10.229 17:38:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:10.229 17:38:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:10.229 17:38:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:10.229 17:38:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:17:10.229 "name": "raid_bdev1", 00:17:10.229 "uuid": "ce088f1b-e78a-4f92-83c0-f1bc98379766", 00:17:10.229 "strip_size_kb": 0, 00:17:10.229 "state": "online", 00:17:10.229 "raid_level": "raid1", 00:17:10.229 "superblock": true, 00:17:10.229 "num_base_bdevs": 2, 00:17:10.229 "num_base_bdevs_discovered": 2, 00:17:10.229 "num_base_bdevs_operational": 2, 00:17:10.229 "base_bdevs_list": [ 00:17:10.229 { 00:17:10.229 "name": "spare", 00:17:10.229 "uuid": "1ef6695d-3b3a-53fd-a2c8-512ff743a86e", 00:17:10.229 "is_configured": true, 00:17:10.229 "data_offset": 256, 00:17:10.229 "data_size": 7936 00:17:10.229 }, 00:17:10.229 { 00:17:10.229 "name": "BaseBdev2", 00:17:10.229 "uuid": "84f144e7-7f26-5250-8bec-51172e731063", 00:17:10.229 "is_configured": true, 00:17:10.229 "data_offset": 256, 00:17:10.229 "data_size": 7936 00:17:10.229 } 00:17:10.229 ] 00:17:10.229 }' 00:17:10.229 17:38:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:17:10.229 17:38:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ none == \r\e\b\u\i\l\d ]] 00:17:10.229 17:38:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:17:10.229 17:38:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ none == \s\p\a\r\e ]] 00:17:10.229 17:38:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@709 -- # break 00:17:10.229 17:38:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@715 -- # verify_raid_bdev_process raid_bdev1 none none 00:17:10.229 17:38:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:17:10.229 17:38:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:17:10.229 17:38:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=none 00:17:10.229 17:38:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:17:10.229 17:38:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:10.229 17:38:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:10.229 17:38:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:10.229 17:38:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:10.229 17:38:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:10.229 17:38:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:17:10.229 "name": "raid_bdev1", 00:17:10.229 "uuid": "ce088f1b-e78a-4f92-83c0-f1bc98379766", 00:17:10.229 "strip_size_kb": 0, 00:17:10.229 "state": "online", 00:17:10.229 "raid_level": "raid1", 00:17:10.229 "superblock": true, 00:17:10.229 "num_base_bdevs": 2, 00:17:10.229 "num_base_bdevs_discovered": 2, 00:17:10.229 "num_base_bdevs_operational": 2, 00:17:10.229 "base_bdevs_list": [ 00:17:10.229 { 00:17:10.229 "name": "spare", 00:17:10.229 "uuid": "1ef6695d-3b3a-53fd-a2c8-512ff743a86e", 00:17:10.229 "is_configured": true, 00:17:10.229 "data_offset": 256, 00:17:10.229 "data_size": 7936 00:17:10.229 }, 00:17:10.229 { 00:17:10.229 "name": "BaseBdev2", 00:17:10.229 "uuid": "84f144e7-7f26-5250-8bec-51172e731063", 00:17:10.229 "is_configured": true, 00:17:10.229 "data_offset": 256, 00:17:10.229 "data_size": 7936 00:17:10.229 } 00:17:10.229 ] 00:17:10.229 }' 00:17:10.229 17:38:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:17:10.229 17:38:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:17:10.229 17:38:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:17:10.229 17:38:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:17:10.229 17:38:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@716 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:17:10.229 17:38:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:17:10.229 17:38:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:17:10.229 17:38:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:10.229 17:38:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:10.229 17:38:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:17:10.229 17:38:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:10.229 17:38:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:10.229 17:38:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:10.229 17:38:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:10.229 17:38:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:10.229 17:38:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:10.229 17:38:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:10.229 17:38:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:10.229 17:38:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:10.489 17:38:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:10.489 "name": "raid_bdev1", 00:17:10.489 "uuid": "ce088f1b-e78a-4f92-83c0-f1bc98379766", 00:17:10.489 "strip_size_kb": 0, 00:17:10.489 "state": "online", 00:17:10.489 "raid_level": "raid1", 00:17:10.489 "superblock": true, 00:17:10.489 "num_base_bdevs": 2, 00:17:10.489 "num_base_bdevs_discovered": 2, 00:17:10.489 "num_base_bdevs_operational": 2, 00:17:10.489 "base_bdevs_list": [ 00:17:10.489 { 00:17:10.489 "name": "spare", 00:17:10.489 "uuid": "1ef6695d-3b3a-53fd-a2c8-512ff743a86e", 00:17:10.489 "is_configured": true, 00:17:10.489 "data_offset": 256, 00:17:10.489 "data_size": 7936 00:17:10.489 }, 00:17:10.489 { 00:17:10.489 "name": "BaseBdev2", 00:17:10.489 "uuid": "84f144e7-7f26-5250-8bec-51172e731063", 00:17:10.489 "is_configured": true, 00:17:10.489 "data_offset": 256, 00:17:10.489 "data_size": 7936 00:17:10.489 } 00:17:10.489 ] 00:17:10.489 }' 00:17:10.489 17:38:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:10.489 17:38:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:10.749 17:38:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@719 -- # rpc_cmd bdev_raid_delete raid_bdev1 00:17:10.749 17:38:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:10.749 17:38:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:10.749 [2024-11-27 17:38:41.849461] bdev_raid.c:2407:raid_bdev_delete: *DEBUG*: delete raid bdev: raid_bdev1 00:17:10.749 [2024-11-27 17:38:41.849539] bdev_raid.c:1895:raid_bdev_deconfigure: *DEBUG*: raid bdev state changing from online to offline 00:17:10.749 [2024-11-27 17:38:41.849642] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:17:10.749 [2024-11-27 17:38:41.849733] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:17:10.749 [2024-11-27 17:38:41.849785] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001200 name raid_bdev1, state offline 00:17:10.749 17:38:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:10.749 17:38:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@720 -- # jq length 00:17:10.749 17:38:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@720 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:10.749 17:38:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:10.749 17:38:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:10.749 17:38:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:10.749 17:38:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@720 -- # [[ 0 == 0 ]] 00:17:10.750 17:38:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@722 -- # '[' false = true ']' 00:17:10.750 17:38:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@743 -- # '[' true = true ']' 00:17:10.750 17:38:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@745 -- # rpc_cmd bdev_passthru_delete spare 00:17:10.750 17:38:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:10.750 17:38:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:10.750 17:38:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:10.750 17:38:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@746 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:17:10.750 17:38:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:10.750 17:38:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:10.750 [2024-11-27 17:38:41.905359] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:17:10.750 [2024-11-27 17:38:41.905416] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:17:10.750 [2024-11-27 17:38:41.905438] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009380 00:17:10.750 [2024-11-27 17:38:41.905451] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:17:10.750 [2024-11-27 17:38:41.907285] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:17:10.750 [2024-11-27 17:38:41.907323] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:17:10.750 [2024-11-27 17:38:41.907373] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:17:10.750 [2024-11-27 17:38:41.907425] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:17:10.750 [2024-11-27 17:38:41.907523] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev2 is claimed 00:17:10.750 spare 00:17:10.750 17:38:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:10.750 17:38:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@747 -- # rpc_cmd bdev_wait_for_examine 00:17:10.750 17:38:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:10.750 17:38:41 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:11.010 [2024-11-27 17:38:42.007407] bdev_raid.c:1730:raid_bdev_configure_cont: *DEBUG*: io device register 0x617000001580 00:17:11.010 [2024-11-27 17:38:42.007429] bdev_raid.c:1731:raid_bdev_configure_cont: *DEBUG*: blockcnt 7936, blocklen 4128 00:17:11.010 [2024-11-27 17:38:42.007515] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000026d0 00:17:11.010 [2024-11-27 17:38:42.007595] bdev_raid.c:1760:raid_bdev_configure_cont: *DEBUG*: raid bdev generic 0x617000001580 00:17:11.010 [2024-11-27 17:38:42.007606] bdev_raid.c:1761:raid_bdev_configure_cont: *DEBUG*: raid bdev is created with name raid_bdev1, raid_bdev 0x617000001580 00:17:11.010 [2024-11-27 17:38:42.007677] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:17:11.010 17:38:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:11.011 17:38:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@749 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 2 00:17:11.011 17:38:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:17:11.011 17:38:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:17:11.011 17:38:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:11.011 17:38:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:11.011 17:38:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=2 00:17:11.011 17:38:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:11.011 17:38:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:11.011 17:38:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:11.011 17:38:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:11.011 17:38:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:11.011 17:38:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:11.011 17:38:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:11.011 17:38:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:11.011 17:38:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:11.011 17:38:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:11.011 "name": "raid_bdev1", 00:17:11.011 "uuid": "ce088f1b-e78a-4f92-83c0-f1bc98379766", 00:17:11.011 "strip_size_kb": 0, 00:17:11.011 "state": "online", 00:17:11.011 "raid_level": "raid1", 00:17:11.011 "superblock": true, 00:17:11.011 "num_base_bdevs": 2, 00:17:11.011 "num_base_bdevs_discovered": 2, 00:17:11.011 "num_base_bdevs_operational": 2, 00:17:11.011 "base_bdevs_list": [ 00:17:11.011 { 00:17:11.011 "name": "spare", 00:17:11.011 "uuid": "1ef6695d-3b3a-53fd-a2c8-512ff743a86e", 00:17:11.011 "is_configured": true, 00:17:11.011 "data_offset": 256, 00:17:11.011 "data_size": 7936 00:17:11.011 }, 00:17:11.011 { 00:17:11.011 "name": "BaseBdev2", 00:17:11.011 "uuid": "84f144e7-7f26-5250-8bec-51172e731063", 00:17:11.011 "is_configured": true, 00:17:11.011 "data_offset": 256, 00:17:11.011 "data_size": 7936 00:17:11.011 } 00:17:11.011 ] 00:17:11.011 }' 00:17:11.011 17:38:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:11.011 17:38:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:11.581 17:38:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@750 -- # verify_raid_bdev_process raid_bdev1 none none 00:17:11.581 17:38:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:17:11.581 17:38:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:17:11.581 17:38:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=none 00:17:11.581 17:38:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:17:11.581 17:38:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:11.581 17:38:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:11.581 17:38:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:11.581 17:38:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:11.581 17:38:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:11.581 17:38:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:17:11.581 "name": "raid_bdev1", 00:17:11.581 "uuid": "ce088f1b-e78a-4f92-83c0-f1bc98379766", 00:17:11.581 "strip_size_kb": 0, 00:17:11.581 "state": "online", 00:17:11.581 "raid_level": "raid1", 00:17:11.581 "superblock": true, 00:17:11.581 "num_base_bdevs": 2, 00:17:11.581 "num_base_bdevs_discovered": 2, 00:17:11.581 "num_base_bdevs_operational": 2, 00:17:11.581 "base_bdevs_list": [ 00:17:11.581 { 00:17:11.581 "name": "spare", 00:17:11.581 "uuid": "1ef6695d-3b3a-53fd-a2c8-512ff743a86e", 00:17:11.581 "is_configured": true, 00:17:11.581 "data_offset": 256, 00:17:11.581 "data_size": 7936 00:17:11.581 }, 00:17:11.581 { 00:17:11.581 "name": "BaseBdev2", 00:17:11.581 "uuid": "84f144e7-7f26-5250-8bec-51172e731063", 00:17:11.581 "is_configured": true, 00:17:11.581 "data_offset": 256, 00:17:11.581 "data_size": 7936 00:17:11.581 } 00:17:11.581 ] 00:17:11.581 }' 00:17:11.581 17:38:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:17:11.581 17:38:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:17:11.581 17:38:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:17:11.581 17:38:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:17:11.581 17:38:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@751 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:11.581 17:38:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:11.581 17:38:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:11.581 17:38:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@751 -- # jq -r '.[].base_bdevs_list[0].name' 00:17:11.581 17:38:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:11.581 17:38:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@751 -- # [[ spare == \s\p\a\r\e ]] 00:17:11.581 17:38:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@754 -- # rpc_cmd bdev_raid_remove_base_bdev spare 00:17:11.581 17:38:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:11.581 17:38:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:11.581 [2024-11-27 17:38:42.624220] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:17:11.581 17:38:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:11.581 17:38:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@755 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:17:11.581 17:38:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:17:11.581 17:38:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:17:11.581 17:38:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:11.581 17:38:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:11.581 17:38:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:17:11.581 17:38:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:11.581 17:38:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:11.581 17:38:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:11.581 17:38:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:11.581 17:38:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:11.581 17:38:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:11.581 17:38:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:11.581 17:38:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:11.581 17:38:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:11.581 17:38:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:11.581 "name": "raid_bdev1", 00:17:11.581 "uuid": "ce088f1b-e78a-4f92-83c0-f1bc98379766", 00:17:11.581 "strip_size_kb": 0, 00:17:11.581 "state": "online", 00:17:11.581 "raid_level": "raid1", 00:17:11.581 "superblock": true, 00:17:11.581 "num_base_bdevs": 2, 00:17:11.581 "num_base_bdevs_discovered": 1, 00:17:11.581 "num_base_bdevs_operational": 1, 00:17:11.581 "base_bdevs_list": [ 00:17:11.581 { 00:17:11.581 "name": null, 00:17:11.582 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:11.582 "is_configured": false, 00:17:11.582 "data_offset": 0, 00:17:11.582 "data_size": 7936 00:17:11.582 }, 00:17:11.582 { 00:17:11.582 "name": "BaseBdev2", 00:17:11.582 "uuid": "84f144e7-7f26-5250-8bec-51172e731063", 00:17:11.582 "is_configured": true, 00:17:11.582 "data_offset": 256, 00:17:11.582 "data_size": 7936 00:17:11.582 } 00:17:11.582 ] 00:17:11.582 }' 00:17:11.582 17:38:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:11.582 17:38:42 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:12.152 17:38:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@756 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 spare 00:17:12.152 17:38:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:12.152 17:38:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:12.152 [2024-11-27 17:38:43.047494] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:17:12.152 [2024-11-27 17:38:43.047714] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:17:12.152 [2024-11-27 17:38:43.047774] bdev_raid.c:3748:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:17:12.152 [2024-11-27 17:38:43.047841] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:17:12.152 [2024-11-27 17:38:43.050569] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d0000027a0 00:17:12.152 [2024-11-27 17:38:43.052391] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:17:12.152 17:38:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:12.152 17:38:43 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@757 -- # sleep 1 00:17:13.092 17:38:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@758 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:17:13.092 17:38:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:17:13.092 17:38:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:17:13.092 17:38:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=spare 00:17:13.092 17:38:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:17:13.092 17:38:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:13.092 17:38:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:13.092 17:38:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:13.092 17:38:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:13.092 17:38:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:13.092 17:38:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:17:13.092 "name": "raid_bdev1", 00:17:13.092 "uuid": "ce088f1b-e78a-4f92-83c0-f1bc98379766", 00:17:13.092 "strip_size_kb": 0, 00:17:13.092 "state": "online", 00:17:13.092 "raid_level": "raid1", 00:17:13.092 "superblock": true, 00:17:13.092 "num_base_bdevs": 2, 00:17:13.092 "num_base_bdevs_discovered": 2, 00:17:13.092 "num_base_bdevs_operational": 2, 00:17:13.092 "process": { 00:17:13.092 "type": "rebuild", 00:17:13.092 "target": "spare", 00:17:13.092 "progress": { 00:17:13.092 "blocks": 2560, 00:17:13.092 "percent": 32 00:17:13.092 } 00:17:13.092 }, 00:17:13.092 "base_bdevs_list": [ 00:17:13.092 { 00:17:13.092 "name": "spare", 00:17:13.092 "uuid": "1ef6695d-3b3a-53fd-a2c8-512ff743a86e", 00:17:13.092 "is_configured": true, 00:17:13.092 "data_offset": 256, 00:17:13.092 "data_size": 7936 00:17:13.092 }, 00:17:13.092 { 00:17:13.092 "name": "BaseBdev2", 00:17:13.092 "uuid": "84f144e7-7f26-5250-8bec-51172e731063", 00:17:13.092 "is_configured": true, 00:17:13.092 "data_offset": 256, 00:17:13.092 "data_size": 7936 00:17:13.092 } 00:17:13.092 ] 00:17:13.092 }' 00:17:13.092 17:38:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:17:13.092 17:38:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:17:13.092 17:38:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:17:13.092 17:38:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:17:13.092 17:38:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@761 -- # rpc_cmd bdev_passthru_delete spare 00:17:13.092 17:38:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:13.092 17:38:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:13.092 [2024-11-27 17:38:44.203335] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:17:13.092 [2024-11-27 17:38:44.256213] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:17:13.092 [2024-11-27 17:38:44.256263] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:17:13.092 [2024-11-27 17:38:44.256278] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:17:13.092 [2024-11-27 17:38:44.256286] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:17:13.092 17:38:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:13.092 17:38:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@762 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:17:13.092 17:38:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:17:13.092 17:38:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:17:13.092 17:38:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:13.092 17:38:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:13.092 17:38:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:17:13.092 17:38:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:13.092 17:38:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:13.092 17:38:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:13.092 17:38:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:13.092 17:38:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:13.092 17:38:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:13.092 17:38:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:13.092 17:38:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:13.352 17:38:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:13.352 17:38:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:13.352 "name": "raid_bdev1", 00:17:13.352 "uuid": "ce088f1b-e78a-4f92-83c0-f1bc98379766", 00:17:13.352 "strip_size_kb": 0, 00:17:13.352 "state": "online", 00:17:13.352 "raid_level": "raid1", 00:17:13.352 "superblock": true, 00:17:13.352 "num_base_bdevs": 2, 00:17:13.352 "num_base_bdevs_discovered": 1, 00:17:13.352 "num_base_bdevs_operational": 1, 00:17:13.352 "base_bdevs_list": [ 00:17:13.352 { 00:17:13.352 "name": null, 00:17:13.352 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:13.352 "is_configured": false, 00:17:13.352 "data_offset": 0, 00:17:13.352 "data_size": 7936 00:17:13.352 }, 00:17:13.352 { 00:17:13.352 "name": "BaseBdev2", 00:17:13.352 "uuid": "84f144e7-7f26-5250-8bec-51172e731063", 00:17:13.352 "is_configured": true, 00:17:13.352 "data_offset": 256, 00:17:13.352 "data_size": 7936 00:17:13.352 } 00:17:13.352 ] 00:17:13.352 }' 00:17:13.352 17:38:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:13.352 17:38:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:13.612 17:38:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@763 -- # rpc_cmd bdev_passthru_create -b spare_delay -p spare 00:17:13.612 17:38:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:13.612 17:38:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:13.612 [2024-11-27 17:38:44.690827] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on spare_delay 00:17:13.612 [2024-11-27 17:38:44.690924] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:17:13.612 [2024-11-27 17:38:44.690967] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x616000009c80 00:17:13.612 [2024-11-27 17:38:44.690996] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:17:13.612 [2024-11-27 17:38:44.691191] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:17:13.612 [2024-11-27 17:38:44.691248] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: spare 00:17:13.612 [2024-11-27 17:38:44.691319] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev spare 00:17:13.612 [2024-11-27 17:38:44.691357] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev spare (4) smaller than existing raid bdev raid_bdev1 (5) 00:17:13.612 [2024-11-27 17:38:44.691395] bdev_raid.c:3748:raid_bdev_examine_sb: *NOTICE*: Re-adding bdev spare to raid bdev raid_bdev1. 00:17:13.612 [2024-11-27 17:38:44.691455] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev spare is claimed 00:17:13.612 [2024-11-27 17:38:44.693738] bdev_raid.c: 265:raid_bdev_create_cb: *DEBUG*: raid_bdev_create_cb, 0x60d000002870 00:17:13.613 [2024-11-27 17:38:44.695488] bdev_raid.c:2931:raid_bdev_process_thread_init: *NOTICE*: Started rebuild on raid bdev raid_bdev1 00:17:13.613 spare 00:17:13.613 17:38:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:13.613 17:38:44 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@764 -- # sleep 1 00:17:14.552 17:38:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@765 -- # verify_raid_bdev_process raid_bdev1 rebuild spare 00:17:14.552 17:38:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:17:14.552 17:38:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=rebuild 00:17:14.552 17:38:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=spare 00:17:14.552 17:38:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:17:14.552 17:38:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:14.552 17:38:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:14.552 17:38:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:14.552 17:38:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:14.552 17:38:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:14.812 17:38:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:17:14.812 "name": "raid_bdev1", 00:17:14.812 "uuid": "ce088f1b-e78a-4f92-83c0-f1bc98379766", 00:17:14.812 "strip_size_kb": 0, 00:17:14.812 "state": "online", 00:17:14.812 "raid_level": "raid1", 00:17:14.812 "superblock": true, 00:17:14.812 "num_base_bdevs": 2, 00:17:14.812 "num_base_bdevs_discovered": 2, 00:17:14.812 "num_base_bdevs_operational": 2, 00:17:14.812 "process": { 00:17:14.812 "type": "rebuild", 00:17:14.812 "target": "spare", 00:17:14.812 "progress": { 00:17:14.812 "blocks": 2560, 00:17:14.812 "percent": 32 00:17:14.812 } 00:17:14.812 }, 00:17:14.812 "base_bdevs_list": [ 00:17:14.812 { 00:17:14.812 "name": "spare", 00:17:14.812 "uuid": "1ef6695d-3b3a-53fd-a2c8-512ff743a86e", 00:17:14.812 "is_configured": true, 00:17:14.812 "data_offset": 256, 00:17:14.812 "data_size": 7936 00:17:14.812 }, 00:17:14.812 { 00:17:14.812 "name": "BaseBdev2", 00:17:14.812 "uuid": "84f144e7-7f26-5250-8bec-51172e731063", 00:17:14.812 "is_configured": true, 00:17:14.812 "data_offset": 256, 00:17:14.812 "data_size": 7936 00:17:14.812 } 00:17:14.812 ] 00:17:14.812 }' 00:17:14.812 17:38:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:17:14.812 17:38:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ rebuild == \r\e\b\u\i\l\d ]] 00:17:14.812 17:38:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:17:14.812 17:38:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ spare == \s\p\a\r\e ]] 00:17:14.812 17:38:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@768 -- # rpc_cmd bdev_passthru_delete spare 00:17:14.812 17:38:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:14.812 17:38:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:14.812 [2024-11-27 17:38:45.858894] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:17:14.812 [2024-11-27 17:38:45.899239] bdev_raid.c:2567:raid_bdev_process_finish_done: *WARNING*: Finished rebuild on raid bdev raid_bdev1: No such device 00:17:14.812 [2024-11-27 17:38:45.899293] bdev_raid.c: 345:raid_bdev_destroy_cb: *DEBUG*: raid_bdev_destroy_cb 00:17:14.812 [2024-11-27 17:38:45.899306] bdev_raid.c:2171:_raid_bdev_remove_base_bdev: *DEBUG*: spare 00:17:14.812 [2024-11-27 17:38:45.899314] bdev_raid.c:2505:raid_bdev_process_finish_target_removed: *ERROR*: Failed to remove target bdev: No such device 00:17:14.812 17:38:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:14.812 17:38:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@769 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:17:14.812 17:38:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:17:14.812 17:38:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:17:14.812 17:38:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:14.812 17:38:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:14.812 17:38:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:17:14.812 17:38:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:14.812 17:38:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:14.812 17:38:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:14.812 17:38:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:14.812 17:38:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:14.812 17:38:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:14.812 17:38:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:14.812 17:38:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:14.812 17:38:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:14.812 17:38:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:14.812 "name": "raid_bdev1", 00:17:14.812 "uuid": "ce088f1b-e78a-4f92-83c0-f1bc98379766", 00:17:14.812 "strip_size_kb": 0, 00:17:14.812 "state": "online", 00:17:14.812 "raid_level": "raid1", 00:17:14.812 "superblock": true, 00:17:14.812 "num_base_bdevs": 2, 00:17:14.812 "num_base_bdevs_discovered": 1, 00:17:14.812 "num_base_bdevs_operational": 1, 00:17:14.812 "base_bdevs_list": [ 00:17:14.812 { 00:17:14.812 "name": null, 00:17:14.812 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:14.812 "is_configured": false, 00:17:14.812 "data_offset": 0, 00:17:14.812 "data_size": 7936 00:17:14.812 }, 00:17:14.812 { 00:17:14.812 "name": "BaseBdev2", 00:17:14.812 "uuid": "84f144e7-7f26-5250-8bec-51172e731063", 00:17:14.812 "is_configured": true, 00:17:14.812 "data_offset": 256, 00:17:14.812 "data_size": 7936 00:17:14.812 } 00:17:14.812 ] 00:17:14.812 }' 00:17:14.812 17:38:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:14.812 17:38:45 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:15.384 17:38:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@770 -- # verify_raid_bdev_process raid_bdev1 none none 00:17:15.384 17:38:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:17:15.384 17:38:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:17:15.384 17:38:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=none 00:17:15.384 17:38:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:17:15.384 17:38:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:15.384 17:38:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:15.384 17:38:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:15.384 17:38:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:15.384 17:38:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:15.384 17:38:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:17:15.384 "name": "raid_bdev1", 00:17:15.384 "uuid": "ce088f1b-e78a-4f92-83c0-f1bc98379766", 00:17:15.384 "strip_size_kb": 0, 00:17:15.384 "state": "online", 00:17:15.384 "raid_level": "raid1", 00:17:15.384 "superblock": true, 00:17:15.384 "num_base_bdevs": 2, 00:17:15.384 "num_base_bdevs_discovered": 1, 00:17:15.384 "num_base_bdevs_operational": 1, 00:17:15.384 "base_bdevs_list": [ 00:17:15.384 { 00:17:15.384 "name": null, 00:17:15.384 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:15.384 "is_configured": false, 00:17:15.384 "data_offset": 0, 00:17:15.384 "data_size": 7936 00:17:15.384 }, 00:17:15.384 { 00:17:15.384 "name": "BaseBdev2", 00:17:15.384 "uuid": "84f144e7-7f26-5250-8bec-51172e731063", 00:17:15.384 "is_configured": true, 00:17:15.384 "data_offset": 256, 00:17:15.384 "data_size": 7936 00:17:15.384 } 00:17:15.384 ] 00:17:15.384 }' 00:17:15.384 17:38:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:17:15.384 17:38:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:17:15.384 17:38:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:17:15.384 17:38:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:17:15.384 17:38:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@773 -- # rpc_cmd bdev_passthru_delete BaseBdev1 00:17:15.384 17:38:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:15.384 17:38:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:15.384 17:38:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:15.384 17:38:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@774 -- # rpc_cmd bdev_passthru_create -b BaseBdev1_malloc -p BaseBdev1 00:17:15.384 17:38:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:15.384 17:38:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:15.384 [2024-11-27 17:38:46.501848] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on BaseBdev1_malloc 00:17:15.384 [2024-11-27 17:38:46.501907] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:17:15.384 [2024-11-27 17:38:46.501929] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x61600000a280 00:17:15.384 [2024-11-27 17:38:46.501939] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:17:15.384 [2024-11-27 17:38:46.502099] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:17:15.384 [2024-11-27 17:38:46.502115] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: BaseBdev1 00:17:15.384 [2024-11-27 17:38:46.502169] bdev_raid.c:3897:raid_bdev_examine_cont: *DEBUG*: raid superblock found on bdev BaseBdev1 00:17:15.384 [2024-11-27 17:38:46.502191] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:17:15.384 [2024-11-27 17:38:46.502204] bdev_raid.c:3709:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:17:15.384 [2024-11-27 17:38:46.502218] bdev_raid.c:3884:raid_bdev_examine_done: *ERROR*: Failed to examine bdev BaseBdev1: Invalid argument 00:17:15.384 BaseBdev1 00:17:15.384 17:38:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:15.384 17:38:46 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@775 -- # sleep 1 00:17:16.767 17:38:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@776 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:17:16.767 17:38:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:17:16.767 17:38:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:17:16.767 17:38:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:16.767 17:38:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:16.767 17:38:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:17:16.767 17:38:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:16.767 17:38:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:16.767 17:38:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:16.767 17:38:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:16.767 17:38:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:16.767 17:38:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:16.767 17:38:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:16.767 17:38:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:16.767 17:38:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:16.767 17:38:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:16.767 "name": "raid_bdev1", 00:17:16.767 "uuid": "ce088f1b-e78a-4f92-83c0-f1bc98379766", 00:17:16.767 "strip_size_kb": 0, 00:17:16.767 "state": "online", 00:17:16.767 "raid_level": "raid1", 00:17:16.767 "superblock": true, 00:17:16.767 "num_base_bdevs": 2, 00:17:16.767 "num_base_bdevs_discovered": 1, 00:17:16.767 "num_base_bdevs_operational": 1, 00:17:16.767 "base_bdevs_list": [ 00:17:16.767 { 00:17:16.767 "name": null, 00:17:16.767 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:16.767 "is_configured": false, 00:17:16.767 "data_offset": 0, 00:17:16.767 "data_size": 7936 00:17:16.767 }, 00:17:16.767 { 00:17:16.767 "name": "BaseBdev2", 00:17:16.767 "uuid": "84f144e7-7f26-5250-8bec-51172e731063", 00:17:16.767 "is_configured": true, 00:17:16.767 "data_offset": 256, 00:17:16.767 "data_size": 7936 00:17:16.767 } 00:17:16.767 ] 00:17:16.767 }' 00:17:16.767 17:38:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:16.767 17:38:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:17.027 17:38:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@777 -- # verify_raid_bdev_process raid_bdev1 none none 00:17:17.027 17:38:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:17:17.027 17:38:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:17:17.027 17:38:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=none 00:17:17.027 17:38:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:17:17.027 17:38:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:17.027 17:38:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:17.027 17:38:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:17.027 17:38:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:17.027 17:38:47 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:17.027 17:38:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:17:17.027 "name": "raid_bdev1", 00:17:17.027 "uuid": "ce088f1b-e78a-4f92-83c0-f1bc98379766", 00:17:17.027 "strip_size_kb": 0, 00:17:17.027 "state": "online", 00:17:17.027 "raid_level": "raid1", 00:17:17.027 "superblock": true, 00:17:17.027 "num_base_bdevs": 2, 00:17:17.027 "num_base_bdevs_discovered": 1, 00:17:17.027 "num_base_bdevs_operational": 1, 00:17:17.027 "base_bdevs_list": [ 00:17:17.027 { 00:17:17.027 "name": null, 00:17:17.027 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:17.027 "is_configured": false, 00:17:17.027 "data_offset": 0, 00:17:17.027 "data_size": 7936 00:17:17.027 }, 00:17:17.027 { 00:17:17.027 "name": "BaseBdev2", 00:17:17.027 "uuid": "84f144e7-7f26-5250-8bec-51172e731063", 00:17:17.027 "is_configured": true, 00:17:17.027 "data_offset": 256, 00:17:17.027 "data_size": 7936 00:17:17.027 } 00:17:17.027 ] 00:17:17.027 }' 00:17:17.027 17:38:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:17:17.027 17:38:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:17:17.027 17:38:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:17:17.027 17:38:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:17:17.027 17:38:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@778 -- # NOT rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:17:17.027 17:38:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@650 -- # local es=0 00:17:17.027 17:38:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:17:17.027 17:38:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:17:17.027 17:38:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:17:17.027 17:38:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:17:17.027 17:38:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:17:17.027 17:38:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@653 -- # rpc_cmd bdev_raid_add_base_bdev raid_bdev1 BaseBdev1 00:17:17.028 17:38:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:17.028 17:38:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:17.028 [2024-11-27 17:38:48.107283] bdev_raid.c:3322:raid_bdev_configure_base_bdev: *DEBUG*: bdev BaseBdev1 is claimed 00:17:17.028 [2024-11-27 17:38:48.107426] bdev_raid.c:3690:raid_bdev_examine_sb: *DEBUG*: raid superblock seq_number on bdev BaseBdev1 (1) smaller than existing raid bdev raid_bdev1 (5) 00:17:17.028 [2024-11-27 17:38:48.107438] bdev_raid.c:3709:raid_bdev_examine_sb: *DEBUG*: raid superblock does not contain this bdev's uuid 00:17:17.028 request: 00:17:17.028 { 00:17:17.028 "base_bdev": "BaseBdev1", 00:17:17.028 "raid_bdev": "raid_bdev1", 00:17:17.028 "method": "bdev_raid_add_base_bdev", 00:17:17.028 "req_id": 1 00:17:17.028 } 00:17:17.028 Got JSON-RPC error response 00:17:17.028 response: 00:17:17.028 { 00:17:17.028 "code": -22, 00:17:17.028 "message": "Failed to add base bdev to RAID bdev: Invalid argument" 00:17:17.028 } 00:17:17.028 17:38:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:17:17.028 17:38:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@653 -- # es=1 00:17:17.028 17:38:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:17:17.028 17:38:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:17:17.028 17:38:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:17:17.028 17:38:48 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@779 -- # sleep 1 00:17:17.967 17:38:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@780 -- # verify_raid_bdev_state raid_bdev1 online raid1 0 1 00:17:17.967 17:38:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@103 -- # local raid_bdev_name=raid_bdev1 00:17:17.967 17:38:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@104 -- # local expected_state=online 00:17:17.967 17:38:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@105 -- # local raid_level=raid1 00:17:17.967 17:38:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@106 -- # local strip_size=0 00:17:17.967 17:38:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@107 -- # local num_base_bdevs_operational=1 00:17:17.967 17:38:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@108 -- # local raid_bdev_info 00:17:17.967 17:38:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@109 -- # local num_base_bdevs 00:17:17.967 17:38:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@110 -- # local num_base_bdevs_discovered 00:17:17.967 17:38:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@111 -- # local tmp 00:17:17.967 17:38:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:17.967 17:38:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:17.967 17:38:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:17.967 17:38:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:17.967 17:38:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:18.226 17:38:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@113 -- # raid_bdev_info='{ 00:17:18.226 "name": "raid_bdev1", 00:17:18.226 "uuid": "ce088f1b-e78a-4f92-83c0-f1bc98379766", 00:17:18.226 "strip_size_kb": 0, 00:17:18.226 "state": "online", 00:17:18.226 "raid_level": "raid1", 00:17:18.226 "superblock": true, 00:17:18.226 "num_base_bdevs": 2, 00:17:18.226 "num_base_bdevs_discovered": 1, 00:17:18.226 "num_base_bdevs_operational": 1, 00:17:18.226 "base_bdevs_list": [ 00:17:18.226 { 00:17:18.226 "name": null, 00:17:18.226 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:18.226 "is_configured": false, 00:17:18.226 "data_offset": 0, 00:17:18.226 "data_size": 7936 00:17:18.226 }, 00:17:18.226 { 00:17:18.226 "name": "BaseBdev2", 00:17:18.226 "uuid": "84f144e7-7f26-5250-8bec-51172e731063", 00:17:18.226 "is_configured": true, 00:17:18.226 "data_offset": 256, 00:17:18.226 "data_size": 7936 00:17:18.226 } 00:17:18.226 ] 00:17:18.226 }' 00:17:18.226 17:38:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@115 -- # xtrace_disable 00:17:18.226 17:38:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:18.485 17:38:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@781 -- # verify_raid_bdev_process raid_bdev1 none none 00:17:18.485 17:38:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@169 -- # local raid_bdev_name=raid_bdev1 00:17:18.485 17:38:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@170 -- # local process_type=none 00:17:18.485 17:38:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@171 -- # local target=none 00:17:18.485 17:38:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@172 -- # local raid_bdev_info 00:17:18.485 17:38:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # jq -r '.[] | select(.name == "raid_bdev1")' 00:17:18.485 17:38:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # rpc_cmd bdev_raid_get_bdevs all 00:17:18.485 17:38:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:18.486 17:38:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:18.486 17:38:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:18.486 17:38:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@174 -- # raid_bdev_info='{ 00:17:18.486 "name": "raid_bdev1", 00:17:18.486 "uuid": "ce088f1b-e78a-4f92-83c0-f1bc98379766", 00:17:18.486 "strip_size_kb": 0, 00:17:18.486 "state": "online", 00:17:18.486 "raid_level": "raid1", 00:17:18.486 "superblock": true, 00:17:18.486 "num_base_bdevs": 2, 00:17:18.486 "num_base_bdevs_discovered": 1, 00:17:18.486 "num_base_bdevs_operational": 1, 00:17:18.486 "base_bdevs_list": [ 00:17:18.486 { 00:17:18.486 "name": null, 00:17:18.486 "uuid": "00000000-0000-0000-0000-000000000000", 00:17:18.486 "is_configured": false, 00:17:18.486 "data_offset": 0, 00:17:18.486 "data_size": 7936 00:17:18.486 }, 00:17:18.486 { 00:17:18.486 "name": "BaseBdev2", 00:17:18.486 "uuid": "84f144e7-7f26-5250-8bec-51172e731063", 00:17:18.486 "is_configured": true, 00:17:18.486 "data_offset": 256, 00:17:18.486 "data_size": 7936 00:17:18.486 } 00:17:18.486 ] 00:17:18.486 }' 00:17:18.486 17:38:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # jq -r '.process.type // "none"' 00:17:18.486 17:38:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@176 -- # [[ none == \n\o\n\e ]] 00:17:18.486 17:38:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # jq -r '.process.target // "none"' 00:17:18.486 17:38:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@177 -- # [[ none == \n\o\n\e ]] 00:17:18.486 17:38:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@784 -- # killprocess 99211 00:17:18.486 17:38:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@950 -- # '[' -z 99211 ']' 00:17:18.486 17:38:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@954 -- # kill -0 99211 00:17:18.486 17:38:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@955 -- # uname 00:17:18.746 17:38:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:17:18.746 17:38:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 99211 00:17:18.746 killing process with pid 99211 00:17:18.746 Received shutdown signal, test time was about 60.000000 seconds 00:17:18.746 00:17:18.746 Latency(us) 00:17:18.746 [2024-11-27T17:38:49.938Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:17:18.746 [2024-11-27T17:38:49.938Z] =================================================================================================================== 00:17:18.746 [2024-11-27T17:38:49.938Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:17:18.746 17:38:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:17:18.746 17:38:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:17:18.746 17:38:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@968 -- # echo 'killing process with pid 99211' 00:17:18.746 17:38:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@969 -- # kill 99211 00:17:18.746 [2024-11-27 17:38:49.710812] bdev_raid.c:1383:raid_bdev_fini_start: *DEBUG*: raid_bdev_fini_start 00:17:18.746 [2024-11-27 17:38:49.710934] bdev_raid.c: 492:_raid_bdev_destruct: *DEBUG*: raid_bdev_destruct 00:17:18.746 17:38:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@974 -- # wait 99211 00:17:18.746 [2024-11-27 17:38:49.710980] bdev_raid.c: 469:raid_bdev_io_device_unregister_cb: *DEBUG*: raid bdev base bdevs is 0, going to free all in destruct 00:17:18.746 [2024-11-27 17:38:49.710990] bdev_raid.c: 380:raid_bdev_cleanup: *DEBUG*: raid_bdev_cleanup, 0x617000001580 name raid_bdev1, state offline 00:17:18.746 [2024-11-27 17:38:49.742434] bdev_raid.c:1409:raid_bdev_exit: *DEBUG*: raid_bdev_exit 00:17:19.006 ************************************ 00:17:19.006 END TEST raid_rebuild_test_sb_md_interleaved 00:17:19.006 17:38:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- bdev/bdev_raid.sh@786 -- # return 0 00:17:19.006 00:17:19.006 real 0m16.101s 00:17:19.006 user 0m21.499s 00:17:19.006 sys 0m1.627s 00:17:19.006 17:38:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@1126 -- # xtrace_disable 00:17:19.006 17:38:49 bdev_raid.raid_rebuild_test_sb_md_interleaved -- common/autotest_common.sh@10 -- # set +x 00:17:19.006 ************************************ 00:17:19.006 17:38:50 bdev_raid -- bdev/bdev_raid.sh@1015 -- # trap - EXIT 00:17:19.006 17:38:50 bdev_raid -- bdev/bdev_raid.sh@1016 -- # cleanup 00:17:19.006 17:38:50 bdev_raid -- bdev/bdev_raid.sh@56 -- # '[' -n 99211 ']' 00:17:19.006 17:38:50 bdev_raid -- bdev/bdev_raid.sh@56 -- # ps -p 99211 00:17:19.006 17:38:50 bdev_raid -- bdev/bdev_raid.sh@60 -- # rm -rf /raidtest 00:17:19.006 00:17:19.006 real 10m7.222s 00:17:19.006 user 14m12.908s 00:17:19.006 sys 1m55.362s 00:17:19.006 17:38:50 bdev_raid -- common/autotest_common.sh@1126 -- # xtrace_disable 00:17:19.006 ************************************ 00:17:19.006 END TEST bdev_raid 00:17:19.006 ************************************ 00:17:19.006 17:38:50 bdev_raid -- common/autotest_common.sh@10 -- # set +x 00:17:19.006 17:38:50 -- spdk/autotest.sh@190 -- # run_test spdkcli_raid /home/vagrant/spdk_repo/spdk/test/spdkcli/raid.sh 00:17:19.006 17:38:50 -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:17:19.006 17:38:50 -- common/autotest_common.sh@1107 -- # xtrace_disable 00:17:19.006 17:38:50 -- common/autotest_common.sh@10 -- # set +x 00:17:19.006 ************************************ 00:17:19.006 START TEST spdkcli_raid 00:17:19.006 ************************************ 00:17:19.006 17:38:50 spdkcli_raid -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/spdkcli/raid.sh 00:17:19.266 * Looking for test storage... 00:17:19.266 * Found test storage at /home/vagrant/spdk_repo/spdk/test/spdkcli 00:17:19.266 17:38:50 spdkcli_raid -- common/autotest_common.sh@1680 -- # [[ y == y ]] 00:17:19.266 17:38:50 spdkcli_raid -- common/autotest_common.sh@1681 -- # lcov --version 00:17:19.266 17:38:50 spdkcli_raid -- common/autotest_common.sh@1681 -- # awk '{print $NF}' 00:17:19.266 17:38:50 spdkcli_raid -- common/autotest_common.sh@1681 -- # lt 1.15 2 00:17:19.266 17:38:50 spdkcli_raid -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:17:19.266 17:38:50 spdkcli_raid -- scripts/common.sh@333 -- # local ver1 ver1_l 00:17:19.266 17:38:50 spdkcli_raid -- scripts/common.sh@334 -- # local ver2 ver2_l 00:17:19.266 17:38:50 spdkcli_raid -- scripts/common.sh@336 -- # IFS=.-: 00:17:19.266 17:38:50 spdkcli_raid -- scripts/common.sh@336 -- # read -ra ver1 00:17:19.266 17:38:50 spdkcli_raid -- scripts/common.sh@337 -- # IFS=.-: 00:17:19.266 17:38:50 spdkcli_raid -- scripts/common.sh@337 -- # read -ra ver2 00:17:19.266 17:38:50 spdkcli_raid -- scripts/common.sh@338 -- # local 'op=<' 00:17:19.266 17:38:50 spdkcli_raid -- scripts/common.sh@340 -- # ver1_l=2 00:17:19.266 17:38:50 spdkcli_raid -- scripts/common.sh@341 -- # ver2_l=1 00:17:19.266 17:38:50 spdkcli_raid -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:17:19.266 17:38:50 spdkcli_raid -- scripts/common.sh@344 -- # case "$op" in 00:17:19.266 17:38:50 spdkcli_raid -- scripts/common.sh@345 -- # : 1 00:17:19.266 17:38:50 spdkcli_raid -- scripts/common.sh@364 -- # (( v = 0 )) 00:17:19.266 17:38:50 spdkcli_raid -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:17:19.266 17:38:50 spdkcli_raid -- scripts/common.sh@365 -- # decimal 1 00:17:19.266 17:38:50 spdkcli_raid -- scripts/common.sh@353 -- # local d=1 00:17:19.266 17:38:50 spdkcli_raid -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:17:19.266 17:38:50 spdkcli_raid -- scripts/common.sh@355 -- # echo 1 00:17:19.266 17:38:50 spdkcli_raid -- scripts/common.sh@365 -- # ver1[v]=1 00:17:19.266 17:38:50 spdkcli_raid -- scripts/common.sh@366 -- # decimal 2 00:17:19.266 17:38:50 spdkcli_raid -- scripts/common.sh@353 -- # local d=2 00:17:19.266 17:38:50 spdkcli_raid -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:17:19.266 17:38:50 spdkcli_raid -- scripts/common.sh@355 -- # echo 2 00:17:19.266 17:38:50 spdkcli_raid -- scripts/common.sh@366 -- # ver2[v]=2 00:17:19.266 17:38:50 spdkcli_raid -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:17:19.266 17:38:50 spdkcli_raid -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:17:19.266 17:38:50 spdkcli_raid -- scripts/common.sh@368 -- # return 0 00:17:19.266 17:38:50 spdkcli_raid -- common/autotest_common.sh@1682 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:17:19.266 17:38:50 spdkcli_raid -- common/autotest_common.sh@1694 -- # export 'LCOV_OPTS= 00:17:19.266 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:17:19.266 --rc genhtml_branch_coverage=1 00:17:19.266 --rc genhtml_function_coverage=1 00:17:19.266 --rc genhtml_legend=1 00:17:19.266 --rc geninfo_all_blocks=1 00:17:19.266 --rc geninfo_unexecuted_blocks=1 00:17:19.266 00:17:19.266 ' 00:17:19.266 17:38:50 spdkcli_raid -- common/autotest_common.sh@1694 -- # LCOV_OPTS=' 00:17:19.266 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:17:19.266 --rc genhtml_branch_coverage=1 00:17:19.266 --rc genhtml_function_coverage=1 00:17:19.266 --rc genhtml_legend=1 00:17:19.266 --rc geninfo_all_blocks=1 00:17:19.266 --rc geninfo_unexecuted_blocks=1 00:17:19.266 00:17:19.266 ' 00:17:19.266 17:38:50 spdkcli_raid -- common/autotest_common.sh@1695 -- # export 'LCOV=lcov 00:17:19.266 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:17:19.266 --rc genhtml_branch_coverage=1 00:17:19.266 --rc genhtml_function_coverage=1 00:17:19.266 --rc genhtml_legend=1 00:17:19.266 --rc geninfo_all_blocks=1 00:17:19.266 --rc geninfo_unexecuted_blocks=1 00:17:19.266 00:17:19.266 ' 00:17:19.267 17:38:50 spdkcli_raid -- common/autotest_common.sh@1695 -- # LCOV='lcov 00:17:19.267 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:17:19.267 --rc genhtml_branch_coverage=1 00:17:19.267 --rc genhtml_function_coverage=1 00:17:19.267 --rc genhtml_legend=1 00:17:19.267 --rc geninfo_all_blocks=1 00:17:19.267 --rc geninfo_unexecuted_blocks=1 00:17:19.267 00:17:19.267 ' 00:17:19.267 17:38:50 spdkcli_raid -- spdkcli/raid.sh@9 -- # source /home/vagrant/spdk_repo/spdk/test/spdkcli/common.sh 00:17:19.267 17:38:50 spdkcli_raid -- spdkcli/common.sh@6 -- # spdkcli_job=/home/vagrant/spdk_repo/spdk/test/spdkcli/spdkcli_job.py 00:17:19.267 17:38:50 spdkcli_raid -- spdkcli/common.sh@7 -- # spdk_clear_config_py=/home/vagrant/spdk_repo/spdk/test/json_config/clear_config.py 00:17:19.267 17:38:50 spdkcli_raid -- spdkcli/raid.sh@10 -- # source /home/vagrant/spdk_repo/spdk/test/iscsi_tgt/common.sh 00:17:19.267 17:38:50 spdkcli_raid -- iscsi_tgt/common.sh@9 -- # ISCSI_BRIDGE=iscsi_br 00:17:19.267 17:38:50 spdkcli_raid -- iscsi_tgt/common.sh@10 -- # INITIATOR_INTERFACE=spdk_init_int 00:17:19.267 17:38:50 spdkcli_raid -- iscsi_tgt/common.sh@11 -- # INITIATOR_BRIDGE=init_br 00:17:19.267 17:38:50 spdkcli_raid -- iscsi_tgt/common.sh@12 -- # TARGET_NAMESPACE=spdk_iscsi_ns 00:17:19.267 17:38:50 spdkcli_raid -- iscsi_tgt/common.sh@13 -- # TARGET_NS_CMD=(ip netns exec "$TARGET_NAMESPACE") 00:17:19.267 17:38:50 spdkcli_raid -- iscsi_tgt/common.sh@14 -- # TARGET_INTERFACE=spdk_tgt_int 00:17:19.267 17:38:50 spdkcli_raid -- iscsi_tgt/common.sh@15 -- # TARGET_INTERFACE2=spdk_tgt_int2 00:17:19.267 17:38:50 spdkcli_raid -- iscsi_tgt/common.sh@16 -- # TARGET_BRIDGE=tgt_br 00:17:19.267 17:38:50 spdkcli_raid -- iscsi_tgt/common.sh@17 -- # TARGET_BRIDGE2=tgt_br2 00:17:19.267 17:38:50 spdkcli_raid -- iscsi_tgt/common.sh@20 -- # TARGET_IP=10.0.0.1 00:17:19.267 17:38:50 spdkcli_raid -- iscsi_tgt/common.sh@21 -- # TARGET_IP2=10.0.0.3 00:17:19.267 17:38:50 spdkcli_raid -- iscsi_tgt/common.sh@22 -- # INITIATOR_IP=10.0.0.2 00:17:19.267 17:38:50 spdkcli_raid -- iscsi_tgt/common.sh@23 -- # ISCSI_PORT=3260 00:17:19.267 17:38:50 spdkcli_raid -- iscsi_tgt/common.sh@24 -- # NETMASK=10.0.0.2/32 00:17:19.267 17:38:50 spdkcli_raid -- iscsi_tgt/common.sh@25 -- # INITIATOR_TAG=2 00:17:19.267 17:38:50 spdkcli_raid -- iscsi_tgt/common.sh@26 -- # INITIATOR_NAME=ANY 00:17:19.267 17:38:50 spdkcli_raid -- iscsi_tgt/common.sh@27 -- # PORTAL_TAG=1 00:17:19.267 17:38:50 spdkcli_raid -- iscsi_tgt/common.sh@28 -- # ISCSI_APP=("${TARGET_NS_CMD[@]}" "${ISCSI_APP[@]}") 00:17:19.267 17:38:50 spdkcli_raid -- iscsi_tgt/common.sh@29 -- # ISCSI_TEST_CORE_MASK=0xF 00:17:19.267 17:38:50 spdkcli_raid -- spdkcli/raid.sh@12 -- # MATCH_FILE=spdkcli_raid.test 00:17:19.267 17:38:50 spdkcli_raid -- spdkcli/raid.sh@13 -- # SPDKCLI_BRANCH=/bdevs 00:17:19.267 17:38:50 spdkcli_raid -- spdkcli/raid.sh@14 -- # dirname /home/vagrant/spdk_repo/spdk/test/spdkcli/raid.sh 00:17:19.267 17:38:50 spdkcli_raid -- spdkcli/raid.sh@14 -- # readlink -f /home/vagrant/spdk_repo/spdk/test/spdkcli 00:17:19.267 17:38:50 spdkcli_raid -- spdkcli/raid.sh@14 -- # testdir=/home/vagrant/spdk_repo/spdk/test/spdkcli 00:17:19.267 17:38:50 spdkcli_raid -- spdkcli/raid.sh@15 -- # . /home/vagrant/spdk_repo/spdk/test/spdkcli/common.sh 00:17:19.267 17:38:50 spdkcli_raid -- spdkcli/common.sh@6 -- # spdkcli_job=/home/vagrant/spdk_repo/spdk/test/spdkcli/spdkcli_job.py 00:17:19.267 17:38:50 spdkcli_raid -- spdkcli/common.sh@7 -- # spdk_clear_config_py=/home/vagrant/spdk_repo/spdk/test/json_config/clear_config.py 00:17:19.267 17:38:50 spdkcli_raid -- spdkcli/raid.sh@17 -- # trap cleanup EXIT 00:17:19.267 17:38:50 spdkcli_raid -- spdkcli/raid.sh@19 -- # timing_enter run_spdk_tgt 00:17:19.267 17:38:50 spdkcli_raid -- common/autotest_common.sh@724 -- # xtrace_disable 00:17:19.267 17:38:50 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:17:19.267 17:38:50 spdkcli_raid -- spdkcli/raid.sh@20 -- # run_spdk_tgt 00:17:19.267 17:38:50 spdkcli_raid -- spdkcli/common.sh@27 -- # spdk_tgt_pid=99874 00:17:19.267 17:38:50 spdkcli_raid -- spdkcli/common.sh@26 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt -m 0x3 -p 0 00:17:19.267 17:38:50 spdkcli_raid -- spdkcli/common.sh@28 -- # waitforlisten 99874 00:17:19.267 17:38:50 spdkcli_raid -- common/autotest_common.sh@831 -- # '[' -z 99874 ']' 00:17:19.267 17:38:50 spdkcli_raid -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:17:19.267 17:38:50 spdkcli_raid -- common/autotest_common.sh@836 -- # local max_retries=100 00:17:19.267 17:38:50 spdkcli_raid -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:17:19.267 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:17:19.267 17:38:50 spdkcli_raid -- common/autotest_common.sh@840 -- # xtrace_disable 00:17:19.267 17:38:50 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:17:19.526 [2024-11-27 17:38:50.496662] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:17:19.526 [2024-11-27 17:38:50.497436] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x3 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid99874 ] 00:17:19.526 [2024-11-27 17:38:50.642998] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 2 00:17:19.526 [2024-11-27 17:38:50.688188] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:17:19.526 [2024-11-27 17:38:50.688231] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 1 00:17:20.464 17:38:51 spdkcli_raid -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:17:20.464 17:38:51 spdkcli_raid -- common/autotest_common.sh@864 -- # return 0 00:17:20.464 17:38:51 spdkcli_raid -- spdkcli/raid.sh@21 -- # timing_exit run_spdk_tgt 00:17:20.464 17:38:51 spdkcli_raid -- common/autotest_common.sh@730 -- # xtrace_disable 00:17:20.464 17:38:51 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:17:20.464 17:38:51 spdkcli_raid -- spdkcli/raid.sh@23 -- # timing_enter spdkcli_create_malloc 00:17:20.464 17:38:51 spdkcli_raid -- common/autotest_common.sh@724 -- # xtrace_disable 00:17:20.464 17:38:51 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:17:20.464 17:38:51 spdkcli_raid -- spdkcli/raid.sh@26 -- # /home/vagrant/spdk_repo/spdk/test/spdkcli/spdkcli_job.py ''\''/bdevs/malloc create 8 512 Malloc1'\'' '\''Malloc1'\'' True 00:17:20.464 '\''/bdevs/malloc create 8 512 Malloc2'\'' '\''Malloc2'\'' True 00:17:20.464 ' 00:17:21.888 Executing command: ['/bdevs/malloc create 8 512 Malloc1', 'Malloc1', True] 00:17:21.888 Executing command: ['/bdevs/malloc create 8 512 Malloc2', 'Malloc2', True] 00:17:21.888 17:38:52 spdkcli_raid -- spdkcli/raid.sh@27 -- # timing_exit spdkcli_create_malloc 00:17:21.888 17:38:52 spdkcli_raid -- common/autotest_common.sh@730 -- # xtrace_disable 00:17:21.888 17:38:52 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:17:21.888 17:38:52 spdkcli_raid -- spdkcli/raid.sh@29 -- # timing_enter spdkcli_create_raid 00:17:21.888 17:38:52 spdkcli_raid -- common/autotest_common.sh@724 -- # xtrace_disable 00:17:21.888 17:38:52 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:17:21.888 17:38:53 spdkcli_raid -- spdkcli/raid.sh@31 -- # /home/vagrant/spdk_repo/spdk/test/spdkcli/spdkcli_job.py ''\''/bdevs/raid_volume create testraid 0 "Malloc1 Malloc2" 4'\'' '\''testraid'\'' True 00:17:21.888 ' 00:17:23.283 Executing command: ['/bdevs/raid_volume create testraid 0 "Malloc1 Malloc2" 4', 'testraid', True] 00:17:23.283 17:38:54 spdkcli_raid -- spdkcli/raid.sh@32 -- # timing_exit spdkcli_create_raid 00:17:23.283 17:38:54 spdkcli_raid -- common/autotest_common.sh@730 -- # xtrace_disable 00:17:23.283 17:38:54 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:17:23.283 17:38:54 spdkcli_raid -- spdkcli/raid.sh@34 -- # timing_enter spdkcli_check_match 00:17:23.283 17:38:54 spdkcli_raid -- common/autotest_common.sh@724 -- # xtrace_disable 00:17:23.283 17:38:54 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:17:23.283 17:38:54 spdkcli_raid -- spdkcli/raid.sh@35 -- # check_match 00:17:23.283 17:38:54 spdkcli_raid -- spdkcli/common.sh@44 -- # /home/vagrant/spdk_repo/spdk/scripts/spdkcli.py ll /bdevs 00:17:23.542 17:38:54 spdkcli_raid -- spdkcli/common.sh@45 -- # /home/vagrant/spdk_repo/spdk/test/app/match/match /home/vagrant/spdk_repo/spdk/test/spdkcli/match_files/spdkcli_raid.test.match 00:17:23.801 17:38:54 spdkcli_raid -- spdkcli/common.sh@46 -- # rm -f /home/vagrant/spdk_repo/spdk/test/spdkcli/match_files/spdkcli_raid.test 00:17:23.801 17:38:54 spdkcli_raid -- spdkcli/raid.sh@36 -- # timing_exit spdkcli_check_match 00:17:23.801 17:38:54 spdkcli_raid -- common/autotest_common.sh@730 -- # xtrace_disable 00:17:23.801 17:38:54 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:17:23.801 17:38:54 spdkcli_raid -- spdkcli/raid.sh@38 -- # timing_enter spdkcli_delete_raid 00:17:23.801 17:38:54 spdkcli_raid -- common/autotest_common.sh@724 -- # xtrace_disable 00:17:23.801 17:38:54 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:17:23.801 17:38:54 spdkcli_raid -- spdkcli/raid.sh@40 -- # /home/vagrant/spdk_repo/spdk/test/spdkcli/spdkcli_job.py ''\''/bdevs/raid_volume delete testraid'\'' '\'''\'' True 00:17:23.801 ' 00:17:24.738 Executing command: ['/bdevs/raid_volume delete testraid', '', True] 00:17:24.738 17:38:55 spdkcli_raid -- spdkcli/raid.sh@41 -- # timing_exit spdkcli_delete_raid 00:17:24.738 17:38:55 spdkcli_raid -- common/autotest_common.sh@730 -- # xtrace_disable 00:17:24.738 17:38:55 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:17:24.738 17:38:55 spdkcli_raid -- spdkcli/raid.sh@43 -- # timing_enter spdkcli_delete_malloc 00:17:24.738 17:38:55 spdkcli_raid -- common/autotest_common.sh@724 -- # xtrace_disable 00:17:24.738 17:38:55 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:17:24.738 17:38:55 spdkcli_raid -- spdkcli/raid.sh@46 -- # /home/vagrant/spdk_repo/spdk/test/spdkcli/spdkcli_job.py ''\''/bdevs/malloc delete Malloc1'\'' '\'''\'' True 00:17:24.738 '\''/bdevs/malloc delete Malloc2'\'' '\'''\'' True 00:17:24.738 ' 00:17:26.115 Executing command: ['/bdevs/malloc delete Malloc1', '', True] 00:17:26.115 Executing command: ['/bdevs/malloc delete Malloc2', '', True] 00:17:26.373 17:38:57 spdkcli_raid -- spdkcli/raid.sh@47 -- # timing_exit spdkcli_delete_malloc 00:17:26.373 17:38:57 spdkcli_raid -- common/autotest_common.sh@730 -- # xtrace_disable 00:17:26.373 17:38:57 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:17:26.373 17:38:57 spdkcli_raid -- spdkcli/raid.sh@49 -- # killprocess 99874 00:17:26.373 17:38:57 spdkcli_raid -- common/autotest_common.sh@950 -- # '[' -z 99874 ']' 00:17:26.373 17:38:57 spdkcli_raid -- common/autotest_common.sh@954 -- # kill -0 99874 00:17:26.373 17:38:57 spdkcli_raid -- common/autotest_common.sh@955 -- # uname 00:17:26.373 17:38:57 spdkcli_raid -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:17:26.373 17:38:57 spdkcli_raid -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 99874 00:17:26.373 killing process with pid 99874 00:17:26.373 17:38:57 spdkcli_raid -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:17:26.373 17:38:57 spdkcli_raid -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:17:26.373 17:38:57 spdkcli_raid -- common/autotest_common.sh@968 -- # echo 'killing process with pid 99874' 00:17:26.373 17:38:57 spdkcli_raid -- common/autotest_common.sh@969 -- # kill 99874 00:17:26.373 17:38:57 spdkcli_raid -- common/autotest_common.sh@974 -- # wait 99874 00:17:26.943 17:38:57 spdkcli_raid -- spdkcli/raid.sh@1 -- # cleanup 00:17:26.943 17:38:57 spdkcli_raid -- spdkcli/common.sh@10 -- # '[' -n 99874 ']' 00:17:26.943 Process with pid 99874 is not found 00:17:26.943 17:38:57 spdkcli_raid -- spdkcli/common.sh@11 -- # killprocess 99874 00:17:26.943 17:38:57 spdkcli_raid -- common/autotest_common.sh@950 -- # '[' -z 99874 ']' 00:17:26.943 17:38:57 spdkcli_raid -- common/autotest_common.sh@954 -- # kill -0 99874 00:17:26.943 /home/vagrant/spdk_repo/spdk/test/common/autotest_common.sh: line 954: kill: (99874) - No such process 00:17:26.943 17:38:57 spdkcli_raid -- common/autotest_common.sh@977 -- # echo 'Process with pid 99874 is not found' 00:17:26.943 17:38:57 spdkcli_raid -- spdkcli/common.sh@13 -- # '[' -n '' ']' 00:17:26.943 17:38:57 spdkcli_raid -- spdkcli/common.sh@16 -- # '[' -n '' ']' 00:17:26.943 17:38:57 spdkcli_raid -- spdkcli/common.sh@19 -- # '[' -n '' ']' 00:17:26.943 17:38:57 spdkcli_raid -- spdkcli/common.sh@22 -- # rm -f /home/vagrant/spdk_repo/spdk/test/spdkcli/spdkcli_raid.test /home/vagrant/spdk_repo/spdk/test/spdkcli/match_files/spdkcli_details_vhost.test /tmp/sample_aio 00:17:26.943 00:17:26.943 real 0m7.719s 00:17:26.943 user 0m16.276s 00:17:26.943 sys 0m1.082s 00:17:26.943 17:38:57 spdkcli_raid -- common/autotest_common.sh@1126 -- # xtrace_disable 00:17:26.943 17:38:57 spdkcli_raid -- common/autotest_common.sh@10 -- # set +x 00:17:26.943 ************************************ 00:17:26.943 END TEST spdkcli_raid 00:17:26.943 ************************************ 00:17:26.943 17:38:57 -- spdk/autotest.sh@191 -- # run_test blockdev_raid5f /home/vagrant/spdk_repo/spdk/test/bdev/blockdev.sh raid5f 00:17:26.943 17:38:57 -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:17:26.943 17:38:57 -- common/autotest_common.sh@1107 -- # xtrace_disable 00:17:26.943 17:38:57 -- common/autotest_common.sh@10 -- # set +x 00:17:26.943 ************************************ 00:17:26.943 START TEST blockdev_raid5f 00:17:26.943 ************************************ 00:17:26.943 17:38:57 blockdev_raid5f -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/test/bdev/blockdev.sh raid5f 00:17:26.943 * Looking for test storage... 00:17:26.943 * Found test storage at /home/vagrant/spdk_repo/spdk/test/bdev 00:17:26.943 17:38:58 blockdev_raid5f -- common/autotest_common.sh@1680 -- # [[ y == y ]] 00:17:26.943 17:38:58 blockdev_raid5f -- common/autotest_common.sh@1681 -- # lcov --version 00:17:26.943 17:38:58 blockdev_raid5f -- common/autotest_common.sh@1681 -- # awk '{print $NF}' 00:17:26.943 17:38:58 blockdev_raid5f -- common/autotest_common.sh@1681 -- # lt 1.15 2 00:17:27.204 17:38:58 blockdev_raid5f -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:17:27.204 17:38:58 blockdev_raid5f -- scripts/common.sh@333 -- # local ver1 ver1_l 00:17:27.204 17:38:58 blockdev_raid5f -- scripts/common.sh@334 -- # local ver2 ver2_l 00:17:27.204 17:38:58 blockdev_raid5f -- scripts/common.sh@336 -- # IFS=.-: 00:17:27.204 17:38:58 blockdev_raid5f -- scripts/common.sh@336 -- # read -ra ver1 00:17:27.204 17:38:58 blockdev_raid5f -- scripts/common.sh@337 -- # IFS=.-: 00:17:27.204 17:38:58 blockdev_raid5f -- scripts/common.sh@337 -- # read -ra ver2 00:17:27.204 17:38:58 blockdev_raid5f -- scripts/common.sh@338 -- # local 'op=<' 00:17:27.204 17:38:58 blockdev_raid5f -- scripts/common.sh@340 -- # ver1_l=2 00:17:27.204 17:38:58 blockdev_raid5f -- scripts/common.sh@341 -- # ver2_l=1 00:17:27.204 17:38:58 blockdev_raid5f -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:17:27.204 17:38:58 blockdev_raid5f -- scripts/common.sh@344 -- # case "$op" in 00:17:27.204 17:38:58 blockdev_raid5f -- scripts/common.sh@345 -- # : 1 00:17:27.204 17:38:58 blockdev_raid5f -- scripts/common.sh@364 -- # (( v = 0 )) 00:17:27.204 17:38:58 blockdev_raid5f -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:17:27.204 17:38:58 blockdev_raid5f -- scripts/common.sh@365 -- # decimal 1 00:17:27.204 17:38:58 blockdev_raid5f -- scripts/common.sh@353 -- # local d=1 00:17:27.204 17:38:58 blockdev_raid5f -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:17:27.204 17:38:58 blockdev_raid5f -- scripts/common.sh@355 -- # echo 1 00:17:27.204 17:38:58 blockdev_raid5f -- scripts/common.sh@365 -- # ver1[v]=1 00:17:27.204 17:38:58 blockdev_raid5f -- scripts/common.sh@366 -- # decimal 2 00:17:27.204 17:38:58 blockdev_raid5f -- scripts/common.sh@353 -- # local d=2 00:17:27.204 17:38:58 blockdev_raid5f -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:17:27.204 17:38:58 blockdev_raid5f -- scripts/common.sh@355 -- # echo 2 00:17:27.204 17:38:58 blockdev_raid5f -- scripts/common.sh@366 -- # ver2[v]=2 00:17:27.204 17:38:58 blockdev_raid5f -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:17:27.204 17:38:58 blockdev_raid5f -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:17:27.204 17:38:58 blockdev_raid5f -- scripts/common.sh@368 -- # return 0 00:17:27.204 17:38:58 blockdev_raid5f -- common/autotest_common.sh@1682 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:17:27.204 17:38:58 blockdev_raid5f -- common/autotest_common.sh@1694 -- # export 'LCOV_OPTS= 00:17:27.204 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:17:27.204 --rc genhtml_branch_coverage=1 00:17:27.204 --rc genhtml_function_coverage=1 00:17:27.204 --rc genhtml_legend=1 00:17:27.204 --rc geninfo_all_blocks=1 00:17:27.204 --rc geninfo_unexecuted_blocks=1 00:17:27.204 00:17:27.204 ' 00:17:27.204 17:38:58 blockdev_raid5f -- common/autotest_common.sh@1694 -- # LCOV_OPTS=' 00:17:27.204 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:17:27.204 --rc genhtml_branch_coverage=1 00:17:27.204 --rc genhtml_function_coverage=1 00:17:27.204 --rc genhtml_legend=1 00:17:27.204 --rc geninfo_all_blocks=1 00:17:27.204 --rc geninfo_unexecuted_blocks=1 00:17:27.204 00:17:27.204 ' 00:17:27.204 17:38:58 blockdev_raid5f -- common/autotest_common.sh@1695 -- # export 'LCOV=lcov 00:17:27.204 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:17:27.204 --rc genhtml_branch_coverage=1 00:17:27.204 --rc genhtml_function_coverage=1 00:17:27.204 --rc genhtml_legend=1 00:17:27.204 --rc geninfo_all_blocks=1 00:17:27.204 --rc geninfo_unexecuted_blocks=1 00:17:27.204 00:17:27.204 ' 00:17:27.205 17:38:58 blockdev_raid5f -- common/autotest_common.sh@1695 -- # LCOV='lcov 00:17:27.205 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:17:27.205 --rc genhtml_branch_coverage=1 00:17:27.205 --rc genhtml_function_coverage=1 00:17:27.205 --rc genhtml_legend=1 00:17:27.205 --rc geninfo_all_blocks=1 00:17:27.205 --rc geninfo_unexecuted_blocks=1 00:17:27.205 00:17:27.205 ' 00:17:27.205 17:38:58 blockdev_raid5f -- bdev/blockdev.sh@10 -- # source /home/vagrant/spdk_repo/spdk/test/bdev/nbd_common.sh 00:17:27.205 17:38:58 blockdev_raid5f -- bdev/nbd_common.sh@6 -- # set -e 00:17:27.205 17:38:58 blockdev_raid5f -- bdev/blockdev.sh@12 -- # rpc_py=rpc_cmd 00:17:27.205 17:38:58 blockdev_raid5f -- bdev/blockdev.sh@13 -- # conf_file=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.json 00:17:27.205 17:38:58 blockdev_raid5f -- bdev/blockdev.sh@14 -- # nonenclosed_conf_file=/home/vagrant/spdk_repo/spdk/test/bdev/nonenclosed.json 00:17:27.205 17:38:58 blockdev_raid5f -- bdev/blockdev.sh@15 -- # nonarray_conf_file=/home/vagrant/spdk_repo/spdk/test/bdev/nonarray.json 00:17:27.205 17:38:58 blockdev_raid5f -- bdev/blockdev.sh@17 -- # export RPC_PIPE_TIMEOUT=30 00:17:27.205 17:38:58 blockdev_raid5f -- bdev/blockdev.sh@17 -- # RPC_PIPE_TIMEOUT=30 00:17:27.205 17:38:58 blockdev_raid5f -- bdev/blockdev.sh@20 -- # : 00:17:27.205 17:38:58 blockdev_raid5f -- bdev/blockdev.sh@669 -- # QOS_DEV_1=Malloc_0 00:17:27.205 17:38:58 blockdev_raid5f -- bdev/blockdev.sh@670 -- # QOS_DEV_2=Null_1 00:17:27.205 17:38:58 blockdev_raid5f -- bdev/blockdev.sh@671 -- # QOS_RUN_TIME=5 00:17:27.205 17:38:58 blockdev_raid5f -- bdev/blockdev.sh@673 -- # uname -s 00:17:27.205 17:38:58 blockdev_raid5f -- bdev/blockdev.sh@673 -- # '[' Linux = Linux ']' 00:17:27.205 17:38:58 blockdev_raid5f -- bdev/blockdev.sh@675 -- # PRE_RESERVED_MEM=0 00:17:27.205 17:38:58 blockdev_raid5f -- bdev/blockdev.sh@681 -- # test_type=raid5f 00:17:27.205 17:38:58 blockdev_raid5f -- bdev/blockdev.sh@682 -- # crypto_device= 00:17:27.205 17:38:58 blockdev_raid5f -- bdev/blockdev.sh@683 -- # dek= 00:17:27.205 17:38:58 blockdev_raid5f -- bdev/blockdev.sh@684 -- # env_ctx= 00:17:27.205 17:38:58 blockdev_raid5f -- bdev/blockdev.sh@685 -- # wait_for_rpc= 00:17:27.205 17:38:58 blockdev_raid5f -- bdev/blockdev.sh@686 -- # '[' -n '' ']' 00:17:27.205 17:38:58 blockdev_raid5f -- bdev/blockdev.sh@689 -- # [[ raid5f == bdev ]] 00:17:27.205 17:38:58 blockdev_raid5f -- bdev/blockdev.sh@689 -- # [[ raid5f == crypto_* ]] 00:17:27.205 17:38:58 blockdev_raid5f -- bdev/blockdev.sh@692 -- # start_spdk_tgt 00:17:27.205 17:38:58 blockdev_raid5f -- bdev/blockdev.sh@47 -- # spdk_tgt_pid=100132 00:17:27.205 17:38:58 blockdev_raid5f -- bdev/blockdev.sh@46 -- # /home/vagrant/spdk_repo/spdk/build/bin/spdk_tgt '' '' 00:17:27.205 17:38:58 blockdev_raid5f -- bdev/blockdev.sh@48 -- # trap 'killprocess "$spdk_tgt_pid"; exit 1' SIGINT SIGTERM EXIT 00:17:27.205 17:38:58 blockdev_raid5f -- bdev/blockdev.sh@49 -- # waitforlisten 100132 00:17:27.205 17:38:58 blockdev_raid5f -- common/autotest_common.sh@831 -- # '[' -z 100132 ']' 00:17:27.205 17:38:58 blockdev_raid5f -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:17:27.205 17:38:58 blockdev_raid5f -- common/autotest_common.sh@836 -- # local max_retries=100 00:17:27.205 17:38:58 blockdev_raid5f -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:17:27.205 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:17:27.205 17:38:58 blockdev_raid5f -- common/autotest_common.sh@840 -- # xtrace_disable 00:17:27.205 17:38:58 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:17:27.205 [2024-11-27 17:38:58.277517] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:17:27.205 [2024-11-27 17:38:58.277758] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid100132 ] 00:17:27.465 [2024-11-27 17:38:58.422401] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:17:27.465 [2024-11-27 17:38:58.467474] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:17:28.035 17:38:59 blockdev_raid5f -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:17:28.035 17:38:59 blockdev_raid5f -- common/autotest_common.sh@864 -- # return 0 00:17:28.035 17:38:59 blockdev_raid5f -- bdev/blockdev.sh@693 -- # case "$test_type" in 00:17:28.035 17:38:59 blockdev_raid5f -- bdev/blockdev.sh@725 -- # setup_raid5f_conf 00:17:28.035 17:38:59 blockdev_raid5f -- bdev/blockdev.sh@279 -- # rpc_cmd 00:17:28.035 17:38:59 blockdev_raid5f -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:28.035 17:38:59 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:17:28.035 Malloc0 00:17:28.035 Malloc1 00:17:28.035 Malloc2 00:17:28.035 17:38:59 blockdev_raid5f -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:28.035 17:38:59 blockdev_raid5f -- bdev/blockdev.sh@736 -- # rpc_cmd bdev_wait_for_examine 00:17:28.035 17:38:59 blockdev_raid5f -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:28.035 17:38:59 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:17:28.035 17:38:59 blockdev_raid5f -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:28.035 17:38:59 blockdev_raid5f -- bdev/blockdev.sh@739 -- # cat 00:17:28.035 17:38:59 blockdev_raid5f -- bdev/blockdev.sh@739 -- # rpc_cmd save_subsystem_config -n accel 00:17:28.035 17:38:59 blockdev_raid5f -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:28.035 17:38:59 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:17:28.035 17:38:59 blockdev_raid5f -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:28.035 17:38:59 blockdev_raid5f -- bdev/blockdev.sh@739 -- # rpc_cmd save_subsystem_config -n bdev 00:17:28.035 17:38:59 blockdev_raid5f -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:28.035 17:38:59 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:17:28.035 17:38:59 blockdev_raid5f -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:28.035 17:38:59 blockdev_raid5f -- bdev/blockdev.sh@739 -- # rpc_cmd save_subsystem_config -n iobuf 00:17:28.035 17:38:59 blockdev_raid5f -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:28.035 17:38:59 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:17:28.035 17:38:59 blockdev_raid5f -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:28.035 17:38:59 blockdev_raid5f -- bdev/blockdev.sh@747 -- # mapfile -t bdevs 00:17:28.035 17:38:59 blockdev_raid5f -- bdev/blockdev.sh@747 -- # rpc_cmd bdev_get_bdevs 00:17:28.035 17:38:59 blockdev_raid5f -- bdev/blockdev.sh@747 -- # jq -r '.[] | select(.claimed == false)' 00:17:28.035 17:38:59 blockdev_raid5f -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:28.035 17:38:59 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:17:28.295 17:38:59 blockdev_raid5f -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:28.295 17:38:59 blockdev_raid5f -- bdev/blockdev.sh@748 -- # mapfile -t bdevs_name 00:17:28.295 17:38:59 blockdev_raid5f -- bdev/blockdev.sh@748 -- # printf '%s\n' '{' ' "name": "raid5f",' ' "aliases": [' ' "f96b21bb-3c7f-413d-a0ec-6e412518c70b"' ' ],' ' "product_name": "Raid Volume",' ' "block_size": 512,' ' "num_blocks": 131072,' ' "uuid": "f96b21bb-3c7f-413d-a0ec-6e412518c70b",' ' "assigned_rate_limits": {' ' "rw_ios_per_sec": 0,' ' "rw_mbytes_per_sec": 0,' ' "r_mbytes_per_sec": 0,' ' "w_mbytes_per_sec": 0' ' },' ' "claimed": false,' ' "zoned": false,' ' "supported_io_types": {' ' "read": true,' ' "write": true,' ' "unmap": false,' ' "flush": false,' ' "reset": true,' ' "nvme_admin": false,' ' "nvme_io": false,' ' "nvme_io_md": false,' ' "write_zeroes": true,' ' "zcopy": false,' ' "get_zone_info": false,' ' "zone_management": false,' ' "zone_append": false,' ' "compare": false,' ' "compare_and_write": false,' ' "abort": false,' ' "seek_hole": false,' ' "seek_data": false,' ' "copy": false,' ' "nvme_iov_md": false' ' },' ' "driver_specific": {' ' "raid": {' ' "uuid": "f96b21bb-3c7f-413d-a0ec-6e412518c70b",' ' "strip_size_kb": 2,' ' "state": "online",' ' "raid_level": "raid5f",' ' "superblock": false,' ' "num_base_bdevs": 3,' ' "num_base_bdevs_discovered": 3,' ' "num_base_bdevs_operational": 3,' ' "base_bdevs_list": [' ' {' ' "name": "Malloc0",' ' "uuid": "6f04ac7a-0675-4d84-9d8b-a0759cd6fb48",' ' "is_configured": true,' ' "data_offset": 0,' ' "data_size": 65536' ' },' ' {' ' "name": "Malloc1",' ' "uuid": "84ad0f12-6a54-4566-aa81-a5edc43373d4",' ' "is_configured": true,' ' "data_offset": 0,' ' "data_size": 65536' ' },' ' {' ' "name": "Malloc2",' ' "uuid": "80158261-fa7b-4b9a-826c-5de5f64db6db",' ' "is_configured": true,' ' "data_offset": 0,' ' "data_size": 65536' ' }' ' ]' ' }' ' }' '}' 00:17:28.295 17:38:59 blockdev_raid5f -- bdev/blockdev.sh@748 -- # jq -r .name 00:17:28.295 17:38:59 blockdev_raid5f -- bdev/blockdev.sh@749 -- # bdev_list=("${bdevs_name[@]}") 00:17:28.295 17:38:59 blockdev_raid5f -- bdev/blockdev.sh@751 -- # hello_world_bdev=raid5f 00:17:28.295 17:38:59 blockdev_raid5f -- bdev/blockdev.sh@752 -- # trap - SIGINT SIGTERM EXIT 00:17:28.295 17:38:59 blockdev_raid5f -- bdev/blockdev.sh@753 -- # killprocess 100132 00:17:28.295 17:38:59 blockdev_raid5f -- common/autotest_common.sh@950 -- # '[' -z 100132 ']' 00:17:28.295 17:38:59 blockdev_raid5f -- common/autotest_common.sh@954 -- # kill -0 100132 00:17:28.295 17:38:59 blockdev_raid5f -- common/autotest_common.sh@955 -- # uname 00:17:28.295 17:38:59 blockdev_raid5f -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:17:28.295 17:38:59 blockdev_raid5f -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 100132 00:17:28.295 killing process with pid 100132 00:17:28.295 17:38:59 blockdev_raid5f -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:17:28.295 17:38:59 blockdev_raid5f -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:17:28.295 17:38:59 blockdev_raid5f -- common/autotest_common.sh@968 -- # echo 'killing process with pid 100132' 00:17:28.295 17:38:59 blockdev_raid5f -- common/autotest_common.sh@969 -- # kill 100132 00:17:28.295 17:38:59 blockdev_raid5f -- common/autotest_common.sh@974 -- # wait 100132 00:17:28.864 17:38:59 blockdev_raid5f -- bdev/blockdev.sh@757 -- # trap cleanup SIGINT SIGTERM EXIT 00:17:28.864 17:38:59 blockdev_raid5f -- bdev/blockdev.sh@759 -- # run_test bdev_hello_world /home/vagrant/spdk_repo/spdk/build/examples/hello_bdev --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json -b raid5f '' 00:17:28.864 17:38:59 blockdev_raid5f -- common/autotest_common.sh@1101 -- # '[' 7 -le 1 ']' 00:17:28.864 17:38:59 blockdev_raid5f -- common/autotest_common.sh@1107 -- # xtrace_disable 00:17:28.864 17:38:59 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:17:28.864 ************************************ 00:17:28.864 START TEST bdev_hello_world 00:17:28.864 ************************************ 00:17:28.864 17:38:59 blockdev_raid5f.bdev_hello_world -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/build/examples/hello_bdev --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json -b raid5f '' 00:17:28.864 [2024-11-27 17:38:59.847828] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:17:28.864 [2024-11-27 17:38:59.847943] [ DPDK EAL parameters: hello_bdev --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid100171 ] 00:17:28.864 [2024-11-27 17:38:59.990861] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:17:28.864 [2024-11-27 17:39:00.042560] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:17:29.124 [2024-11-27 17:39:00.235977] hello_bdev.c: 222:hello_start: *NOTICE*: Successfully started the application 00:17:29.124 [2024-11-27 17:39:00.236024] hello_bdev.c: 231:hello_start: *NOTICE*: Opening the bdev raid5f 00:17:29.124 [2024-11-27 17:39:00.236039] hello_bdev.c: 244:hello_start: *NOTICE*: Opening io channel 00:17:29.124 [2024-11-27 17:39:00.236328] hello_bdev.c: 138:hello_write: *NOTICE*: Writing to the bdev 00:17:29.124 [2024-11-27 17:39:00.236453] hello_bdev.c: 117:write_complete: *NOTICE*: bdev io write completed successfully 00:17:29.124 [2024-11-27 17:39:00.236469] hello_bdev.c: 84:hello_read: *NOTICE*: Reading io 00:17:29.124 [2024-11-27 17:39:00.236514] hello_bdev.c: 65:read_complete: *NOTICE*: Read string from bdev : Hello World! 00:17:29.124 00:17:29.124 [2024-11-27 17:39:00.236530] hello_bdev.c: 74:read_complete: *NOTICE*: Stopping app 00:17:29.385 00:17:29.385 real 0m0.716s 00:17:29.385 user 0m0.402s 00:17:29.385 sys 0m0.199s 00:17:29.385 ************************************ 00:17:29.385 END TEST bdev_hello_world 00:17:29.385 ************************************ 00:17:29.385 17:39:00 blockdev_raid5f.bdev_hello_world -- common/autotest_common.sh@1126 -- # xtrace_disable 00:17:29.385 17:39:00 blockdev_raid5f.bdev_hello_world -- common/autotest_common.sh@10 -- # set +x 00:17:29.385 17:39:00 blockdev_raid5f -- bdev/blockdev.sh@760 -- # run_test bdev_bounds bdev_bounds '' 00:17:29.385 17:39:00 blockdev_raid5f -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:17:29.385 17:39:00 blockdev_raid5f -- common/autotest_common.sh@1107 -- # xtrace_disable 00:17:29.385 17:39:00 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:17:29.385 ************************************ 00:17:29.385 START TEST bdev_bounds 00:17:29.385 ************************************ 00:17:29.385 17:39:00 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@1125 -- # bdev_bounds '' 00:17:29.385 17:39:00 blockdev_raid5f.bdev_bounds -- bdev/blockdev.sh@289 -- # bdevio_pid=100197 00:17:29.385 17:39:00 blockdev_raid5f.bdev_bounds -- bdev/blockdev.sh@288 -- # /home/vagrant/spdk_repo/spdk/test/bdev/bdevio/bdevio -w -s 0 --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json '' 00:17:29.385 17:39:00 blockdev_raid5f.bdev_bounds -- bdev/blockdev.sh@290 -- # trap 'cleanup; killprocess $bdevio_pid; exit 1' SIGINT SIGTERM EXIT 00:17:29.385 17:39:00 blockdev_raid5f.bdev_bounds -- bdev/blockdev.sh@291 -- # echo 'Process bdevio pid: 100197' 00:17:29.385 Process bdevio pid: 100197 00:17:29.385 17:39:00 blockdev_raid5f.bdev_bounds -- bdev/blockdev.sh@292 -- # waitforlisten 100197 00:17:29.385 17:39:00 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@831 -- # '[' -z 100197 ']' 00:17:29.385 17:39:00 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:17:29.385 17:39:00 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@836 -- # local max_retries=100 00:17:29.385 17:39:00 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:17:29.385 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:17:29.385 17:39:00 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@840 -- # xtrace_disable 00:17:29.385 17:39:00 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@10 -- # set +x 00:17:29.646 [2024-11-27 17:39:00.646653] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:17:29.646 [2024-11-27 17:39:00.646874] [ DPDK EAL parameters: bdevio --no-shconf -c 0x7 -m 0 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid100197 ] 00:17:29.646 [2024-11-27 17:39:00.795383] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 3 00:17:29.905 [2024-11-27 17:39:00.842022] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 1 00:17:29.905 [2024-11-27 17:39:00.842152] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:17:29.905 [2024-11-27 17:39:00.842308] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 2 00:17:30.473 17:39:01 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:17:30.474 17:39:01 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@864 -- # return 0 00:17:30.474 17:39:01 blockdev_raid5f.bdev_bounds -- bdev/blockdev.sh@293 -- # /home/vagrant/spdk_repo/spdk/test/bdev/bdevio/tests.py perform_tests 00:17:30.474 I/O targets: 00:17:30.474 raid5f: 131072 blocks of 512 bytes (64 MiB) 00:17:30.474 00:17:30.474 00:17:30.474 CUnit - A unit testing framework for C - Version 2.1-3 00:17:30.474 http://cunit.sourceforge.net/ 00:17:30.474 00:17:30.474 00:17:30.474 Suite: bdevio tests on: raid5f 00:17:30.474 Test: blockdev write read block ...passed 00:17:30.474 Test: blockdev write zeroes read block ...passed 00:17:30.474 Test: blockdev write zeroes read no split ...passed 00:17:30.474 Test: blockdev write zeroes read split ...passed 00:17:30.734 Test: blockdev write zeroes read split partial ...passed 00:17:30.734 Test: blockdev reset ...passed 00:17:30.734 Test: blockdev write read 8 blocks ...passed 00:17:30.734 Test: blockdev write read size > 128k ...passed 00:17:30.734 Test: blockdev write read invalid size ...passed 00:17:30.734 Test: blockdev write read offset + nbytes == size of blockdev ...passed 00:17:30.734 Test: blockdev write read offset + nbytes > size of blockdev ...passed 00:17:30.734 Test: blockdev write read max offset ...passed 00:17:30.734 Test: blockdev write read 2 blocks on overlapped address offset ...passed 00:17:30.734 Test: blockdev writev readv 8 blocks ...passed 00:17:30.734 Test: blockdev writev readv 30 x 1block ...passed 00:17:30.734 Test: blockdev writev readv block ...passed 00:17:30.734 Test: blockdev writev readv size > 128k ...passed 00:17:30.734 Test: blockdev writev readv size > 128k in two iovs ...passed 00:17:30.734 Test: blockdev comparev and writev ...passed 00:17:30.734 Test: blockdev nvme passthru rw ...passed 00:17:30.734 Test: blockdev nvme passthru vendor specific ...passed 00:17:30.734 Test: blockdev nvme admin passthru ...passed 00:17:30.734 Test: blockdev copy ...passed 00:17:30.734 00:17:30.734 Run Summary: Type Total Ran Passed Failed Inactive 00:17:30.734 suites 1 1 n/a 0 0 00:17:30.734 tests 23 23 23 0 0 00:17:30.734 asserts 130 130 130 0 n/a 00:17:30.734 00:17:30.734 Elapsed time = 0.319 seconds 00:17:30.734 0 00:17:30.734 17:39:01 blockdev_raid5f.bdev_bounds -- bdev/blockdev.sh@294 -- # killprocess 100197 00:17:30.734 17:39:01 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@950 -- # '[' -z 100197 ']' 00:17:30.734 17:39:01 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@954 -- # kill -0 100197 00:17:30.734 17:39:01 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@955 -- # uname 00:17:30.734 17:39:01 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:17:30.734 17:39:01 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 100197 00:17:30.734 killing process with pid 100197 00:17:30.734 17:39:01 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:17:30.734 17:39:01 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:17:30.734 17:39:01 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@968 -- # echo 'killing process with pid 100197' 00:17:30.734 17:39:01 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@969 -- # kill 100197 00:17:30.734 17:39:01 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@974 -- # wait 100197 00:17:30.995 17:39:02 blockdev_raid5f.bdev_bounds -- bdev/blockdev.sh@295 -- # trap - SIGINT SIGTERM EXIT 00:17:30.995 00:17:30.995 real 0m1.448s 00:17:30.995 user 0m3.421s 00:17:30.995 sys 0m0.359s 00:17:30.995 ************************************ 00:17:30.995 END TEST bdev_bounds 00:17:30.995 ************************************ 00:17:30.995 17:39:02 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@1126 -- # xtrace_disable 00:17:30.995 17:39:02 blockdev_raid5f.bdev_bounds -- common/autotest_common.sh@10 -- # set +x 00:17:30.995 17:39:02 blockdev_raid5f -- bdev/blockdev.sh@761 -- # run_test bdev_nbd nbd_function_test /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json raid5f '' 00:17:30.995 17:39:02 blockdev_raid5f -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:17:30.995 17:39:02 blockdev_raid5f -- common/autotest_common.sh@1107 -- # xtrace_disable 00:17:30.995 17:39:02 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:17:30.995 ************************************ 00:17:30.995 START TEST bdev_nbd 00:17:30.995 ************************************ 00:17:30.995 17:39:02 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@1125 -- # nbd_function_test /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json raid5f '' 00:17:30.995 17:39:02 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@299 -- # uname -s 00:17:30.995 17:39:02 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@299 -- # [[ Linux == Linux ]] 00:17:30.995 17:39:02 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@301 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:17:30.995 17:39:02 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@302 -- # local conf=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.json 00:17:30.995 17:39:02 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@303 -- # bdev_all=('raid5f') 00:17:30.995 17:39:02 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@303 -- # local bdev_all 00:17:30.995 17:39:02 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@304 -- # local bdev_num=1 00:17:30.995 17:39:02 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@308 -- # [[ -e /sys/module/nbd ]] 00:17:30.995 17:39:02 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@310 -- # nbd_all=('/dev/nbd0' '/dev/nbd1' '/dev/nbd10' '/dev/nbd11' '/dev/nbd12' '/dev/nbd13' '/dev/nbd14' '/dev/nbd15' '/dev/nbd2' '/dev/nbd3' '/dev/nbd4' '/dev/nbd5' '/dev/nbd6' '/dev/nbd7' '/dev/nbd8' '/dev/nbd9') 00:17:30.995 17:39:02 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@310 -- # local nbd_all 00:17:30.995 17:39:02 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@311 -- # bdev_num=1 00:17:30.995 17:39:02 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@313 -- # nbd_list=('/dev/nbd0') 00:17:30.995 17:39:02 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@313 -- # local nbd_list 00:17:30.995 17:39:02 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@314 -- # bdev_list=('raid5f') 00:17:30.995 17:39:02 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@314 -- # local bdev_list 00:17:30.995 17:39:02 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@317 -- # nbd_pid=100251 00:17:30.995 17:39:02 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@316 -- # /home/vagrant/spdk_repo/spdk/test/app/bdev_svc/bdev_svc -r /var/tmp/spdk-nbd.sock -i 0 --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json '' 00:17:30.995 17:39:02 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@318 -- # trap 'cleanup; killprocess $nbd_pid' SIGINT SIGTERM EXIT 00:17:30.995 17:39:02 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@319 -- # waitforlisten 100251 /var/tmp/spdk-nbd.sock 00:17:30.995 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock... 00:17:30.995 17:39:02 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@831 -- # '[' -z 100251 ']' 00:17:30.995 17:39:02 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk-nbd.sock 00:17:30.995 17:39:02 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@836 -- # local max_retries=100 00:17:30.995 17:39:02 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock...' 00:17:30.995 17:39:02 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@840 -- # xtrace_disable 00:17:30.995 17:39:02 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@10 -- # set +x 00:17:30.995 [2024-11-27 17:39:02.174719] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:17:30.995 [2024-11-27 17:39:02.174881] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:17:31.255 [2024-11-27 17:39:02.319178] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:17:31.255 [2024-11-27 17:39:02.362834] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:17:31.826 17:39:02 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:17:31.826 17:39:02 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@864 -- # return 0 00:17:31.826 17:39:02 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@321 -- # nbd_rpc_start_stop_verify /var/tmp/spdk-nbd.sock raid5f 00:17:31.826 17:39:02 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@113 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:17:31.826 17:39:02 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@114 -- # bdev_list=('raid5f') 00:17:31.826 17:39:02 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@114 -- # local bdev_list 00:17:31.826 17:39:02 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@116 -- # nbd_start_disks_without_nbd_idx /var/tmp/spdk-nbd.sock raid5f 00:17:31.826 17:39:02 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@22 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:17:31.826 17:39:02 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@23 -- # bdev_list=('raid5f') 00:17:31.826 17:39:02 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@23 -- # local bdev_list 00:17:31.826 17:39:02 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@24 -- # local i 00:17:31.826 17:39:02 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@25 -- # local nbd_device 00:17:31.826 17:39:02 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i = 0 )) 00:17:31.826 17:39:02 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i < 1 )) 00:17:31.826 17:39:02 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@28 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk raid5f 00:17:32.085 17:39:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@28 -- # nbd_device=/dev/nbd0 00:17:32.085 17:39:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@30 -- # basename /dev/nbd0 00:17:32.085 17:39:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@30 -- # waitfornbd nbd0 00:17:32.085 17:39:03 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:17:32.085 17:39:03 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@869 -- # local i 00:17:32.085 17:39:03 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:17:32.085 17:39:03 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:17:32.085 17:39:03 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:17:32.085 17:39:03 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@873 -- # break 00:17:32.085 17:39:03 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:17:32.085 17:39:03 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:17:32.085 17:39:03 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:17:32.085 1+0 records in 00:17:32.085 1+0 records out 00:17:32.085 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000381938 s, 10.7 MB/s 00:17:32.085 17:39:03 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:17:32.085 17:39:03 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@886 -- # size=4096 00:17:32.085 17:39:03 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:17:32.085 17:39:03 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:17:32.085 17:39:03 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@889 -- # return 0 00:17:32.085 17:39:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i++ )) 00:17:32.085 17:39:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@27 -- # (( i < 1 )) 00:17:32.085 17:39:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@118 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:17:32.345 17:39:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@118 -- # nbd_disks_json='[ 00:17:32.345 { 00:17:32.345 "nbd_device": "/dev/nbd0", 00:17:32.345 "bdev_name": "raid5f" 00:17:32.345 } 00:17:32.345 ]' 00:17:32.345 17:39:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@119 -- # nbd_disks_name=($(echo "${nbd_disks_json}" | jq -r '.[] | .nbd_device')) 00:17:32.345 17:39:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@119 -- # echo '[ 00:17:32.345 { 00:17:32.345 "nbd_device": "/dev/nbd0", 00:17:32.345 "bdev_name": "raid5f" 00:17:32.345 } 00:17:32.345 ]' 00:17:32.345 17:39:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@119 -- # jq -r '.[] | .nbd_device' 00:17:32.345 17:39:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@120 -- # nbd_stop_disks /var/tmp/spdk-nbd.sock /dev/nbd0 00:17:32.345 17:39:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:17:32.345 17:39:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:17:32.345 17:39:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@50 -- # local nbd_list 00:17:32.345 17:39:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@51 -- # local i 00:17:32.345 17:39:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:17:32.345 17:39:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd0 00:17:32.605 17:39:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:17:32.605 17:39:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:17:32.605 17:39:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:17:32.605 17:39:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:17:32.605 17:39:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:17:32.605 17:39:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:17:32.605 17:39:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@41 -- # break 00:17:32.605 17:39:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@45 -- # return 0 00:17:32.605 17:39:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@122 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:17:32.605 17:39:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:17:32.605 17:39:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:17:32.865 17:39:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:17:32.865 17:39:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # echo '[]' 00:17:32.865 17:39:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:17:32.865 17:39:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:17:32.865 17:39:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # echo '' 00:17:32.865 17:39:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:17:32.865 17:39:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # true 00:17:32.865 17:39:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # count=0 00:17:32.865 17:39:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@66 -- # echo 0 00:17:32.865 17:39:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@122 -- # count=0 00:17:32.865 17:39:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@123 -- # '[' 0 -ne 0 ']' 00:17:32.865 17:39:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@127 -- # return 0 00:17:32.865 17:39:03 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@322 -- # nbd_rpc_data_verify /var/tmp/spdk-nbd.sock raid5f /dev/nbd0 00:17:32.865 17:39:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@90 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:17:32.865 17:39:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@91 -- # bdev_list=('raid5f') 00:17:32.866 17:39:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@91 -- # local bdev_list 00:17:32.866 17:39:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@92 -- # nbd_list=('/dev/nbd0') 00:17:32.866 17:39:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@92 -- # local nbd_list 00:17:32.866 17:39:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@94 -- # nbd_start_disks /var/tmp/spdk-nbd.sock raid5f /dev/nbd0 00:17:32.866 17:39:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:17:32.866 17:39:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@10 -- # bdev_list=('raid5f') 00:17:32.866 17:39:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@10 -- # local bdev_list 00:17:32.866 17:39:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0') 00:17:32.866 17:39:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@11 -- # local nbd_list 00:17:32.866 17:39:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@12 -- # local i 00:17:32.866 17:39:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:17:32.866 17:39:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:17:32.866 17:39:03 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@15 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk raid5f /dev/nbd0 00:17:33.126 /dev/nbd0 00:17:33.126 17:39:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:17:33.126 17:39:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:17:33.126 17:39:04 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:17:33.126 17:39:04 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@869 -- # local i 00:17:33.126 17:39:04 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:17:33.126 17:39:04 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:17:33.126 17:39:04 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:17:33.126 17:39:04 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@873 -- # break 00:17:33.126 17:39:04 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:17:33.126 17:39:04 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:17:33.126 17:39:04 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdtest bs=4096 count=1 iflag=direct 00:17:33.126 1+0 records in 00:17:33.126 1+0 records out 00:17:33.126 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.00279096 s, 1.5 MB/s 00:17:33.126 17:39:04 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@886 -- # stat -c %s /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:17:33.126 17:39:04 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@886 -- # size=4096 00:17:33.126 17:39:04 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@887 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/nbdtest 00:17:33.126 17:39:04 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:17:33.126 17:39:04 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@889 -- # return 0 00:17:33.126 17:39:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:17:33.126 17:39:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@14 -- # (( i < 1 )) 00:17:33.126 17:39:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@95 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:17:33.126 17:39:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:17:33.126 17:39:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:17:33.387 17:39:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[ 00:17:33.387 { 00:17:33.387 "nbd_device": "/dev/nbd0", 00:17:33.387 "bdev_name": "raid5f" 00:17:33.387 } 00:17:33.387 ]' 00:17:33.387 17:39:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # echo '[ 00:17:33.387 { 00:17:33.387 "nbd_device": "/dev/nbd0", 00:17:33.387 "bdev_name": "raid5f" 00:17:33.387 } 00:17:33.387 ]' 00:17:33.387 17:39:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:17:33.387 17:39:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # nbd_disks_name=/dev/nbd0 00:17:33.387 17:39:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # echo /dev/nbd0 00:17:33.387 17:39:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:17:33.387 17:39:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # count=1 00:17:33.387 17:39:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@66 -- # echo 1 00:17:33.387 17:39:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@95 -- # count=1 00:17:33.387 17:39:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@96 -- # '[' 1 -ne 1 ']' 00:17:33.387 17:39:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@100 -- # nbd_dd_data_verify /dev/nbd0 write 00:17:33.387 17:39:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0') 00:17:33.387 17:39:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@70 -- # local nbd_list 00:17:33.387 17:39:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@71 -- # local operation=write 00:17:33.387 17:39:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@72 -- # local tmp_file=/home/vagrant/spdk_repo/spdk/test/bdev/nbdrandtest 00:17:33.387 17:39:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@74 -- # '[' write = write ']' 00:17:33.387 17:39:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@76 -- # dd if=/dev/urandom of=/home/vagrant/spdk_repo/spdk/test/bdev/nbdrandtest bs=4096 count=256 00:17:33.387 256+0 records in 00:17:33.387 256+0 records out 00:17:33.387 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.00877532 s, 119 MB/s 00:17:33.387 17:39:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:17:33.387 17:39:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@78 -- # dd if=/home/vagrant/spdk_repo/spdk/test/bdev/nbdrandtest of=/dev/nbd0 bs=4096 count=256 oflag=direct 00:17:33.387 256+0 records in 00:17:33.387 256+0 records out 00:17:33.387 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0287109 s, 36.5 MB/s 00:17:33.387 17:39:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@101 -- # nbd_dd_data_verify /dev/nbd0 verify 00:17:33.387 17:39:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0') 00:17:33.387 17:39:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@70 -- # local nbd_list 00:17:33.387 17:39:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@71 -- # local operation=verify 00:17:33.387 17:39:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@72 -- # local tmp_file=/home/vagrant/spdk_repo/spdk/test/bdev/nbdrandtest 00:17:33.387 17:39:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@74 -- # '[' verify = write ']' 00:17:33.387 17:39:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@80 -- # '[' verify = verify ']' 00:17:33.387 17:39:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:17:33.387 17:39:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /home/vagrant/spdk_repo/spdk/test/bdev/nbdrandtest /dev/nbd0 00:17:33.387 17:39:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@85 -- # rm /home/vagrant/spdk_repo/spdk/test/bdev/nbdrandtest 00:17:33.387 17:39:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@103 -- # nbd_stop_disks /var/tmp/spdk-nbd.sock /dev/nbd0 00:17:33.387 17:39:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:17:33.387 17:39:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:17:33.387 17:39:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@50 -- # local nbd_list 00:17:33.387 17:39:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@51 -- # local i 00:17:33.387 17:39:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:17:33.387 17:39:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd0 00:17:33.647 17:39:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:17:33.647 17:39:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:17:33.647 17:39:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:17:33.647 17:39:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:17:33.647 17:39:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:17:33.647 17:39:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:17:33.647 17:39:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@41 -- # break 00:17:33.647 17:39:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@45 -- # return 0 00:17:33.647 17:39:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@104 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:17:33.647 17:39:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:17:33.647 17:39:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@63 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:17:33.907 17:39:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:17:33.907 17:39:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:17:33.907 17:39:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # echo '[]' 00:17:33.907 17:39:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:17:33.907 17:39:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # echo '' 00:17:33.907 17:39:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:17:33.907 17:39:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # true 00:17:33.907 17:39:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@65 -- # count=0 00:17:33.907 17:39:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@66 -- # echo 0 00:17:33.907 17:39:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@104 -- # count=0 00:17:33.907 17:39:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@105 -- # '[' 0 -ne 0 ']' 00:17:33.907 17:39:04 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@109 -- # return 0 00:17:33.907 17:39:05 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@323 -- # nbd_with_lvol_verify /var/tmp/spdk-nbd.sock /dev/nbd0 00:17:33.907 17:39:05 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@131 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:17:33.907 17:39:05 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@132 -- # local nbd=/dev/nbd0 00:17:33.907 17:39:05 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@134 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create -b malloc_lvol_verify 16 512 00:17:34.167 malloc_lvol_verify 00:17:34.167 17:39:05 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@135 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_lvol_create_lvstore malloc_lvol_verify lvs 00:17:34.426 c5ff0a6c-b931-48ec-b458-400d99108ae9 00:17:34.426 17:39:05 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@136 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_lvol_create lvol 4 -l lvs 00:17:34.426 70a50b41-23fa-48ca-b3fa-93d19a0be1c8 00:17:34.426 17:39:05 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@137 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk lvs/lvol /dev/nbd0 00:17:34.686 /dev/nbd0 00:17:34.686 17:39:05 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@139 -- # wait_for_nbd_set_capacity /dev/nbd0 00:17:34.686 17:39:05 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@146 -- # local nbd=nbd0 00:17:34.686 17:39:05 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@148 -- # [[ -e /sys/block/nbd0/size ]] 00:17:34.686 17:39:05 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@150 -- # (( 8192 == 0 )) 00:17:34.686 17:39:05 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@141 -- # mkfs.ext4 /dev/nbd0 00:17:34.686 mke2fs 1.47.0 (5-Feb-2023) 00:17:34.686 Discarding device blocks: 0/4096 done 00:17:34.686 Creating filesystem with 4096 1k blocks and 1024 inodes 00:17:34.686 00:17:34.686 Allocating group tables: 0/1 done 00:17:34.686 Writing inode tables: 0/1 done 00:17:34.686 Creating journal (1024 blocks): done 00:17:34.686 Writing superblocks and filesystem accounting information: 0/1 done 00:17:34.686 00:17:34.686 17:39:05 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@142 -- # nbd_stop_disks /var/tmp/spdk-nbd.sock /dev/nbd0 00:17:34.686 17:39:05 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:17:34.686 17:39:05 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0') 00:17:34.686 17:39:05 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@50 -- # local nbd_list 00:17:34.686 17:39:05 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@51 -- # local i 00:17:34.686 17:39:05 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:17:34.687 17:39:05 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@54 -- # /home/vagrant/spdk_repo/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd0 00:17:34.947 17:39:06 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:17:34.947 17:39:06 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:17:34.947 17:39:06 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:17:34.947 17:39:06 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:17:34.947 17:39:06 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:17:34.947 17:39:06 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:17:34.947 17:39:06 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@41 -- # break 00:17:34.947 17:39:06 blockdev_raid5f.bdev_nbd -- bdev/nbd_common.sh@45 -- # return 0 00:17:34.947 17:39:06 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@325 -- # killprocess 100251 00:17:34.947 17:39:06 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@950 -- # '[' -z 100251 ']' 00:17:34.947 17:39:06 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@954 -- # kill -0 100251 00:17:34.947 17:39:06 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@955 -- # uname 00:17:34.947 17:39:06 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:17:34.947 17:39:06 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 100251 00:17:34.947 17:39:06 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:17:34.947 17:39:06 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:17:34.947 killing process with pid 100251 00:17:34.947 17:39:06 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@968 -- # echo 'killing process with pid 100251' 00:17:34.947 17:39:06 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@969 -- # kill 100251 00:17:34.947 17:39:06 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@974 -- # wait 100251 00:17:35.207 17:39:06 blockdev_raid5f.bdev_nbd -- bdev/blockdev.sh@326 -- # trap - SIGINT SIGTERM EXIT 00:17:35.207 00:17:35.207 real 0m4.279s 00:17:35.207 user 0m6.185s 00:17:35.207 sys 0m1.260s 00:17:35.207 17:39:06 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@1126 -- # xtrace_disable 00:17:35.207 ************************************ 00:17:35.207 END TEST bdev_nbd 00:17:35.207 17:39:06 blockdev_raid5f.bdev_nbd -- common/autotest_common.sh@10 -- # set +x 00:17:35.207 ************************************ 00:17:35.468 17:39:06 blockdev_raid5f -- bdev/blockdev.sh@762 -- # [[ y == y ]] 00:17:35.468 17:39:06 blockdev_raid5f -- bdev/blockdev.sh@763 -- # '[' raid5f = nvme ']' 00:17:35.468 17:39:06 blockdev_raid5f -- bdev/blockdev.sh@763 -- # '[' raid5f = gpt ']' 00:17:35.468 17:39:06 blockdev_raid5f -- bdev/blockdev.sh@767 -- # run_test bdev_fio fio_test_suite '' 00:17:35.468 17:39:06 blockdev_raid5f -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:17:35.468 17:39:06 blockdev_raid5f -- common/autotest_common.sh@1107 -- # xtrace_disable 00:17:35.468 17:39:06 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:17:35.468 ************************************ 00:17:35.468 START TEST bdev_fio 00:17:35.468 ************************************ 00:17:35.468 17:39:06 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1125 -- # fio_test_suite '' 00:17:35.468 17:39:06 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@330 -- # local env_context 00:17:35.468 17:39:06 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@334 -- # pushd /home/vagrant/spdk_repo/spdk/test/bdev 00:17:35.468 /home/vagrant/spdk_repo/spdk/test/bdev /home/vagrant/spdk_repo/spdk 00:17:35.468 17:39:06 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@335 -- # trap 'rm -f ./*.state; popd; exit 1' SIGINT SIGTERM EXIT 00:17:35.468 17:39:06 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@338 -- # echo '' 00:17:35.468 17:39:06 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@338 -- # sed s/--env-context=// 00:17:35.468 17:39:06 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@338 -- # env_context= 00:17:35.468 17:39:06 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@339 -- # fio_config_gen /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio verify AIO '' 00:17:35.468 17:39:06 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1280 -- # local config_file=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio 00:17:35.468 17:39:06 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1281 -- # local workload=verify 00:17:35.468 17:39:06 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1282 -- # local bdev_type=AIO 00:17:35.468 17:39:06 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1283 -- # local env_context= 00:17:35.468 17:39:06 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1284 -- # local fio_dir=/usr/src/fio 00:17:35.468 17:39:06 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1286 -- # '[' -e /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio ']' 00:17:35.468 17:39:06 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1291 -- # '[' -z verify ']' 00:17:35.468 17:39:06 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1295 -- # '[' -n '' ']' 00:17:35.468 17:39:06 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1299 -- # touch /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio 00:17:35.468 17:39:06 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1301 -- # cat 00:17:35.468 17:39:06 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1313 -- # '[' verify == verify ']' 00:17:35.468 17:39:06 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1314 -- # cat 00:17:35.468 17:39:06 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1323 -- # '[' AIO == AIO ']' 00:17:35.468 17:39:06 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1324 -- # /usr/src/fio/fio --version 00:17:35.468 17:39:06 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1324 -- # [[ fio-3.35 == *\f\i\o\-\3* ]] 00:17:35.468 17:39:06 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1325 -- # echo serialize_overlap=1 00:17:35.468 17:39:06 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@340 -- # for b in "${bdevs_name[@]}" 00:17:35.468 17:39:06 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@341 -- # echo '[job_raid5f]' 00:17:35.468 17:39:06 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@342 -- # echo filename=raid5f 00:17:35.468 17:39:06 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@346 -- # local 'fio_params=--ioengine=spdk_bdev --iodepth=8 --bs=4k --runtime=10 /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio --verify_state_save=0 --spdk_json_conf=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.json' 00:17:35.468 17:39:06 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@348 -- # run_test bdev_fio_rw_verify fio_bdev --ioengine=spdk_bdev --iodepth=8 --bs=4k --runtime=10 /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio --verify_state_save=0 --spdk_json_conf=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.json --spdk_mem=0 --aux-path=/home/vagrant/spdk_repo/spdk/../output 00:17:35.468 17:39:06 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1101 -- # '[' 11 -le 1 ']' 00:17:35.468 17:39:06 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1107 -- # xtrace_disable 00:17:35.468 17:39:06 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@10 -- # set +x 00:17:35.468 ************************************ 00:17:35.468 START TEST bdev_fio_rw_verify 00:17:35.468 ************************************ 00:17:35.468 17:39:06 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1125 -- # fio_bdev --ioengine=spdk_bdev --iodepth=8 --bs=4k --runtime=10 /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio --verify_state_save=0 --spdk_json_conf=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.json --spdk_mem=0 --aux-path=/home/vagrant/spdk_repo/spdk/../output 00:17:35.468 17:39:06 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1356 -- # fio_plugin /home/vagrant/spdk_repo/spdk/build/fio/spdk_bdev --ioengine=spdk_bdev --iodepth=8 --bs=4k --runtime=10 /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio --verify_state_save=0 --spdk_json_conf=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.json --spdk_mem=0 --aux-path=/home/vagrant/spdk_repo/spdk/../output 00:17:35.468 17:39:06 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1337 -- # local fio_dir=/usr/src/fio 00:17:35.468 17:39:06 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1339 -- # sanitizers=('libasan' 'libclang_rt.asan') 00:17:35.468 17:39:06 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1339 -- # local sanitizers 00:17:35.468 17:39:06 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1340 -- # local plugin=/home/vagrant/spdk_repo/spdk/build/fio/spdk_bdev 00:17:35.468 17:39:06 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1341 -- # shift 00:17:35.468 17:39:06 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1343 -- # local asan_lib= 00:17:35.468 17:39:06 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1344 -- # for sanitizer in "${sanitizers[@]}" 00:17:35.468 17:39:06 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1345 -- # ldd /home/vagrant/spdk_repo/spdk/build/fio/spdk_bdev 00:17:35.468 17:39:06 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1345 -- # grep libasan 00:17:35.468 17:39:06 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1345 -- # awk '{print $3}' 00:17:35.728 17:39:06 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1345 -- # asan_lib=/usr/lib64/libasan.so.8 00:17:35.728 17:39:06 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1346 -- # [[ -n /usr/lib64/libasan.so.8 ]] 00:17:35.728 17:39:06 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1347 -- # break 00:17:35.729 17:39:06 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1352 -- # LD_PRELOAD='/usr/lib64/libasan.so.8 /home/vagrant/spdk_repo/spdk/build/fio/spdk_bdev' 00:17:35.729 17:39:06 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1352 -- # /usr/src/fio/fio --ioengine=spdk_bdev --iodepth=8 --bs=4k --runtime=10 /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio --verify_state_save=0 --spdk_json_conf=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.json --spdk_mem=0 --aux-path=/home/vagrant/spdk_repo/spdk/../output 00:17:35.729 job_raid5f: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=spdk_bdev, iodepth=8 00:17:35.729 fio-3.35 00:17:35.729 Starting 1 thread 00:17:47.950 00:17:47.950 job_raid5f: (groupid=0, jobs=1): err= 0: pid=100430: Wed Nov 27 17:39:17 2024 00:17:47.950 read: IOPS=12.9k, BW=50.2MiB/s (52.6MB/s)(502MiB/10001msec) 00:17:47.950 slat (nsec): min=16789, max=66817, avg=18104.02, stdev=1611.12 00:17:47.950 clat (usec): min=10, max=356, avg=125.02, stdev=42.85 00:17:47.950 lat (usec): min=29, max=374, avg=143.13, stdev=43.01 00:17:47.950 clat percentiles (usec): 00:17:47.951 | 50.000th=[ 130], 99.000th=[ 204], 99.900th=[ 227], 99.990th=[ 251], 00:17:47.951 | 99.999th=[ 338] 00:17:47.951 write: IOPS=13.5k, BW=52.6MiB/s (55.2MB/s)(520MiB/9880msec); 0 zone resets 00:17:47.951 slat (usec): min=7, max=271, avg=15.74, stdev= 3.63 00:17:47.951 clat (usec): min=55, max=1857, avg=286.10, stdev=42.12 00:17:47.951 lat (usec): min=70, max=2129, avg=301.84, stdev=43.36 00:17:47.951 clat percentiles (usec): 00:17:47.951 | 50.000th=[ 293], 99.000th=[ 359], 99.900th=[ 635], 99.990th=[ 1401], 00:17:47.951 | 99.999th=[ 1745] 00:17:47.951 bw ( KiB/s): min=50576, max=55360, per=98.62%, avg=53170.63, stdev=1721.17, samples=19 00:17:47.951 iops : min=12644, max=13840, avg=13292.63, stdev=430.33, samples=19 00:17:47.951 lat (usec) : 20=0.01%, 50=0.01%, 100=17.24%, 250=40.84%, 500=41.85% 00:17:47.951 lat (usec) : 750=0.04%, 1000=0.02% 00:17:47.951 lat (msec) : 2=0.02% 00:17:47.951 cpu : usr=98.76%, sys=0.59%, ctx=21, majf=0, minf=13565 00:17:47.951 IO depths : 1=7.6%, 2=20.0%, 4=55.1%, 8=17.4%, 16=0.0%, 32=0.0%, >=64=0.0% 00:17:47.951 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:17:47.951 complete : 0=0.0%, 4=90.0%, 8=10.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:17:47.951 issued rwts: total=128523,133166,0,0 short=0,0,0,0 dropped=0,0,0,0 00:17:47.951 latency : target=0, window=0, percentile=100.00%, depth=8 00:17:47.951 00:17:47.951 Run status group 0 (all jobs): 00:17:47.951 READ: bw=50.2MiB/s (52.6MB/s), 50.2MiB/s-50.2MiB/s (52.6MB/s-52.6MB/s), io=502MiB (526MB), run=10001-10001msec 00:17:47.951 WRITE: bw=52.6MiB/s (55.2MB/s), 52.6MiB/s-52.6MiB/s (55.2MB/s-55.2MB/s), io=520MiB (545MB), run=9880-9880msec 00:17:47.951 ----------------------------------------------------- 00:17:47.951 Suppressions used: 00:17:47.951 count bytes template 00:17:47.951 1 7 /usr/src/fio/parse.c 00:17:47.951 651 62496 /usr/src/fio/iolog.c 00:17:47.951 1 8 libtcmalloc_minimal.so 00:17:47.951 1 904 libcrypto.so 00:17:47.951 ----------------------------------------------------- 00:17:47.951 00:17:47.951 00:17:47.951 real 0m11.190s 00:17:47.951 user 0m11.532s 00:17:47.951 sys 0m0.590s 00:17:47.951 17:39:17 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@1126 -- # xtrace_disable 00:17:47.951 17:39:17 blockdev_raid5f.bdev_fio.bdev_fio_rw_verify -- common/autotest_common.sh@10 -- # set +x 00:17:47.951 ************************************ 00:17:47.951 END TEST bdev_fio_rw_verify 00:17:47.951 ************************************ 00:17:47.951 17:39:17 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@349 -- # rm -f 00:17:47.951 17:39:17 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@350 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio 00:17:47.951 17:39:17 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@353 -- # fio_config_gen /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio trim '' '' 00:17:47.951 17:39:17 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1280 -- # local config_file=/home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio 00:17:47.951 17:39:17 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1281 -- # local workload=trim 00:17:47.951 17:39:17 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1282 -- # local bdev_type= 00:17:47.951 17:39:17 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1283 -- # local env_context= 00:17:47.951 17:39:17 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1284 -- # local fio_dir=/usr/src/fio 00:17:47.951 17:39:17 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1286 -- # '[' -e /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio ']' 00:17:47.951 17:39:17 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1291 -- # '[' -z trim ']' 00:17:47.951 17:39:17 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1295 -- # '[' -n '' ']' 00:17:47.951 17:39:17 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1299 -- # touch /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio 00:17:47.951 17:39:17 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1301 -- # cat 00:17:47.951 17:39:17 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1313 -- # '[' trim == verify ']' 00:17:47.951 17:39:17 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1328 -- # '[' trim == trim ']' 00:17:47.951 17:39:17 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1329 -- # echo rw=trimwrite 00:17:47.951 17:39:17 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@354 -- # printf '%s\n' '{' ' "name": "raid5f",' ' "aliases": [' ' "f96b21bb-3c7f-413d-a0ec-6e412518c70b"' ' ],' ' "product_name": "Raid Volume",' ' "block_size": 512,' ' "num_blocks": 131072,' ' "uuid": "f96b21bb-3c7f-413d-a0ec-6e412518c70b",' ' "assigned_rate_limits": {' ' "rw_ios_per_sec": 0,' ' "rw_mbytes_per_sec": 0,' ' "r_mbytes_per_sec": 0,' ' "w_mbytes_per_sec": 0' ' },' ' "claimed": false,' ' "zoned": false,' ' "supported_io_types": {' ' "read": true,' ' "write": true,' ' "unmap": false,' ' "flush": false,' ' "reset": true,' ' "nvme_admin": false,' ' "nvme_io": false,' ' "nvme_io_md": false,' ' "write_zeroes": true,' ' "zcopy": false,' ' "get_zone_info": false,' ' "zone_management": false,' ' "zone_append": false,' ' "compare": false,' ' "compare_and_write": false,' ' "abort": false,' ' "seek_hole": false,' ' "seek_data": false,' ' "copy": false,' ' "nvme_iov_md": false' ' },' ' "driver_specific": {' ' "raid": {' ' "uuid": "f96b21bb-3c7f-413d-a0ec-6e412518c70b",' ' "strip_size_kb": 2,' ' "state": "online",' ' "raid_level": "raid5f",' ' "superblock": false,' ' "num_base_bdevs": 3,' ' "num_base_bdevs_discovered": 3,' ' "num_base_bdevs_operational": 3,' ' "base_bdevs_list": [' ' {' ' "name": "Malloc0",' ' "uuid": "6f04ac7a-0675-4d84-9d8b-a0759cd6fb48",' ' "is_configured": true,' ' "data_offset": 0,' ' "data_size": 65536' ' },' ' {' ' "name": "Malloc1",' ' "uuid": "84ad0f12-6a54-4566-aa81-a5edc43373d4",' ' "is_configured": true,' ' "data_offset": 0,' ' "data_size": 65536' ' },' ' {' ' "name": "Malloc2",' ' "uuid": "80158261-fa7b-4b9a-826c-5de5f64db6db",' ' "is_configured": true,' ' "data_offset": 0,' ' "data_size": 65536' ' }' ' ]' ' }' ' }' '}' 00:17:47.951 17:39:17 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@354 -- # jq -r 'select(.supported_io_types.unmap == true) | .name' 00:17:47.951 17:39:17 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@354 -- # [[ -n '' ]] 00:17:47.951 17:39:17 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@360 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/bdev.fio 00:17:47.951 /home/vagrant/spdk_repo/spdk 00:17:47.951 17:39:17 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@361 -- # popd 00:17:47.951 17:39:17 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@362 -- # trap - SIGINT SIGTERM EXIT 00:17:47.951 17:39:17 blockdev_raid5f.bdev_fio -- bdev/blockdev.sh@363 -- # return 0 00:17:47.951 00:17:47.951 real 0m11.505s 00:17:47.951 user 0m11.657s 00:17:47.951 sys 0m0.749s 00:17:47.951 17:39:17 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@1126 -- # xtrace_disable 00:17:47.951 17:39:17 blockdev_raid5f.bdev_fio -- common/autotest_common.sh@10 -- # set +x 00:17:47.951 ************************************ 00:17:47.951 END TEST bdev_fio 00:17:47.951 ************************************ 00:17:47.951 17:39:17 blockdev_raid5f -- bdev/blockdev.sh@774 -- # trap cleanup SIGINT SIGTERM EXIT 00:17:47.951 17:39:17 blockdev_raid5f -- bdev/blockdev.sh@776 -- # run_test bdev_verify /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json -q 128 -o 4096 -w verify -t 5 -C -m 0x3 '' 00:17:47.951 17:39:17 blockdev_raid5f -- common/autotest_common.sh@1101 -- # '[' 16 -le 1 ']' 00:17:47.951 17:39:17 blockdev_raid5f -- common/autotest_common.sh@1107 -- # xtrace_disable 00:17:47.951 17:39:17 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:17:47.951 ************************************ 00:17:47.951 START TEST bdev_verify 00:17:47.951 ************************************ 00:17:47.951 17:39:18 blockdev_raid5f.bdev_verify -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json -q 128 -o 4096 -w verify -t 5 -C -m 0x3 '' 00:17:47.951 [2024-11-27 17:39:18.097560] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:17:47.951 [2024-11-27 17:39:18.097670] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x3 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid100586 ] 00:17:47.951 [2024-11-27 17:39:18.243098] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 2 00:17:47.951 [2024-11-27 17:39:18.299789] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:17:47.951 [2024-11-27 17:39:18.299917] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 1 00:17:47.952 Running I/O for 5 seconds... 00:17:49.458 15025.00 IOPS, 58.69 MiB/s [2024-11-27T17:39:21.587Z] 13154.50 IOPS, 51.38 MiB/s [2024-11-27T17:39:22.524Z] 12466.33 IOPS, 48.70 MiB/s [2024-11-27T17:39:23.905Z] 12144.50 IOPS, 47.44 MiB/s [2024-11-27T17:39:23.905Z] 11953.00 IOPS, 46.69 MiB/s 00:17:52.713 Latency(us) 00:17:52.713 [2024-11-27T17:39:23.905Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:17:52.713 Job: raid5f (Core Mask 0x1, workload: verify, depth: 128, IO size: 4096) 00:17:52.713 Verification LBA range: start 0x0 length 0x2000 00:17:52.713 raid5f : 5.03 5059.64 19.76 0.00 0.00 37911.94 131.47 34113.06 00:17:52.713 Job: raid5f (Core Mask 0x2, workload: verify, depth: 128, IO size: 4096) 00:17:52.713 Verification LBA range: start 0x2000 length 0x2000 00:17:52.713 raid5f : 5.02 6887.25 26.90 0.00 0.00 27890.45 116.26 34799.90 00:17:52.713 [2024-11-27T17:39:23.905Z] =================================================================================================================== 00:17:52.713 [2024-11-27T17:39:23.905Z] Total : 11946.90 46.67 0.00 0.00 32136.97 116.26 34799.90 00:17:52.713 00:17:52.713 real 0m5.765s 00:17:52.713 user 0m10.713s 00:17:52.713 sys 0m0.236s 00:17:52.713 17:39:23 blockdev_raid5f.bdev_verify -- common/autotest_common.sh@1126 -- # xtrace_disable 00:17:52.714 17:39:23 blockdev_raid5f.bdev_verify -- common/autotest_common.sh@10 -- # set +x 00:17:52.714 ************************************ 00:17:52.714 END TEST bdev_verify 00:17:52.714 ************************************ 00:17:52.714 17:39:23 blockdev_raid5f -- bdev/blockdev.sh@777 -- # run_test bdev_verify_big_io /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json -q 128 -o 65536 -w verify -t 5 -C -m 0x3 '' 00:17:52.714 17:39:23 blockdev_raid5f -- common/autotest_common.sh@1101 -- # '[' 16 -le 1 ']' 00:17:52.714 17:39:23 blockdev_raid5f -- common/autotest_common.sh@1107 -- # xtrace_disable 00:17:52.714 17:39:23 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:17:52.714 ************************************ 00:17:52.714 START TEST bdev_verify_big_io 00:17:52.714 ************************************ 00:17:52.714 17:39:23 blockdev_raid5f.bdev_verify_big_io -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json -q 128 -o 65536 -w verify -t 5 -C -m 0x3 '' 00:17:52.974 [2024-11-27 17:39:23.943682] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:17:52.974 [2024-11-27 17:39:23.943799] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x3 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid100668 ] 00:17:52.974 [2024-11-27 17:39:24.091068] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 2 00:17:52.974 [2024-11-27 17:39:24.146245] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:17:52.974 [2024-11-27 17:39:24.146298] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 1 00:17:53.234 Running I/O for 5 seconds... 00:17:55.566 633.00 IOPS, 39.56 MiB/s [2024-11-27T17:39:27.734Z] 761.00 IOPS, 47.56 MiB/s [2024-11-27T17:39:28.674Z] 803.00 IOPS, 50.19 MiB/s [2024-11-27T17:39:29.615Z] 793.25 IOPS, 49.58 MiB/s [2024-11-27T17:39:29.874Z] 812.40 IOPS, 50.77 MiB/s 00:17:58.682 Latency(us) 00:17:58.682 [2024-11-27T17:39:29.874Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:17:58.682 Job: raid5f (Core Mask 0x1, workload: verify, depth: 128, IO size: 65536) 00:17:58.682 Verification LBA range: start 0x0 length 0x200 00:17:58.682 raid5f : 5.28 361.09 22.57 0.00 0.00 8797270.27 194.96 379135.78 00:17:58.682 Job: raid5f (Core Mask 0x2, workload: verify, depth: 128, IO size: 65536) 00:17:58.682 Verification LBA range: start 0x200 length 0x200 00:17:58.682 raid5f : 5.23 461.09 28.82 0.00 0.00 6947392.76 152.93 307704.40 00:17:58.682 [2024-11-27T17:39:29.874Z] =================================================================================================================== 00:17:58.682 [2024-11-27T17:39:29.874Z] Total : 822.17 51.39 0.00 0.00 7763515.19 152.93 379135.78 00:17:58.682 00:17:58.682 real 0m6.009s 00:17:58.682 user 0m11.204s 00:17:58.682 sys 0m0.238s 00:17:58.682 17:39:29 blockdev_raid5f.bdev_verify_big_io -- common/autotest_common.sh@1126 -- # xtrace_disable 00:17:58.682 17:39:29 blockdev_raid5f.bdev_verify_big_io -- common/autotest_common.sh@10 -- # set +x 00:17:58.682 ************************************ 00:17:58.682 END TEST bdev_verify_big_io 00:17:58.682 ************************************ 00:17:58.943 17:39:29 blockdev_raid5f -- bdev/blockdev.sh@778 -- # run_test bdev_write_zeroes /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json -q 128 -o 4096 -w write_zeroes -t 1 '' 00:17:58.943 17:39:29 blockdev_raid5f -- common/autotest_common.sh@1101 -- # '[' 13 -le 1 ']' 00:17:58.943 17:39:29 blockdev_raid5f -- common/autotest_common.sh@1107 -- # xtrace_disable 00:17:58.943 17:39:29 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:17:58.943 ************************************ 00:17:58.943 START TEST bdev_write_zeroes 00:17:58.943 ************************************ 00:17:58.943 17:39:29 blockdev_raid5f.bdev_write_zeroes -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json -q 128 -o 4096 -w write_zeroes -t 1 '' 00:17:58.943 [2024-11-27 17:39:30.033616] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:17:58.943 [2024-11-27 17:39:30.033735] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid100750 ] 00:17:59.203 [2024-11-27 17:39:30.180863] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:17:59.203 [2024-11-27 17:39:30.228238] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:17:59.462 Running I/O for 1 seconds... 00:18:00.404 30375.00 IOPS, 118.65 MiB/s 00:18:00.404 Latency(us) 00:18:00.404 [2024-11-27T17:39:31.596Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:18:00.404 Job: raid5f (Core Mask 0x1, workload: write_zeroes, depth: 128, IO size: 4096) 00:18:00.404 raid5f : 1.01 30338.41 118.51 0.00 0.00 4206.39 1330.75 5723.67 00:18:00.404 [2024-11-27T17:39:31.596Z] =================================================================================================================== 00:18:00.404 [2024-11-27T17:39:31.596Z] Total : 30338.41 118.51 0.00 0.00 4206.39 1330.75 5723.67 00:18:00.665 00:18:00.665 real 0m1.745s 00:18:00.665 user 0m1.394s 00:18:00.665 sys 0m0.227s 00:18:00.665 17:39:31 blockdev_raid5f.bdev_write_zeroes -- common/autotest_common.sh@1126 -- # xtrace_disable 00:18:00.665 17:39:31 blockdev_raid5f.bdev_write_zeroes -- common/autotest_common.sh@10 -- # set +x 00:18:00.665 ************************************ 00:18:00.665 END TEST bdev_write_zeroes 00:18:00.665 ************************************ 00:18:00.665 17:39:31 blockdev_raid5f -- bdev/blockdev.sh@781 -- # run_test bdev_json_nonenclosed /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/nonenclosed.json -q 128 -o 4096 -w write_zeroes -t 1 '' 00:18:00.665 17:39:31 blockdev_raid5f -- common/autotest_common.sh@1101 -- # '[' 13 -le 1 ']' 00:18:00.665 17:39:31 blockdev_raid5f -- common/autotest_common.sh@1107 -- # xtrace_disable 00:18:00.665 17:39:31 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:18:00.665 ************************************ 00:18:00.665 START TEST bdev_json_nonenclosed 00:18:00.665 ************************************ 00:18:00.665 17:39:31 blockdev_raid5f.bdev_json_nonenclosed -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/nonenclosed.json -q 128 -o 4096 -w write_zeroes -t 1 '' 00:18:00.926 [2024-11-27 17:39:31.857855] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:18:00.926 [2024-11-27 17:39:31.857998] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid100786 ] 00:18:00.926 [2024-11-27 17:39:32.002415] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:18:00.926 [2024-11-27 17:39:32.059619] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:18:00.926 [2024-11-27 17:39:32.059741] json_config.c: 608:json_config_prepare_ctx: *ERROR*: Invalid JSON configuration: not enclosed in {}. 00:18:00.926 [2024-11-27 17:39:32.059768] rpc.c: 190:spdk_rpc_server_finish: *ERROR*: No server listening on provided address: 00:18:00.926 [2024-11-27 17:39:32.059789] app.c:1061:spdk_app_stop: *WARNING*: spdk_app_stop'd on non-zero 00:18:01.187 00:18:01.187 real 0m0.404s 00:18:01.187 user 0m0.170s 00:18:01.187 sys 0m0.130s 00:18:01.187 17:39:32 blockdev_raid5f.bdev_json_nonenclosed -- common/autotest_common.sh@1126 -- # xtrace_disable 00:18:01.187 17:39:32 blockdev_raid5f.bdev_json_nonenclosed -- common/autotest_common.sh@10 -- # set +x 00:18:01.187 ************************************ 00:18:01.187 END TEST bdev_json_nonenclosed 00:18:01.187 ************************************ 00:18:01.187 17:39:32 blockdev_raid5f -- bdev/blockdev.sh@784 -- # run_test bdev_json_nonarray /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/nonarray.json -q 128 -o 4096 -w write_zeroes -t 1 '' 00:18:01.187 17:39:32 blockdev_raid5f -- common/autotest_common.sh@1101 -- # '[' 13 -le 1 ']' 00:18:01.187 17:39:32 blockdev_raid5f -- common/autotest_common.sh@1107 -- # xtrace_disable 00:18:01.187 17:39:32 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:18:01.187 ************************************ 00:18:01.187 START TEST bdev_json_nonarray 00:18:01.187 ************************************ 00:18:01.187 17:39:32 blockdev_raid5f.bdev_json_nonarray -- common/autotest_common.sh@1125 -- # /home/vagrant/spdk_repo/spdk/build/examples/bdevperf --json /home/vagrant/spdk_repo/spdk/test/bdev/nonarray.json -q 128 -o 4096 -w write_zeroes -t 1 '' 00:18:01.187 [2024-11-27 17:39:32.343901] Starting SPDK v24.09.1-pre git sha1 b18e1bd62 / DPDK 22.11.4 initialization... 00:18:01.187 [2024-11-27 17:39:32.344050] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --iova-mode=pa --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid100812 ] 00:18:01.447 [2024-11-27 17:39:32.489816] app.c: 917:spdk_app_start: *NOTICE*: Total cores available: 1 00:18:01.447 [2024-11-27 17:39:32.541585] reactor.c: 990:reactor_run: *NOTICE*: Reactor started on core 0 00:18:01.447 [2024-11-27 17:39:32.541710] json_config.c: 614:json_config_prepare_ctx: *ERROR*: Invalid JSON configuration: 'subsystems' should be an array. 00:18:01.447 [2024-11-27 17:39:32.541732] rpc.c: 190:spdk_rpc_server_finish: *ERROR*: No server listening on provided address: 00:18:01.447 [2024-11-27 17:39:32.541748] app.c:1061:spdk_app_stop: *WARNING*: spdk_app_stop'd on non-zero 00:18:01.708 00:18:01.708 real 0m0.402s 00:18:01.708 user 0m0.172s 00:18:01.708 sys 0m0.126s 00:18:01.708 17:39:32 blockdev_raid5f.bdev_json_nonarray -- common/autotest_common.sh@1126 -- # xtrace_disable 00:18:01.708 17:39:32 blockdev_raid5f.bdev_json_nonarray -- common/autotest_common.sh@10 -- # set +x 00:18:01.708 ************************************ 00:18:01.708 END TEST bdev_json_nonarray 00:18:01.708 ************************************ 00:18:01.708 17:39:32 blockdev_raid5f -- bdev/blockdev.sh@786 -- # [[ raid5f == bdev ]] 00:18:01.708 17:39:32 blockdev_raid5f -- bdev/blockdev.sh@793 -- # [[ raid5f == gpt ]] 00:18:01.708 17:39:32 blockdev_raid5f -- bdev/blockdev.sh@797 -- # [[ raid5f == crypto_sw ]] 00:18:01.708 17:39:32 blockdev_raid5f -- bdev/blockdev.sh@809 -- # trap - SIGINT SIGTERM EXIT 00:18:01.708 17:39:32 blockdev_raid5f -- bdev/blockdev.sh@810 -- # cleanup 00:18:01.708 17:39:32 blockdev_raid5f -- bdev/blockdev.sh@23 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/aiofile 00:18:01.708 17:39:32 blockdev_raid5f -- bdev/blockdev.sh@24 -- # rm -f /home/vagrant/spdk_repo/spdk/test/bdev/bdev.json 00:18:01.708 17:39:32 blockdev_raid5f -- bdev/blockdev.sh@26 -- # [[ raid5f == rbd ]] 00:18:01.708 17:39:32 blockdev_raid5f -- bdev/blockdev.sh@30 -- # [[ raid5f == daos ]] 00:18:01.708 17:39:32 blockdev_raid5f -- bdev/blockdev.sh@34 -- # [[ raid5f = \g\p\t ]] 00:18:01.708 17:39:32 blockdev_raid5f -- bdev/blockdev.sh@40 -- # [[ raid5f == xnvme ]] 00:18:01.708 00:18:01.708 real 0m34.799s 00:18:01.708 user 0m47.260s 00:18:01.708 sys 0m4.579s 00:18:01.708 17:39:32 blockdev_raid5f -- common/autotest_common.sh@1126 -- # xtrace_disable 00:18:01.708 17:39:32 blockdev_raid5f -- common/autotest_common.sh@10 -- # set +x 00:18:01.708 ************************************ 00:18:01.708 END TEST blockdev_raid5f 00:18:01.708 ************************************ 00:18:01.708 17:39:32 -- spdk/autotest.sh@194 -- # uname -s 00:18:01.708 17:39:32 -- spdk/autotest.sh@194 -- # [[ Linux == Linux ]] 00:18:01.708 17:39:32 -- spdk/autotest.sh@195 -- # [[ 0 -eq 1 ]] 00:18:01.708 17:39:32 -- spdk/autotest.sh@195 -- # [[ 0 -eq 1 ]] 00:18:01.708 17:39:32 -- spdk/autotest.sh@207 -- # '[' 0 -eq 1 ']' 00:18:01.708 17:39:32 -- spdk/autotest.sh@252 -- # '[' 0 -eq 1 ']' 00:18:01.708 17:39:32 -- spdk/autotest.sh@256 -- # timing_exit lib 00:18:01.708 17:39:32 -- common/autotest_common.sh@730 -- # xtrace_disable 00:18:01.708 17:39:32 -- common/autotest_common.sh@10 -- # set +x 00:18:01.708 17:39:32 -- spdk/autotest.sh@258 -- # '[' 0 -eq 1 ']' 00:18:01.708 17:39:32 -- spdk/autotest.sh@263 -- # '[' 0 -eq 1 ']' 00:18:01.708 17:39:32 -- spdk/autotest.sh@272 -- # '[' 0 -eq 1 ']' 00:18:01.708 17:39:32 -- spdk/autotest.sh@307 -- # '[' 0 -eq 1 ']' 00:18:01.708 17:39:32 -- spdk/autotest.sh@311 -- # '[' 0 -eq 1 ']' 00:18:01.708 17:39:32 -- spdk/autotest.sh@315 -- # '[' 0 -eq 1 ']' 00:18:01.708 17:39:32 -- spdk/autotest.sh@320 -- # '[' 0 -eq 1 ']' 00:18:01.708 17:39:32 -- spdk/autotest.sh@329 -- # '[' 0 -eq 1 ']' 00:18:01.708 17:39:32 -- spdk/autotest.sh@334 -- # '[' 0 -eq 1 ']' 00:18:01.708 17:39:32 -- spdk/autotest.sh@338 -- # '[' 0 -eq 1 ']' 00:18:01.708 17:39:32 -- spdk/autotest.sh@342 -- # '[' 0 -eq 1 ']' 00:18:01.708 17:39:32 -- spdk/autotest.sh@346 -- # '[' 0 -eq 1 ']' 00:18:01.708 17:39:32 -- spdk/autotest.sh@351 -- # '[' 0 -eq 1 ']' 00:18:01.708 17:39:32 -- spdk/autotest.sh@355 -- # '[' 0 -eq 1 ']' 00:18:01.708 17:39:32 -- spdk/autotest.sh@362 -- # [[ 0 -eq 1 ]] 00:18:01.708 17:39:32 -- spdk/autotest.sh@366 -- # [[ 0 -eq 1 ]] 00:18:01.708 17:39:32 -- spdk/autotest.sh@370 -- # [[ 0 -eq 1 ]] 00:18:01.708 17:39:32 -- spdk/autotest.sh@374 -- # [[ '' -eq 1 ]] 00:18:01.708 17:39:32 -- spdk/autotest.sh@381 -- # trap - SIGINT SIGTERM EXIT 00:18:01.708 17:39:32 -- spdk/autotest.sh@383 -- # timing_enter post_cleanup 00:18:01.708 17:39:32 -- common/autotest_common.sh@724 -- # xtrace_disable 00:18:01.708 17:39:32 -- common/autotest_common.sh@10 -- # set +x 00:18:01.708 17:39:32 -- spdk/autotest.sh@384 -- # autotest_cleanup 00:18:01.708 17:39:32 -- common/autotest_common.sh@1392 -- # local autotest_es=0 00:18:01.708 17:39:32 -- common/autotest_common.sh@1393 -- # xtrace_disable 00:18:01.708 17:39:32 -- common/autotest_common.sh@10 -- # set +x 00:18:04.251 INFO: APP EXITING 00:18:04.251 INFO: killing all VMs 00:18:04.251 INFO: killing vhost app 00:18:04.251 INFO: EXIT DONE 00:18:04.821 0000:00:03.0 (1af4 1001): Active devices: mount@vda:vda2,mount@vda:vda3,mount@vda:vda5, so not binding PCI dev 00:18:04.821 Waiting for block devices as requested 00:18:04.821 0000:00:11.0 (1b36 0010): uio_pci_generic -> nvme 00:18:04.821 0000:00:10.0 (1b36 0010): uio_pci_generic -> nvme 00:18:05.763 0000:00:03.0 (1af4 1001): Active devices: mount@vda:vda2,mount@vda:vda3,mount@vda:vda5, so not binding PCI dev 00:18:05.763 Cleaning 00:18:05.763 Removing: /var/run/dpdk/spdk0/config 00:18:05.763 Removing: /var/run/dpdk/spdk0/fbarray_memseg-2048k-0-0 00:18:05.763 Removing: /var/run/dpdk/spdk0/fbarray_memseg-2048k-0-1 00:18:05.763 Removing: /var/run/dpdk/spdk0/fbarray_memseg-2048k-0-2 00:18:05.763 Removing: /var/run/dpdk/spdk0/fbarray_memseg-2048k-0-3 00:18:05.763 Removing: /var/run/dpdk/spdk0/fbarray_memzone 00:18:05.763 Removing: /var/run/dpdk/spdk0/hugepage_info 00:18:05.763 Removing: /dev/shm/spdk_tgt_trace.pid68812 00:18:05.763 Removing: /var/run/dpdk/spdk0 00:18:05.763 Removing: /var/run/dpdk/spdk_pid100132 00:18:05.763 Removing: /var/run/dpdk/spdk_pid100171 00:18:05.763 Removing: /var/run/dpdk/spdk_pid100197 00:18:05.763 Removing: /var/run/dpdk/spdk_pid100420 00:18:05.763 Removing: /var/run/dpdk/spdk_pid100586 00:18:06.024 Removing: /var/run/dpdk/spdk_pid100668 00:18:06.024 Removing: /var/run/dpdk/spdk_pid100750 00:18:06.024 Removing: /var/run/dpdk/spdk_pid100786 00:18:06.024 Removing: /var/run/dpdk/spdk_pid100812 00:18:06.024 Removing: /var/run/dpdk/spdk_pid68637 00:18:06.024 Removing: /var/run/dpdk/spdk_pid68812 00:18:06.024 Removing: /var/run/dpdk/spdk_pid69019 00:18:06.024 Removing: /var/run/dpdk/spdk_pid69106 00:18:06.024 Removing: /var/run/dpdk/spdk_pid69135 00:18:06.024 Removing: /var/run/dpdk/spdk_pid69252 00:18:06.024 Removing: /var/run/dpdk/spdk_pid69270 00:18:06.024 Removing: /var/run/dpdk/spdk_pid69458 00:18:06.024 Removing: /var/run/dpdk/spdk_pid69538 00:18:06.024 Removing: /var/run/dpdk/spdk_pid69623 00:18:06.024 Removing: /var/run/dpdk/spdk_pid69723 00:18:06.024 Removing: /var/run/dpdk/spdk_pid69809 00:18:06.024 Removing: /var/run/dpdk/spdk_pid69852 00:18:06.024 Removing: /var/run/dpdk/spdk_pid69885 00:18:06.024 Removing: /var/run/dpdk/spdk_pid69960 00:18:06.024 Removing: /var/run/dpdk/spdk_pid70067 00:18:06.024 Removing: /var/run/dpdk/spdk_pid70498 00:18:06.024 Removing: /var/run/dpdk/spdk_pid70551 00:18:06.024 Removing: /var/run/dpdk/spdk_pid70603 00:18:06.024 Removing: /var/run/dpdk/spdk_pid70619 00:18:06.024 Removing: /var/run/dpdk/spdk_pid70695 00:18:06.024 Removing: /var/run/dpdk/spdk_pid70706 00:18:06.024 Removing: /var/run/dpdk/spdk_pid70786 00:18:06.024 Removing: /var/run/dpdk/spdk_pid70791 00:18:06.024 Removing: /var/run/dpdk/spdk_pid70844 00:18:06.024 Removing: /var/run/dpdk/spdk_pid70862 00:18:06.024 Removing: /var/run/dpdk/spdk_pid70910 00:18:06.024 Removing: /var/run/dpdk/spdk_pid70922 00:18:06.024 Removing: /var/run/dpdk/spdk_pid71066 00:18:06.024 Removing: /var/run/dpdk/spdk_pid71103 00:18:06.024 Removing: /var/run/dpdk/spdk_pid71189 00:18:06.024 Removing: /var/run/dpdk/spdk_pid72384 00:18:06.024 Removing: /var/run/dpdk/spdk_pid72579 00:18:06.024 Removing: /var/run/dpdk/spdk_pid72714 00:18:06.024 Removing: /var/run/dpdk/spdk_pid73324 00:18:06.024 Removing: /var/run/dpdk/spdk_pid73524 00:18:06.024 Removing: /var/run/dpdk/spdk_pid73659 00:18:06.024 Removing: /var/run/dpdk/spdk_pid74269 00:18:06.024 Removing: /var/run/dpdk/spdk_pid74587 00:18:06.024 Removing: /var/run/dpdk/spdk_pid74717 00:18:06.024 Removing: /var/run/dpdk/spdk_pid76064 00:18:06.024 Removing: /var/run/dpdk/spdk_pid76306 00:18:06.024 Removing: /var/run/dpdk/spdk_pid76439 00:18:06.024 Removing: /var/run/dpdk/spdk_pid77781 00:18:06.024 Removing: /var/run/dpdk/spdk_pid78023 00:18:06.024 Removing: /var/run/dpdk/spdk_pid78158 00:18:06.024 Removing: /var/run/dpdk/spdk_pid79499 00:18:06.024 Removing: /var/run/dpdk/spdk_pid79939 00:18:06.024 Removing: /var/run/dpdk/spdk_pid80068 00:18:06.024 Removing: /var/run/dpdk/spdk_pid81510 00:18:06.024 Removing: /var/run/dpdk/spdk_pid81758 00:18:06.285 Removing: /var/run/dpdk/spdk_pid81893 00:18:06.285 Removing: /var/run/dpdk/spdk_pid83327 00:18:06.285 Removing: /var/run/dpdk/spdk_pid83582 00:18:06.285 Removing: /var/run/dpdk/spdk_pid83711 00:18:06.285 Removing: /var/run/dpdk/spdk_pid85147 00:18:06.285 Removing: /var/run/dpdk/spdk_pid85623 00:18:06.285 Removing: /var/run/dpdk/spdk_pid85752 00:18:06.285 Removing: /var/run/dpdk/spdk_pid85885 00:18:06.285 Removing: /var/run/dpdk/spdk_pid86285 00:18:06.285 Removing: /var/run/dpdk/spdk_pid87003 00:18:06.285 Removing: /var/run/dpdk/spdk_pid87372 00:18:06.285 Removing: /var/run/dpdk/spdk_pid88044 00:18:06.285 Removing: /var/run/dpdk/spdk_pid88470 00:18:06.285 Removing: /var/run/dpdk/spdk_pid89207 00:18:06.285 Removing: /var/run/dpdk/spdk_pid89599 00:18:06.285 Removing: /var/run/dpdk/spdk_pid91514 00:18:06.285 Removing: /var/run/dpdk/spdk_pid91941 00:18:06.285 Removing: /var/run/dpdk/spdk_pid92370 00:18:06.285 Removing: /var/run/dpdk/spdk_pid94404 00:18:06.285 Removing: /var/run/dpdk/spdk_pid94879 00:18:06.285 Removing: /var/run/dpdk/spdk_pid95364 00:18:06.285 Removing: /var/run/dpdk/spdk_pid96411 00:18:06.285 Removing: /var/run/dpdk/spdk_pid96717 00:18:06.285 Removing: /var/run/dpdk/spdk_pid97643 00:18:06.285 Removing: /var/run/dpdk/spdk_pid97960 00:18:06.285 Removing: /var/run/dpdk/spdk_pid98893 00:18:06.285 Removing: /var/run/dpdk/spdk_pid99211 00:18:06.285 Removing: /var/run/dpdk/spdk_pid99874 00:18:06.285 Clean 00:18:06.285 17:39:37 -- common/autotest_common.sh@1451 -- # return 0 00:18:06.285 17:39:37 -- spdk/autotest.sh@385 -- # timing_exit post_cleanup 00:18:06.285 17:39:37 -- common/autotest_common.sh@730 -- # xtrace_disable 00:18:06.285 17:39:37 -- common/autotest_common.sh@10 -- # set +x 00:18:06.545 17:39:37 -- spdk/autotest.sh@387 -- # timing_exit autotest 00:18:06.545 17:39:37 -- common/autotest_common.sh@730 -- # xtrace_disable 00:18:06.545 17:39:37 -- common/autotest_common.sh@10 -- # set +x 00:18:06.545 17:39:37 -- spdk/autotest.sh@388 -- # chmod a+r /home/vagrant/spdk_repo/spdk/../output/timing.txt 00:18:06.545 17:39:37 -- spdk/autotest.sh@390 -- # [[ -f /home/vagrant/spdk_repo/spdk/../output/udev.log ]] 00:18:06.545 17:39:37 -- spdk/autotest.sh@390 -- # rm -f /home/vagrant/spdk_repo/spdk/../output/udev.log 00:18:06.545 17:39:37 -- spdk/autotest.sh@392 -- # [[ y == y ]] 00:18:06.545 17:39:37 -- spdk/autotest.sh@394 -- # hostname 00:18:06.545 17:39:37 -- spdk/autotest.sh@394 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -c --no-external -d /home/vagrant/spdk_repo/spdk -t fedora39-cloud-1721788873-2326 -o /home/vagrant/spdk_repo/spdk/../output/cov_test.info 00:18:06.805 geninfo: WARNING: invalid characters removed from testname! 00:18:33.375 17:40:01 -- spdk/autotest.sh@395 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -a /home/vagrant/spdk_repo/spdk/../output/cov_base.info -a /home/vagrant/spdk_repo/spdk/../output/cov_test.info -o /home/vagrant/spdk_repo/spdk/../output/cov_total.info 00:18:33.375 17:40:04 -- spdk/autotest.sh@396 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -r /home/vagrant/spdk_repo/spdk/../output/cov_total.info '*/dpdk/*' -o /home/vagrant/spdk_repo/spdk/../output/cov_total.info 00:18:35.288 17:40:06 -- spdk/autotest.sh@400 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -r /home/vagrant/spdk_repo/spdk/../output/cov_total.info --ignore-errors unused,unused '/usr/*' -o /home/vagrant/spdk_repo/spdk/../output/cov_total.info 00:18:37.200 17:40:08 -- spdk/autotest.sh@401 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -r /home/vagrant/spdk_repo/spdk/../output/cov_total.info '*/examples/vmd/*' -o /home/vagrant/spdk_repo/spdk/../output/cov_total.info 00:18:39.110 17:40:10 -- spdk/autotest.sh@402 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -r /home/vagrant/spdk_repo/spdk/../output/cov_total.info '*/app/spdk_lspci/*' -o /home/vagrant/spdk_repo/spdk/../output/cov_total.info 00:18:41.650 17:40:12 -- spdk/autotest.sh@403 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -r /home/vagrant/spdk_repo/spdk/../output/cov_total.info '*/app/spdk_top/*' -o /home/vagrant/spdk_repo/spdk/../output/cov_total.info 00:18:43.562 17:40:14 -- spdk/autotest.sh@404 -- # rm -f cov_base.info cov_test.info OLD_STDOUT OLD_STDERR 00:18:43.562 17:40:14 -- common/autotest_common.sh@1680 -- $ [[ y == y ]] 00:18:43.562 17:40:14 -- common/autotest_common.sh@1681 -- $ lcov --version 00:18:43.562 17:40:14 -- common/autotest_common.sh@1681 -- $ awk '{print $NF}' 00:18:43.562 17:40:14 -- common/autotest_common.sh@1681 -- $ lt 1.15 2 00:18:43.562 17:40:14 -- scripts/common.sh@373 -- $ cmp_versions 1.15 '<' 2 00:18:43.562 17:40:14 -- scripts/common.sh@333 -- $ local ver1 ver1_l 00:18:43.562 17:40:14 -- scripts/common.sh@334 -- $ local ver2 ver2_l 00:18:43.562 17:40:14 -- scripts/common.sh@336 -- $ IFS=.-: 00:18:43.562 17:40:14 -- scripts/common.sh@336 -- $ read -ra ver1 00:18:43.562 17:40:14 -- scripts/common.sh@337 -- $ IFS=.-: 00:18:43.562 17:40:14 -- scripts/common.sh@337 -- $ read -ra ver2 00:18:43.562 17:40:14 -- scripts/common.sh@338 -- $ local 'op=<' 00:18:43.562 17:40:14 -- scripts/common.sh@340 -- $ ver1_l=2 00:18:43.562 17:40:14 -- scripts/common.sh@341 -- $ ver2_l=1 00:18:43.562 17:40:14 -- scripts/common.sh@343 -- $ local lt=0 gt=0 eq=0 v 00:18:43.562 17:40:14 -- scripts/common.sh@344 -- $ case "$op" in 00:18:43.562 17:40:14 -- scripts/common.sh@345 -- $ : 1 00:18:43.563 17:40:14 -- scripts/common.sh@364 -- $ (( v = 0 )) 00:18:43.563 17:40:14 -- scripts/common.sh@364 -- $ (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:18:43.563 17:40:14 -- scripts/common.sh@365 -- $ decimal 1 00:18:43.563 17:40:14 -- scripts/common.sh@353 -- $ local d=1 00:18:43.563 17:40:14 -- scripts/common.sh@354 -- $ [[ 1 =~ ^[0-9]+$ ]] 00:18:43.563 17:40:14 -- scripts/common.sh@355 -- $ echo 1 00:18:43.563 17:40:14 -- scripts/common.sh@365 -- $ ver1[v]=1 00:18:43.563 17:40:14 -- scripts/common.sh@366 -- $ decimal 2 00:18:43.563 17:40:14 -- scripts/common.sh@353 -- $ local d=2 00:18:43.563 17:40:14 -- scripts/common.sh@354 -- $ [[ 2 =~ ^[0-9]+$ ]] 00:18:43.563 17:40:14 -- scripts/common.sh@355 -- $ echo 2 00:18:43.563 17:40:14 -- scripts/common.sh@366 -- $ ver2[v]=2 00:18:43.563 17:40:14 -- scripts/common.sh@367 -- $ (( ver1[v] > ver2[v] )) 00:18:43.563 17:40:14 -- scripts/common.sh@368 -- $ (( ver1[v] < ver2[v] )) 00:18:43.563 17:40:14 -- scripts/common.sh@368 -- $ return 0 00:18:43.563 17:40:14 -- common/autotest_common.sh@1682 -- $ lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:18:43.563 17:40:14 -- common/autotest_common.sh@1694 -- $ export 'LCOV_OPTS= 00:18:43.563 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:18:43.563 --rc genhtml_branch_coverage=1 00:18:43.563 --rc genhtml_function_coverage=1 00:18:43.563 --rc genhtml_legend=1 00:18:43.563 --rc geninfo_all_blocks=1 00:18:43.563 --rc geninfo_unexecuted_blocks=1 00:18:43.563 00:18:43.563 ' 00:18:43.563 17:40:14 -- common/autotest_common.sh@1694 -- $ LCOV_OPTS=' 00:18:43.563 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:18:43.563 --rc genhtml_branch_coverage=1 00:18:43.563 --rc genhtml_function_coverage=1 00:18:43.563 --rc genhtml_legend=1 00:18:43.563 --rc geninfo_all_blocks=1 00:18:43.563 --rc geninfo_unexecuted_blocks=1 00:18:43.563 00:18:43.563 ' 00:18:43.563 17:40:14 -- common/autotest_common.sh@1695 -- $ export 'LCOV=lcov 00:18:43.563 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:18:43.563 --rc genhtml_branch_coverage=1 00:18:43.563 --rc genhtml_function_coverage=1 00:18:43.563 --rc genhtml_legend=1 00:18:43.563 --rc geninfo_all_blocks=1 00:18:43.563 --rc geninfo_unexecuted_blocks=1 00:18:43.563 00:18:43.563 ' 00:18:43.563 17:40:14 -- common/autotest_common.sh@1695 -- $ LCOV='lcov 00:18:43.563 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:18:43.563 --rc genhtml_branch_coverage=1 00:18:43.563 --rc genhtml_function_coverage=1 00:18:43.563 --rc genhtml_legend=1 00:18:43.563 --rc geninfo_all_blocks=1 00:18:43.563 --rc geninfo_unexecuted_blocks=1 00:18:43.563 00:18:43.563 ' 00:18:43.563 17:40:14 -- common/autobuild_common.sh@15 -- $ source /home/vagrant/spdk_repo/spdk/scripts/common.sh 00:18:43.563 17:40:14 -- scripts/common.sh@15 -- $ shopt -s extglob 00:18:43.563 17:40:14 -- scripts/common.sh@544 -- $ [[ -e /bin/wpdk_common.sh ]] 00:18:43.563 17:40:14 -- scripts/common.sh@552 -- $ [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:18:43.563 17:40:14 -- scripts/common.sh@553 -- $ source /etc/opt/spdk-pkgdep/paths/export.sh 00:18:43.563 17:40:14 -- paths/export.sh@2 -- $ PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/vagrant/.local/bin:/home/vagrant/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:18:43.563 17:40:14 -- paths/export.sh@3 -- $ PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/vagrant/.local/bin:/home/vagrant/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:18:43.563 17:40:14 -- paths/export.sh@4 -- $ PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/vagrant/.local/bin:/home/vagrant/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:18:43.563 17:40:14 -- paths/export.sh@5 -- $ export PATH 00:18:43.563 17:40:14 -- paths/export.sh@6 -- $ echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/vagrant/.local/bin:/home/vagrant/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:18:43.563 17:40:14 -- common/autobuild_common.sh@478 -- $ out=/home/vagrant/spdk_repo/spdk/../output 00:18:43.563 17:40:14 -- common/autobuild_common.sh@479 -- $ date +%s 00:18:43.563 17:40:14 -- common/autobuild_common.sh@479 -- $ mktemp -dt spdk_1732729214.XXXXXX 00:18:43.563 17:40:14 -- common/autobuild_common.sh@479 -- $ SPDK_WORKSPACE=/tmp/spdk_1732729214.ThkSOE 00:18:43.563 17:40:14 -- common/autobuild_common.sh@481 -- $ [[ -n '' ]] 00:18:43.563 17:40:14 -- common/autobuild_common.sh@485 -- $ '[' -n v22.11.4 ']' 00:18:43.563 17:40:14 -- common/autobuild_common.sh@486 -- $ dirname /home/vagrant/spdk_repo/dpdk/build 00:18:43.563 17:40:14 -- common/autobuild_common.sh@486 -- $ scanbuild_exclude=' --exclude /home/vagrant/spdk_repo/dpdk' 00:18:43.563 17:40:14 -- common/autobuild_common.sh@492 -- $ scanbuild_exclude+=' --exclude /home/vagrant/spdk_repo/spdk/xnvme --exclude /tmp' 00:18:43.563 17:40:14 -- common/autobuild_common.sh@494 -- $ scanbuild='scan-build -o /home/vagrant/spdk_repo/spdk/../output/scan-build-tmp --exclude /home/vagrant/spdk_repo/dpdk --exclude /home/vagrant/spdk_repo/spdk/xnvme --exclude /tmp --status-bugs' 00:18:43.563 17:40:14 -- common/autobuild_common.sh@495 -- $ get_config_params 00:18:43.563 17:40:14 -- common/autotest_common.sh@407 -- $ xtrace_disable 00:18:43.563 17:40:14 -- common/autotest_common.sh@10 -- $ set +x 00:18:43.563 17:40:14 -- common/autobuild_common.sh@495 -- $ config_params='--enable-debug --enable-werror --with-rdma --with-idxd --with-fio=/usr/src/fio --with-iscsi-initiator --disable-unit-tests --enable-ubsan --enable-asan --enable-coverage --with-ublk --with-raid5f --with-dpdk=/home/vagrant/spdk_repo/dpdk/build' 00:18:43.563 17:40:14 -- common/autobuild_common.sh@497 -- $ start_monitor_resources 00:18:43.563 17:40:14 -- pm/common@17 -- $ local monitor 00:18:43.563 17:40:14 -- pm/common@19 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:18:43.563 17:40:14 -- pm/common@19 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:18:43.563 17:40:14 -- pm/common@25 -- $ sleep 1 00:18:43.563 17:40:14 -- pm/common@21 -- $ date +%s 00:18:43.563 17:40:14 -- pm/common@21 -- $ date +%s 00:18:43.563 17:40:14 -- pm/common@21 -- $ /home/vagrant/spdk_repo/spdk/scripts/perf/pm/collect-cpu-load -d /home/vagrant/spdk_repo/spdk/../output/power -l -p monitor.autopackage.sh.1732729214 00:18:43.563 17:40:14 -- pm/common@21 -- $ /home/vagrant/spdk_repo/spdk/scripts/perf/pm/collect-vmstat -d /home/vagrant/spdk_repo/spdk/../output/power -l -p monitor.autopackage.sh.1732729214 00:18:43.563 Redirecting to /home/vagrant/spdk_repo/spdk/../output/power/monitor.autopackage.sh.1732729214_collect-cpu-load.pm.log 00:18:43.563 Redirecting to /home/vagrant/spdk_repo/spdk/../output/power/monitor.autopackage.sh.1732729214_collect-vmstat.pm.log 00:18:44.505 17:40:15 -- common/autobuild_common.sh@498 -- $ trap stop_monitor_resources EXIT 00:18:44.505 17:40:15 -- spdk/autopackage.sh@10 -- $ [[ 0 -eq 1 ]] 00:18:44.505 17:40:15 -- spdk/autopackage.sh@14 -- $ timing_finish 00:18:44.505 17:40:15 -- common/autotest_common.sh@736 -- $ flamegraph=/usr/local/FlameGraph/flamegraph.pl 00:18:44.505 17:40:15 -- common/autotest_common.sh@737 -- $ [[ -x /usr/local/FlameGraph/flamegraph.pl ]] 00:18:44.505 17:40:15 -- common/autotest_common.sh@740 -- $ /usr/local/FlameGraph/flamegraph.pl --title 'Build Timing' --nametype Step: --countname seconds /home/vagrant/spdk_repo/spdk/../output/timing.txt 00:18:44.766 17:40:15 -- spdk/autopackage.sh@1 -- $ stop_monitor_resources 00:18:44.766 17:40:15 -- pm/common@29 -- $ signal_monitor_resources TERM 00:18:44.766 17:40:15 -- pm/common@40 -- $ local monitor pid pids signal=TERM 00:18:44.766 17:40:15 -- pm/common@42 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:18:44.766 17:40:15 -- pm/common@43 -- $ [[ -e /home/vagrant/spdk_repo/spdk/../output/power/collect-cpu-load.pid ]] 00:18:44.766 17:40:15 -- pm/common@44 -- $ pid=102374 00:18:44.766 17:40:15 -- pm/common@50 -- $ kill -TERM 102374 00:18:44.766 17:40:15 -- pm/common@42 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:18:44.766 17:40:15 -- pm/common@43 -- $ [[ -e /home/vagrant/spdk_repo/spdk/../output/power/collect-vmstat.pid ]] 00:18:44.766 17:40:15 -- pm/common@44 -- $ pid=102376 00:18:44.766 17:40:15 -- pm/common@50 -- $ kill -TERM 102376 00:18:44.766 + [[ -n 6165 ]] 00:18:44.766 + sudo kill 6165 00:18:44.776 [Pipeline] } 00:18:44.794 [Pipeline] // timeout 00:18:44.801 [Pipeline] } 00:18:44.816 [Pipeline] // stage 00:18:44.823 [Pipeline] } 00:18:44.839 [Pipeline] // catchError 00:18:44.850 [Pipeline] stage 00:18:44.852 [Pipeline] { (Stop VM) 00:18:44.867 [Pipeline] sh 00:18:45.156 + vagrant halt 00:18:47.800 ==> default: Halting domain... 00:18:55.948 [Pipeline] sh 00:18:56.232 + vagrant destroy -f 00:18:58.772 ==> default: Removing domain... 00:18:58.785 [Pipeline] sh 00:18:59.069 + mv output /var/jenkins/workspace/raid-vg-autotest/output 00:18:59.079 [Pipeline] } 00:18:59.093 [Pipeline] // stage 00:18:59.099 [Pipeline] } 00:18:59.113 [Pipeline] // dir 00:18:59.118 [Pipeline] } 00:18:59.131 [Pipeline] // wrap 00:18:59.137 [Pipeline] } 00:18:59.149 [Pipeline] // catchError 00:18:59.159 [Pipeline] stage 00:18:59.161 [Pipeline] { (Epilogue) 00:18:59.175 [Pipeline] sh 00:18:59.461 + jbp/jenkins/jjb-config/jobs/scripts/compress_artifacts.sh 00:19:03.677 [Pipeline] catchError 00:19:03.679 [Pipeline] { 00:19:03.694 [Pipeline] sh 00:19:03.980 + jbp/jenkins/jjb-config/jobs/scripts/check_artifacts_size.sh 00:19:03.980 Artifacts sizes are good 00:19:03.989 [Pipeline] } 00:19:04.003 [Pipeline] // catchError 00:19:04.014 [Pipeline] archiveArtifacts 00:19:04.021 Archiving artifacts 00:19:04.119 [Pipeline] cleanWs 00:19:04.132 [WS-CLEANUP] Deleting project workspace... 00:19:04.132 [WS-CLEANUP] Deferred wipeout is used... 00:19:04.139 [WS-CLEANUP] done 00:19:04.141 [Pipeline] } 00:19:04.157 [Pipeline] // stage 00:19:04.162 [Pipeline] } 00:19:04.176 [Pipeline] // node 00:19:04.182 [Pipeline] End of Pipeline 00:19:04.224 Finished: SUCCESS